Pouch latent tree models (PLTMs) are a class of probabilistic graphical models that generalizes the Gaussian mixture models (GMMs). PLTMs produce multiple clusterings simultaneously and have been shown better than GMMs for cluster analysis in previous studies. However, due to the considerably higher number of possible structures, the training of PLTMs is more time-demanding than GMMs. This thus has limited the application of PLTMs on only small data sets. In this paper, we consider using GPUs to exploit two parallelism opportunities, namely data parallelism and element-wise parallelism, for PTLMs. We focus on clique tree propagation, since this exact inference procedure is a strenuous task and is recurrently called for each data sample and each model structure during PLTM training. Our experiments with real-world data sets show that the GPU-accelerated implementation procedure can achieve up to 52x speedup over the sequential implementation running on CPUs. The experiment results signify promising potential for further improvement on the full training of PLTMs with GPUs. Copyright © 2018 IFIP International Federation for Information Processing.
|Title of host publication||Network and parallel computing: 15th IFIP WG 10.3 International Conference, NPC 2018, Muroran, Japan, November 29 – December 1, 2018, Proceedings|
|Editors||Feng ZHANG, Jidong ZHAI, Marc SNIR, Hai JIN, Hironori KASAHARA, Mateo VALERO|
|Place of Publication||Cham|
|Publication status||Published - 2018|
CitationPoon, L. K. M. (2018). GPU-accelerated clique tree propagation for pouch latent tree models. In F. Zhang, J. Zhai, M. Snir, H. Jin, H. Kasahara, & M. Valero (Eds.), Network and parallel computing: 15th IFIP WG 10.3 International Conference, NPC 2018, Muroran, Japan, November 29 – December 1, 2018, Proceedings (pp. 90-102). Cham: Springer.
- GPU acceleration
- Clique tree propagation
- Pouch latent tree models
- Parallel computing
- Probabilistic graphical models