Abstract
Sum-Product networks (SPNs) are expressive deep architectures for representing probability distributions, yet allowing exact and efficient inference. SPNs have been successfully applied in several domains, however always as black-box distribution estimators. In this paper, we argue that due to their recursive definition, SPNs can also be naturally employed as hierarchical feature extractors and thus for unsupervised representation learning. Moreover, when converted into Max-Product Networks (MPNs), it is possible to decode such representations back into the original input space. In this way, MPNs can be interpreted as a kind of generative autoencoder, even if they were never trained to reconstruct the input data. We show how these learned representations, if visualized, indeed correspond to "meaningful parts" of the training data. They also yield a large improvement when used in structured prediction tasks. As shown in extensive experiments, SPN and MPN encoding and decoding schemes prove very competitive against the ones employing RBMs and other stacked autoencoder architectures.
| Original language | English |
|---|---|
| Title of host publication | International Conference on Learning Representations (ICLR 2017) - Workshop Track |
| Number of pages | 10 |
| Publication status | Published - 26 Apr 2017 |
| Event | 5th International Conference on Learning Representations - Palais des Congrès Neptune, Toulon, France Duration: 24 Apr 2017 → 26 Apr 2017 https://iclr.cc/archive/www/2017.html |
Conference
| Conference | 5th International Conference on Learning Representations |
|---|---|
| Abbreviated title | ICLR 2017 |
| Country/Territory | France |
| City | Toulon |
| Period | 24/04/17 → 26/04/17 |
| Internet address |
Keywords / Materials (for Non-textual outputs)
- unsupervised learning
- structured prediction