Encoding and decoding representations with sum- And max-product networks

Antonio Vergari, Robert Peharz, Nicola Di Mauro, Floriana Esposito

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Sum-Product networks (SPNs) are expressive deep architectures for representing probability distributions, yet allowing exact and efficient inference. SPNs have been successfully applied in several domains, however always as black-box distribution estimators. In this paper, we argue that due to their recursive definition, SPNs can also be naturally employed as hierarchical feature extractors and thus for unsupervised representation learning. Moreover, when converted into Max-Product Networks (MPNs), it is possible to decode such representations back into the original input space. In this way, MPNs can be interpreted as a kind of generative autoencoder, even if they were never trained to reconstruct the input data. We show how these learned representations, if visualized, indeed correspond to "meaningful parts" of the training data. They also yield a large improvement when used in structured prediction tasks. As shown in extensive experiments, SPN and MPN encoding and decoding schemes prove very competitive against the ones employing RBMs and other stacked autoencoder architectures.
Original languageEnglish
Title of host publicationInternational Conference on Learning Representations (ICLR 2017) - Workshop Track
Number of pages10
Publication statusPublished - 26 Apr 2017
Event5th International Conference on Learning Representations - Palais des Congrès Neptune, Toulon, France
Duration: 24 Apr 201726 Apr 2017
https://iclr.cc/archive/www/2017.html

Conference

Conference5th International Conference on Learning Representations
Abbreviated titleICLR 2017
Country/TerritoryFrance
CityToulon
Period24/04/1726/04/17
Internet address

Keywords / Materials (for Non-textual outputs)

  • unsupervised learning
  • structured prediction

Fingerprint

Dive into the research topics of 'Encoding and decoding representations with sum- And max-product networks'. Together they form a unique fingerprint.

Cite this