Title | ||
---|---|---|
Sum-Product Autoencoding: Encoding and Decoding Representations Using Sum-Product Networks. |
Abstract | ||
---|---|---|
Sum-Product Networks (SPNs) are a deep probabilistic architecture that up to now has been successfully employed for tractable inference. Here, we extend their scope towards unsupervised representation learning: we encode samples into continuous and categorical embeddings and show that they can also be decoded back into the original input space by leveraging MPE inference. We characterize when this Sum-Product Autoencoding (SPAE) leads to equivalent reconstructions and extend it towards dealing with missing embedding information. Our experimental results on several multi-label classification problems demonstrate that SPAE is competitive with state-of-the-art autoencoder architectures, even if the SPNs were never trained to reconstruct their inputs. |
Year | Venue | Field |
---|---|---|
2018 | THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | Computer science,Arithmetic,Artificial intelligence,Decoding methods,Machine learning,Encoding (memory) |
DocType | Citations | PageRank |
Conference | 4 | 0.39 |
References | Authors | |
0 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Antonio Vergari | 1 | 31 | 11.86 |
Robert Peharz | 2 | 86 | 12.30 |
Nicola Di Mauro | 3 | 380 | 47.66 |
Alejandro Molina | 4 | 46 | 15.04 |
Kristian Kersting | 5 | 1932 | 154.03 |
Floriana Esposito | 6 | 2434 | 277.96 |