Hybrid Learning Systems utilizing Sum Product Networks
We have recently witnessed a considerable interest in probabilistic models within deep learning, leading to e.g. generative adversarial networks, deep generative networks, neural auto-regressive density estimators and Pixel-RNNs/C...
ver más
¿Tienes un proyecto y buscas un partner? Gracias a nuestro motor inteligente podemos recomendarte los mejores socios y ponerte en contacto con ellos. Te lo explicamos en este video
Información proyecto HYBSPN
Duración del proyecto: 21 meses
Fecha Inicio: 2018-03-09
Fecha Fin: 2019-12-31
Fecha límite de participación
Sin fecha límite de participación.
Descripción del proyecto
We have recently witnessed a considerable interest in probabilistic models within deep learning, leading to e.g. generative adversarial networks, deep generative networks, neural auto-regressive density estimators and Pixel-RNNs/CNNs. Furthermore, sum-product networks (SPNs) are a recent deep architecture with a unique advantage over the aforementioned models: they allow both exact and efficient inference, implemented in terms of simple network passes. However, SPNs are a constrained type of neural network and do not reach the full flexibility of the deep learning tool kit available to date. This calls for hybrid learning systems which exploit the superior inference properties of SPNs within other deep learning approaches.
In this project, I will investigate two such approaches. First, I will structurally combine a deep learning architecture (front-end), which extracts a representation from a set of inputs, controlling the parameters of an SPN (back-end) over a set of outputs. This yields a hybrid conditional SPN which facilitates full inference over the output space, and which is naturally applied in structural prediction tasks. Such hybrid SPNs can be expected to be highly expressive and to set new state-of-the-art results in e.g. semantic image segmentation.
The second approach is to use SPNs as variational distributions, i.e. for approximating a given target distribution by minimizing Kullback-Leibler divergence. On the one hand, this allows to capture intractable models with SPNs, with the goal to enable fast amortized approximate inference. On the other hand, this approach allows to use hybrid conditional SPNs as so-called inference networks for intractable generative models with latent variables, for the purpose of variational posterior inference and learning. This approach would represent a substantial improvement over state-of-the-art approaches, which are usually limited to expensive inference via Monte Carlo estimation.