Research Topic

Design of Learning Systems

This project aims at identifying means to reduce the search space in learning systems as one way to improve the corresponding learning processes. To this end, we study the geometric properties of various connectionistic models known within the field of machine learning using information geometry and algebraic statistics. Our goal is to find distinguished architectures of learning systems based on their expressive power and learning performance.
This kind of model selection is motivated by experimental and theoretical work on restricted Boltzmann machines and deep belief networks, popular learning systems which evermore demand a profound mathematical investigation. This project targets especially the development of design principles for our embodied AI project.


Selection Criteria for Neuromanifolds of Stochastic Dynamics

Within many formal models of neuronal systems, individual neurons are modelled as nodes which receive inputs from other nodes in a network and generate an output that can be stochastic in general. This way the dynamics of the whole network can be described as a stochastic transition in each time step, mathematically formalized in terms of a stochastic matrix. Well-known models of this kind are Boltzmann machines, their generalizations, and policy matrices within reinforcement learning. In order to study such learning systems it is helpful to consider not only one stochastic matrix but a parametrized family of matrices, which forms a geometric object, referred to as neuromanifold within information geometry. Learning crucially depends on the shape of the neuromanifold. This information geometric view, which has been proposed by Amari, suggests to select appropriate neuromanifolds and to define corresponding learning processes as gradient flows on these manifolds. We do not only focus on manifolds that are directly induced by a neuronal model, but study general sets that satisfy natural optimality conditions.

Two dimensional sets containing all deterministic policies

Deterministic policies or near to deterministic policies are optimal for a variety of reinforcement learning problems, they represent dynamics with maximal predictive information as considered in robotics and also dynamics of neural networks with maximal network information flow. It is always possible to construct a two dimensional set that reaches all deterministic policies and on which natural gradient optimization works very efficiently.

inBook
2020 Repository Open Access
Johannes Müller and Marius Zeinhofer

Deep Ritz revisited

In: ICLR 2020 workshop on integration of deep neural models and differential equations : Millennium Hall, Addis Ababa, Ethiopia ; 26th April 2020
[S. L.] : ICLR, 2020.
inJournal
2022 Repository Open Access
Csongor Várady, Riccardo Volpi, Luigi Malagò and Nihat Ay

Natural reweighted wake-sleep

In: Neural networks, 155 (2022), pp. 574-591
MiS Preprint
2020 Repository Open Access
Csongor Varady, Nihat Ay, Riccardo Volpi and Luigi Malagò

Natural Wake-Sleep Algorithm

inJournal
2023 Journal Open Access
Nihat Ay

On the locality of the natural gradient for learning in deep Bayesian networks

In: Information geometry, 6 (2023) 1, pp. 1-49
inBook
2020 Repository Open Access
Johannes Müller

On the space-time expressivity of ResNets

In: ICLR 2020 workshop on integration of deep neural models and differential equations : Millennium Hall, Addis Ababa, Ethiopia ; 26th April 2020
[S. L.] : ICLR, 2020.
inBook
2019 Repository Open Access
Nihat Ay, Johannes Rauh and Guido Montúfar

A continuity result for optimal memoryless planning in POMDPs

In: RLDM 2019 : 4th multidisciplinary conference on reinforcement learning and decision making ; July 7-10, 2019 ; Montréal, Canada
Montréal, Canada : University, 2019. - pp. 362-365
inBook
2019 Repository Open Access
Guido Montúfar, Johannes Rauh and Nihat Ay

Task-agnostic constraining in average reward POMDPs

In: Task-agnostic reinforcement learning : workshop at ICLR, 06 May 2019, New Orleans
[S. L.] : ICLR, 2019.
Preprint
2018 Repository Open Access
Guido Montúfar

Illustration of maxout layer upper bound [Suppl. to: On the number of linear regions of deep neural networks]

inBook
2018 Repository Open Access
Guido Montúfar, Johannes Rauh and Nihat Ay

Uncertainty and stochasticity of optimal policies

In: Proceedings of the 11th workshop on uncertainty processing WUPES '18, June 6-9, 2018 / Václav Kratochvíl (ed.)
Praha : MatfyzPress, 2018. - pp. 133-140
inJournal
2017 Journal Open Access
Guido Montúfar and Jason Morton

Dimension of marginals of Kronecker product models

In: SIAM journal on applied algebra and geometry, 1 (2017) 1, pp. 126-151
inBook
2017 Repository Open Access
Guido Montúfar and Johannes Rauh

Geometry of policy improvement

In: Geometric science of information : Third International Conference, GSI 2017, Paris, France, November 7-9, 2017, proceedings / Frank Nielsen... (eds.)
Cham : Springer, 2017. - pp. 282-290
(Lecture notes in computer science ; 10589)
inBook
2015 Repository Open Access
Guido Montúfar and Johannes Rauh

Hierarchical models as marginals of hierarchical models

In: Proceedings of the 10th workshop on uncertainty processing WUPES '15, Moninec, Czech Republic, September 16-19, 2015 / Václav Kratochvíl (ed.)
Praha : Oeconomica, 2015. - pp. 131-145
inBook
2017 Repository Open Access
Guido Montúfar

Notes on the number of linear regions of deep neural networks

In: 2017 international conference on sampling theory and applications (SampTA) / Gholamreza Anbarjafari... (eds.)
Piscataway, NJ : IEEE, 2017. - pp. 156-159
inJournal
2017 Repository Open Access
Guido Montúfar, Jason Morton and Johannes Rauh

Restricted Boltzmann machines [In: Algebraic statistics ; 16 April - 22 April 2017 ; report no. 20/2017]

In: Oberwolfach reports, 14 (2017) 2, pp. 1241-1242
inBook
2015 Repository Open Access
Guido Montúfar and Johannes Rauh

Mode poset probability polytopes

In: Proceedings of the 10th workshop on uncertainty processing WUPES '15, Moninec, Czech Republic, September 16-19, 2015 / Václav Kratochvíl (ed.)
Praha : Oeconomica, 2015. - pp. 147-154
inBook
2015 Repository Open Access
Guido Montúfar

A comparison of neural network architectures

In: Deep learning Workshop, ICML '15, Vauban Hall at Lille Grande Palais, France, July 10 and 11, 2015
2015.
inJournal
2015 Journal Open Access
Guido Montúfar, Keyan Ghazi-Zahedi and Nihat Ay

A theory of cheap control in embodied systems

In: PLoS computational biology, 11 (2015) 9, e1004427
inBook
2015 Repository Open Access
Guido Montúfar

Deep narrow Boltzmann machines are universal approximators

In: Third international conference on learning representations - ICLR 2015 : May 7-9 2015, San Diego, CA. USA
San Diego : ICLR, 2015.
inJournal
2015 Journal Open Access
Guido Montúfar and Jason Morton

Discrete restricted Boltzmann machines

In: Journal of machine learning research, 16 (2015), pp. 653-672
inJournal
2015 Journal Open Access
Nihat Ay

Geometric design principles for brains of embodied agents

In: Künstliche Intelligenz : KI, 29 (2015) 4, pp. 389-399
Preprint
2015 Repository Open Access
Guido Montúfar, Keyan Ghazi-Zahedi and Nihat Ay

Geometry and determinism of optimal stationary control in partially observable Markov decision processes

inJournal
2015 Journal Open Access
Guido Montúfar, Nihat Ay and Keyan Ghazi-Zahedi

Geometry and expressive power of conditional restricted Boltzmann machines

In: Journal of machine learning research, 16 (2015), pp. 2405-2436
inJournal
2017 Repository Open Access
Guido Montúfar and Johannes Rauh

Hierarchical models as marginals of hierarchical models

In: International journal of approximate reasoning, 88 (2017), pp. 531-546
Preprint
2015 Repository Open Access
Guido Montúfar

Universal approximation of Markov kernels by shallow stochastic feedforward networks

inJournal
2015 Repository Open Access
Guido Montúfar and Jason Morton

When does a mixture of products contain a product of mixtures?

In: SIAM journal on discrete mathematics, 29 (2015) 1, pp. 321-347
inBook
2014
Guido Montúfar and Jason Morton

Geometry of hidden-visible products of statistical models

In: Algebraic Statistics 2014 : May 19-22
Chicago, IL : Illinois Institute of Technology, 2014.
inJournal
2014 Journal Open Access
Guido Montúfar, Johannes Rauh and Nihat Ay

On the Fisher metric of conditional probability polytopes

In: Entropy, 16 (2014) 6, pp. 3207-3233
inBook
2014 Repository Open Access
Razvan Pascanu, Guido Montúfar and Yoshua Bengio

On the number of inference regions of deep feed forward networks with piece-wise linear activations

In: Second international conference on learning representations - ICLR 2014 : 14-16 April 2014, Banff, Canada
Banff : ICLR, 2014.
inBook
2014 Repository Open Access
Guido Montúfar, Razvan Pascanu, Kyunghyun Cho and Yoshua Bengio

On the number of linear regions of deep neural networks

In: NIPS 2014 : Proceedings of the 27th international conference on neural information processing systems - volume 2 ; Montreal, Quebec, Canada, December 8th-13th
Cambridge, MA : MIT Press, 2014. - pp. 2924-2932
inJournal
2014 Repository Open Access
Johannes Rauh and Nihat Ay

Robustness, canalyzing functions and systems design

In: Theory in biosciences, 133 (2014) 2, pp. 63-78
inJournal
2014 Journal Open Access
Guido Montúfar and Johannes Rauh

Scaling of model approximation errors and expected entropy distances

In: Kybernetika, 50 (2014) 2, pp. 234-245
inJournal
2014 Repository Open Access
Guido Montúfar

Universal approximation depth and errors of narrow belief networks with discrete units

In: Neural computation, 26 (2014) 7, pp. 1386-1407
inBook
2013 Repository Open Access
Guido Montúfar, Johannes Rauh and Nihat Ay

Maximal information divergence from statistical models defined by neural networks

In: Geometric science of information : first international conference, GSI 2013, Paris, France, August 28-30, 2013. Proceedings / Frank Nielsen... (eds.)
Berlin [u. a.] : Springer, 2013. - pp. 759-766
(Lecture notes in computer science ; 8085)
inJournal
2013 Journal Open Access
Guido Montúfar

Mixture decompositions of exponential families using a decomposition of their sample spaces

In: Kybernetika, 49 (2013) 1, pp. 23-39
inBook
2013 Repository Open Access
Nihat Ay, Guido Montúfar and Johannes Rauh

Selection criteria for neuromanifolds of stochastic dynamics

In: Advances in cognitive neurodynamics III : proceedings of the 3rd International Conference on Cognitive Neurodynamics 2011 ; [June 9-13, 2011, Hilton Niseko Village, Hokkaido, Japan] / Yoko Yamaguchi (ed.)
Dordrecht : Springer, 2013. - pp. 147-154
(Advances in cognitive neurodynamics)
Academic
2012
Guido Montúfar

On the expressive power of discrete mixture models, restricted Boltzmann machines, and deep belief networks - a unified mathematical treatment

Dissertation, Universität Leipzig, 2012
inBook
2012 Repository Open Access
Guido Montúfar and Johannes Rauh

Scaling of model approximation errors and expected entropy distances

In: Proceedings of the 9th workshop on uncertainty processing WUPES '12 : Marianske Lazne, Czech Republik ; 12-15th September 2012
Praha : Academy of Sciences of the Czech Republik / Institute of Information Theory and Automation, 2012. - pp. 137-148
inBook
2011 Repository Open Access
Guido Montúfar, Johannes Rauh and Nihat Ay

Expressive power and approximation errors of restricted Boltzmann machines

In: Advances in neural information processing systems 24 : NIPS 2011 ; 25th annual conference on neural information processing systems 2011, Granada, Spain December 12th - 15th / John Shawe-Taylor (ed.)
La Jolla, CA : Neural Information Processing Systems, 2011. - pp. 415-423
inJournal
2011 Repository Open Access
Guido Montúfar and Nihat Ay

Refinements of universal approximation results for deep belief networks and restricted Boltzmann machines

In: Neural computation, 23 (2011) 5, pp. 1306-1319
inBook
2010 Repository Open Access
Guido Montúfar

Mixture models and representational power of RBM's, DBN's, and DBM's

In: NIPS 2010 : Deep learning and unsupervised feature learning workshop ; December 19, 2010, Hilton, Vancouver, Canada
[s. l.] : NIPS, 2010. - pp. 1-9
Academic
2010 Repository Open Access
Thomas Kahle

On boundaries of statistical models

Dissertation, Universität Leipzig, 2010
inJournal
2006 Journal Open Access
Nihat Ay and Andreas Knauf

Maximizing multi-information

In: Kybernetika, 42 (2006) 5, pp. 517-538
Academic
2001
Nihat Ay

Aspekte einer Theorie pragmatischer Informationsstrukturierung

Dissertation, Universität Leipzig, 2001