What excites me is uncovering the basic principles behind intelligent systems acting in the world. How can a system learn to perceive the relevant things out there, and act according to some internal goals? If you open the gory black box of an intelligent system, which mechanisms will you find that lie at the basis of all these complex perceptions and actions? I focus on the artificial side of intelligence, while drawing inspiration from natural intelligence. Lately I'm intrigued by the power of self-supervised sequence learning, and all the (un)expected capabilities arising from it, such as in-context learning, hidden world models, action hierarchies etc. Using insights from optimal control, theoretical neuroscience and optimization, I'm trying to make sense of it all.
More concretely, my research focuses on (i) encovering hidden world-models in transformer models, (ii) making reinforcement learning more sample efficient by introducing model-based credit assignment techniques for learning policies grounded in causality and (iii) unifying action, perception and learning by bridging optimal control with theories of learning in the brain, leading to novel learning algorithms for deep and recurrent neural networks that alleviate important drawbacks of current neural network training methods.
Besides my love for science, I'm a passionate musician and on the weekends, you can find me either skiing or hiking in the beautiful Swiss Alps, or experimenting with homemade pizza.