I am currently a research fellow at Singapore University of Technology and Design working with Georgios Piliouras and Antonios Varvitsiotis.
My research develops theory and algorithms for learning in strategic, structured, and dynamic environments. I study how autonomous agents learn to act in multi-agent settings where structure, feedback loops and non-standard or non-stationary objectives shape long-term behavior. My work integrates tools from reinforcement learning, game theory, online learning, stochastic optimization, and dynamical systems to build principled and robust multi-agent learning systems. See below for a more detailed overview of my research.
Previously, I was a postdoctoral fellow at ETH Zurich working with Niao He in the department of computer science. I obtained my PhD in applied mathematics and computer science from Institut Polytechnique de Paris at Télécom Paris under the supervision of Pascal Bianchi and Walid Hachem. I received my engineering Master's degree in applied mathematics and computer science from Télécom Paris and a Master's degree in data science from Université Paris Saclay.
Here is my CV for more information.
PhD in Applied Mathematics and CS, 2021
Institut Polytechnique de Paris (Télécom Paris)
MSc in Data Science, 2018
Université Paris Saclay
MSc in Applied Mathematics and CS, 2018
Télécom Paris
Strategic Learning in Structured and Dynamic Multi-Agent Environments
with General Utility and Behaviorally Aligned Objectives
My research develops theory and algorithms for learning in strategic, structured and dynamic environments. I study how autonomous agents learn to act in multi-agent settings where feedback loops, structure, non-standard and non-stationary objectives shape long-term behavior. My work integrates tools from reinforcement learning, game theory, online learning, stochastic optimization, and dynamical systems to build principled and robust multi-agent learning systems.
Below is an overview of my main research directions, with corresponding publications and preprints.
- Multi-Agent Learning in Strategic, Structured and Dynamic Environments
I study learning in multi-agent environments where agents interact strategically—whether cooperatively or competitively. My work investigates decentralized learning dynamics and convergence to equilibria in structured settings. These include structured games such as (constrained) Markov potential games for multi-agent RL and zero-sum linear quadratic games, structured strategy spaces defined via generalized simplices in symmetric cones, and multi-agent linear dynamical systems with arbitrary disturbances and non-stationary time-varying objectives.- Beyond Expected Returns: General Utility and Behaviorally Aligned RL
I develop theory and algorithms for policy optimization beyond standard additive cumulative rewards, including non-linear and reward-free objectives. Drawing on insights from psychology and behavioral economics, I introduced a policy gradient framework based on cumulative prospect theory to model cognitive and psychological biases in sequential decision-making and capture the diversity of human attitudes toward risk. This work contributes to the design of human-centered RL systems that better align with complex preferences. Applications include personalized healthcare, where psychological factors influence clinical decisions, energy, finance, traffic management as well as domains such as legal and ethical decision-making, cybersecurity, and human-robot interaction.- A Dynamical Systems and Stochastic Approximation View of Adaptive Optimization
During my PhD, I studied the theoretical properties of adaptive gradient methods (e.g., Adam, Adagrad, RMSProp) using dynamical systems and stochastic approximation techniques, providing unified insights into their convergence and stability.
See below my list of publications organized along these complementary research directions.
Multi-Agent Learning in Strategic, Structured and Dynamic Environments
Expected Returns and Beyond: General Utility and Behaviorally Aligned RL
A Dynamical Systems and Stochastic Approximation View of Adaptive Optimization
Conferences: NeurIPS, ICML, ICLR, COLT, AISTATS, IEEE CDC.
Journals: Journal of Machine Learning Research (JMLR), Transactions on Machine Learning Research (TMLR), Mathematical Programming, SIAM Journal on Optimization (SIOPT), Journal of Optimization Theory and Applications (JOTA), Systems & Control Letters, Mathematics of Control, Signals, and Systems (MCSS), Stochastic Systems, IEEE Transactions on Automatic Control, IEEE Control Systems Letters.
ETH Zurich (2022-2024):
Télécom Paris (2018-2021): Teaching Assistant