Anas Barakat

Anas Barakat

Research Fellow

Singapore University of Technology and Design

I am currently a research fellow at Singapore University of Technology and Design working with Georgios Piliouras and Antonios Varvitsiotis.

My research develops theory and algorithms for learning in strategic, structured, and dynamic environments. I study how autonomous agents learn to act in multi-agent settings where structure, feedback loops and non-standard or non-stationary objectives shape long-term behavior. My work integrates tools from reinforcement learning, game theory, online learning, stochastic optimization, and dynamical systems to build principled and robust multi-agent learning systems. See below for a more detailed overview of my research.

Previously, I was a postdoctoral fellow at ETH Zurich working with Niao He in the department of computer science. I obtained my PhD in applied mathematics and computer science from Institut Polytechnique de Paris at Télécom Paris under the supervision of Pascal Bianchi and Walid Hachem. I received my engineering Master's degree in applied mathematics and computer science from Télécom Paris and a Master's degree in data science from Université Paris Saclay.

Here is my CV for more information.

Interests

  • Multi-Agent Learning
  • Reinforcement Learning
  • Stochastic Optimization

Education

  • PhD in Applied Mathematics and CS, 2021

    Institut Polytechnique de Paris (Télécom Paris)

  • MSc in Data Science, 2018

    Université Paris Saclay

  • MSc in Applied Mathematics and CS, 2018

    Télécom Paris

Research

Strategic Learning in Structured and Dynamic Multi-Agent Environments
with General Utility and Behaviorally Aligned Objectives

My research develops theory and algorithms for learning in strategic, structured and dynamic environments. I study how autonomous agents learn to act in multi-agent settings where feedback loops, structure, non-standard and non-stationary objectives shape long-term behavior. My work integrates tools from reinforcement learning, game theory, online learning, stochastic optimization, and dynamical systems to build principled and robust multi-agent learning systems.

Below is an overview of my main research directions, with corresponding publications and preprints.

- Multi-Agent Learning in Strategic, Structured and Dynamic Environments

I study learning in multi-agent environments where agents interact strategically—whether cooperatively or competitively. My work investigates decentralized learning dynamics and convergence to equilibria in structured settings. These include structured games such as (constrained) Markov potential games for multi-agent RL and zero-sum linear quadratic games, structured strategy spaces defined via generalized simplices in symmetric cones, and multi-agent linear dynamical systems with arbitrary disturbances and non-stationary time-varying objectives.

- Beyond Expected Returns: General Utility and Behaviorally Aligned RL

I develop theory and algorithms for policy optimization beyond standard additive cumulative rewards, including non-linear and reward-free objectives. Drawing on insights from psychology and behavioral economics, I introduced a policy gradient framework based on cumulative prospect theory to model cognitive and psychological biases in sequential decision-making and capture the diversity of human attitudes toward risk. This work contributes to the design of human-centered RL systems that better align with complex preferences. Applications include personalized healthcare, where psychological factors influence clinical decisions, energy, finance, traffic management as well as domains such as legal and ethical decision-making, cybersecurity, and human-robot interaction.

- A Dynamical Systems and Stochastic Approximation View of Adaptive Optimization

During my PhD, I studied the theoretical properties of adaptive gradient methods (e.g., Adam, Adagrad, RMSProp) using dynamical systems and stochastic approximation techniques, providing unified insights into their convergence and stability.

See below my list of publications organized along these complementary research directions.

Talks

Invited talk - Learning Theory and Applications Workshop, Nanyang Technological University
Invited talk - Finance and RL Talks
Invited talk - 4th Symposium on Machine Learning and Dynamical Systems
Talk - SUTD Group Meeting
Invited talk - ICCOPT 2022

Reviewing

  • Conferences: NeurIPS, ICML, ICLR, COLT, AISTATS, IEEE CDC.

  • Journals: Journal of Machine Learning Research (JMLR), Transactions on Machine Learning Research (TMLR), Mathematical Programming, SIAM Journal on Optimization (SIOPT), Journal of Optimization Theory and Applications (JOTA), Systems & Control Letters, Mathematics of Control, Signals, and Systems (MCSS), Stochastic Systems, IEEE Transactions on Automatic Control, IEEE Control Systems Letters.

Teaching

ETH Zurich (2022-2024):

Télécom Paris (2018-2021): Teaching Assistant