Anas Barakat

Anas Barakat

Research Fellow

Singapore University of Technology and Design

I am currently a Research Fellow at Singapore University of Technology and Design, working with Georgios Piliouras and Antonios Varvitsiotis.

My primary research interests are in reinforcement learning and multi-agent learning. I develop theory and algorithms for sequential decision-making and learning under uncertainty. I study how learning dynamics evolve in structured, non-stationary, and strategic environments, and I design algorithms with provable guarantees on their long-term behavior. More recently, I have been exploring how these ideas can inform post-training and alignment in large language models. My work draws on tools from game theory, online learning, stochastic optimization and dynamical systems.

Previously, I was a postdoctoral fellow at ETH Zurich working with Niao He in the department of computer science. I obtained my PhD in applied mathematics and computer science from Institut Polytechnique de Paris at Télécom Paris under the supervision of Pascal Bianchi and Walid Hachem. I received my engineering Master's degree in applied mathematics and computer science from Télécom Paris and a Master's degree in data science from Université Paris Saclay.

Here is my CV for more information.

News

  • 02/2026: Check out our new work: `Why Pass@k Optimization Can Degrade Pass@1: Prompt Interference in LLM Post-training’, arxiv link.
  • 02/2026: ‘Optimistic Online Learning in Symmetric Cone Games’ accepted to Transactions on Machine Learning Research, link.
  • 01/2026: ‘Convex Markov Games and Beyond: New Proof of Existence, Characterization and Learning Algorithms for Nash Equilibria’ accepted to AISTATS 2026, arxiv link.
  • 10/2025: Designing and teaching a new course with John Lazarsfeld and Iosif Sakos. Check out our website: Online Learning and Learning in Games.
  • 09/2025: `On the Global Optimality of Policy Gradient Methods in General Utility Reinforcement Learning’ accepted to NeurIPS 2025.

Interests

  • Multi-Agent Learning
  • Reinforcement Learning
  • Stochastic Optimization

Education

  • PhD in Applied Mathematics and CS, 2021

    Institut Polytechnique de Paris (Télécom Paris)

  • MSc in Data Science, 2018

    Université Paris Saclay

  • MSc in Applied Mathematics and CS, 2018

    Télécom Paris

Talks

Invited talk - 5th Symposium on Machine Learning and Dynamical Systems
Invited talk - ICCOPT 2025
Invited talk - Learning Theory and Applications Workshop, NTU
Invited talk - Finance and RL Talks
Invited talk - 4th Symposium on Machine Learning and Dynamical Systems

Reviewing

  • Conferences: NeurIPS, ICML, ICLR, AISTATS, EC.

  • Journals: Journal of Machine Learning Research (JMLR), Transactions on Machine Learning Research (TMLR), Mathematical Programming, SIAM Journal on Optimization (SIOPT), Journal of Optimization Theory and Applications (JOTA), IEEE Transactions on Automatic Control.

Teaching

SUTD (2024-2025):

ETH Zurich (2022-2024):

Télécom Paris (2018-2021): Teaching Assistant

Optimization for Machine Learning (graduate level), Statistics (graduate level), Probabilities (undergraduate level).