Jérôme Garnier-Brun

Marie Skłodowska-Curie Postdoctoral Researcher
Department of Computing Sciences
Università Bocconi, Milan, Italy

Jérôme Garnier-Brun

About me

Hello!

My name is Jérôme Garnier-Brun and I am currently a Marie Skłodowska-Curie postdoctoral research fellow at Università Bocconi in Milan, Italy. I am working with Marc Mézard as part of the Machine Learning & Statistical Physics group of the Department of Computing Sciences on questions relating to the role of structure in data for machine learning. I am also affiliated to the Department of Finance, where I collaborate with Claudio Tebaldi on the application of statistical physics to economics and finance.

Before that, I did my PhD between École polytechnique and Capital Fund Management, within the chair of Econophysics and Complex Systems under the supervision of Jean-Philippe Bouchaud and Michael Benzaquen. There, I mainly worked on the problems at the interface between the physics of spin-glasses and disordered socioeconomic systems.

Outside of research, I enjoy hiking, trail running and cycling in the summer, and skiing and ski mountaineering in the winter. I am also a poor but enthusiastic footballer and a (self-proclaimed) decent cook.

Recent news

  • April 2024: I am happy to announce I have been awarded a Marie Skłodowska-Curie fellowship for my project SPIRAL (Structured Physics-Inspired Representations and dAta models for efficient Learning), extending my stay at Bocconi under the supervision of Marc Mézard!
  • December 2024: New preprint with Ruben Zakine and Michael Benzaquen on a nonequilibrium socioeconomic agent-based model is out!
  • September 2024: I presented our work on how transformers learn structured data at the Rockin'AI summer school in Roccella Ionica, Calabria.
  • August 2024: New preprint with Marc Mézard, Emanuele Moscato and Luca Saglietti on how transformers learn structured data is out!

Contact

Email: jerome DOT garnier AT unibocconi DOT it

The role of the structure of data in machine learning

In the last few years, the success of deep learning, and in particular Large Language Models (LLMs), has been largely driven by the availability of large datasets and the development of powerful computing hardware. However, the role of the structure of data in machine learning is still not well understood. In particular, the question of how to exploit the structure of data to improve the performance of machine learning algorithms remains largely open.

We tackle this question by leveraging simple, tree-based models of data. In this context, we have notably understood how transformers sequentially learn the underlying structure of the data model, and seem to implement the optimal inference algorithm.

Attention matrices visualization

Visualization of attention patterns in a four layer transformer trained on tree-based hierarchical data.

Generative diffusion

In addition to LLMs for text, the generative AI revolution has also been enabled by so-called diffusion models for image creation. By training complex architectures to denoise images, one can indeed produce seemingly novel visuals by following a reverse diffusion process starting from pure noise. Despite their effectiveness and widespread adoption, the role of memorized training data in what these models produce is still unclear, posing important challenges from the perspective of privacy and intellectual property. More generally, the way meaningful new samples are being sculpted along the reverse diffusion process remains poorly understood.

We investigate these issues using the tools of statistical physics, and explore the role the structure of data may play there.

Disordered socioeconomic toy models

Nonequilibrium field theories: from active matter to spatial agent-based models

Preprints

Garnier-Brun J., Zakine R. & Benzaquen M. (2024).
From nonequilibrium to equilibrium: insights from a two-population occupation model.
arXiv preprint, arXiv:2412.14996 [arXiv]
Garnier-Brun J., Mézard M., Moscato E. & Saglietti L. (2024).
How transformers learn structured data: insights from hierarchical filtering.
arXiv preprint, arXiv:2408.15138 [arXiv]

Journal Publications

Garnier-Brun J., Benzaquen M. & Bouchaud J.-P. (2024).
Unlearnable games and 'satisficing' decisions: a simple model for a complex world.
Phys. Rev. X 14, 021039 [OA journal]
Featured in New Scientist [link]
Zakine R., Garnier-Brun J., Becharat A.-C. & Benzaquen M. (2024).
Socioeconomic agents as active matter in nonequilibrium Sakoda-Schelling models.
Phys. Rev. E 109, 044310 [journal, arXiv]
Dechant A., Garnier-Brun J. & Sasa S.-I. (2023).
Thermodynamic bounds on correlation times.
Phys. Rev. Lett. 131, 167101 (Editor's Suggestion) [journal, arXiv]
Garnier-Brun J., Bouchaud J.-P. & Benzaquen M. (2023).
Bounded rationality and animal spirits: a fluctuation-response approach to Slutsky matrices.
J. Phys. Complex. 4, 015004 [OA journal]
Alemany I., Rose J. N., Garnier-Brun J., Scott A. D., & Doorly D. J. (2022).
Random walk diffusion simulations in semi-permeable layered media with varying diffusivity.
Sci. Rep. 12, 10759 [OA journal]
Garnier-Brun J., Benzaquen M., Ciliberti S., & Bouchaud J.-P. (2021).
A new spin on optimal portfolios and ecological equilibria.
J. Stat. Mech. 093408 (featured in Highlights) [OA journal]