Matéo Mahaut

Profile Picture

I'm Matéo, a final year PhD student in the COLT group at Universitat Pompeu Fabra under the supervision of Marco Baroni. I'm interested in multi-agent cooperation, emergent communication, and interpretability. Before that, I got my engineering diploma at the École Nationale Supérieure de Cognitique in Bordeaux, and interned at the FLOWERS Inria lab, and the Institut de Neurosciences de la Timone in Marseille.

News

Teaching

Reinforcement Learning - Madagascar Machine Learning Summer School

December 2023 & December 2022

I gave an introductory course on Reinforcement Learning at the Madagascar Machine Learning Summer School, covering the basics of RL, key algorithms, and applications to LLMs.

Computational Semantics - Universitat Pompeu Fabra

Fall 2024

As a teaching assistant, I helped students understand the fundamentals of machine learning, supervised lab sessions, and provided support for course projects.

Projects

LLM interpretability - mecanisms of factual memorisation

Ongoing work, feel free to contact me with questions and ideas!

We analyze different ways facts are stored and accessed by LLMs. A first work ( published at ACL) compared methods estimating model confidence in a given fact. Current extensions explore the memorisation process. We're for example interested in how representation of a sentence will vary depending on fine-tuning.

Referential communication in pre-trained populations

Project Image

This project looked at whether very different state of the art foundation models could build a common language for referential communication. Our results show that a common representation can very rapidly emerge, and that those protocols can generalise in a variety of ways (Accepted at TMLR). Follow up works include understanding how models build such similar representations throughout the different layers.

Repetitions in Language Models

In this project, we investigate the mechanisms behind repetition phenomena in large language models. Our findings show that not all repetitions are generated by the same underlying processes—distinct mechanisms can sustain repetition depending on the context and model architecture. Ongoing works include looking at the apparition of this phenomenon in the training process, and the mapping of different circuits to different functions.
Read the paper: "Repetitions are not all alike: distinct mechanisms sustain repetition in language models"

Publications

Repetitions are not all alike: distinct mechanisms sustain repetition in language models

Mahaut, M., & Franzon, F. (2025). Repetitions are not all alike: distinct mechanisms sustain repetition in language models. arXiv preprint arXiv:2504.01100.

ArXiv

Factual Confidence of LLMs: on Reliability and Robustness of Current Estimators

Aina, L., Czarnowska, P., Hardalov, M., Müller, T., Màrquez, L. (2024). Factual Confidence of LLMs: on Reliability and Robustness of Current Estimators. ACL 2024, Bangkok Thaïland.

ArXiv GitHub Poster

Referential communication in heterogeneous communities of pre-trained visual deep networks

Mahaut, M., Franzon, F., Dessì, R., & Baroni, M. (2023). Referential communication in heterogeneous communities of pre-trained visual deep networks. arXiv preprint arXiv:2302.08913.

ArXiv GitHub Poster

Social network structure shapes innovation: experience-sharing in RL with SAPIENS

Nisioti, E., Mahaut, M., Oudeyer, P. Y., Momennejad, I., & Moulin-Frier, C. (2022). Social network structure shapes innovation: experience-sharing in RL with SAPIENS. arXiv preprint arXiv:2206.05060.

ArXiv

Team performance analysis of a collaborative spatial orientation mission in Mars analogue environment

Prebot, B., Cavel, C., Calice, L., Mahaut, M., Leduque, A., & Salotti, J. M. (2019, October). Team performance analysis of a collaborative spatial orientation mission in Mars analogue environment. In 70th International Astronautical Congress (p. 7).

Researchgate