News

Jul 08, 2025 Our paper, EuroBERT: Scaling Multilingual Encoders for European Languages is accepted to COLM 2025.
Jul 02, 2025 We release Should We Still Pretrain Encoders with Masked Language Modeling?, a large-scale study comparing causal and bidirectional pretraining objectives for text representation learning.
Jun 23, 2025 EuroBERT is nominated for the 2025 Datacraft Awards in the AI and Society category.
May 26, 2025 We are invited to give a talk about EuroBERT at Meta AI Research.
May 13, 2025 We are invited to give a talk about EuroBERT at FOR.
Apr 30, 2025 We are invited to give a talk about EuroBERT at Cohere.
Mar 10, 2025 We release EuroBERT, a fully open-source family of encoder models specialized in European languages.
Jan 22, 2025 MMTEB: Massive Multilingual Text Embedding Benchmark is accepted to ICLR 2025.
Nov 15, 2024 Presented our work, Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis, as both a poster and oral at WMT 2024, held in conjunction with EMNLP 2024.
Sep 24, 2024 Our paper, Is Preference Alignment Always the Best Option to Enhance LLM-Based Translation? An Empirical Analysis, gets accepted to WMT 2024 as a featured paper (second best overall).
Sep 02, 2024 Our paper, Towards Trustworthy Reranking: A Simple yet Effective Abstention Mechanism, is accepted for publication in TMLR to appear in the October 2024 issue.
Feb 20, 2024 We release Towards Trustworthy Reranking: A Simple yet Effective Abstention Mechanism, a paper proposing a lightweight abstention mechanism for information retrieval, designed to fit insustrial needs.
Nov 06, 2023 I begin my CIFRE PhD in collaboration with CentraleSupélec, Université Paris-Saclay, and the Artefact Research Center, under the supervision of Pierre Colombo, Céline Hudelot, and Emmanuel Malherbe.