Edouard Grave

Edouard Grave is a French AI researcher, Chief Language Officer at Kyutai, and a co-author on the fastText library, the original LLaMA paper, and Moshi.
Edouard Grave

Edouard Grave

Edouard Grave is a French computer scientist, born June 28, 1986. He is the Chief Language Officer of Kyutai, the Paris-based nonprofit AI research lab founded in November 2023, where he leads work on large language models, and a co-author on the fastText library, the original LLaMA paper at Meta AI FAIR Paris, and the Moshi speech-text foundation model. As of May 2026, he is one of the senior researchers shaping the open-research direction at Kyutai, alongside ex-FAIR Paris colleagues Hervé Jegou and Patrick Pérez.

At a glance

Origins

Grave was born on June 28, 1986 in France. He entered École Polytechnique in 2006 as a member of the X2006 promotion. His admission and the 2009 engineer's-degree graduation are recorded in the Légifrance decrees that publish each Polytechnique class roster. The Polytechnique years overlapped with the early 2010s expansion of machine-learning research in Paris that produced much of the cohort he would later work with at FAIR and at Kyutai.

After Polytechnique, Grave moved into doctoral research at the boundary of statistics and machine learning, the path that took him through Inria, three postdoctoral positions, and into industrial AI research.

Career

Grave began his PhD in 2010 at Université Pierre et Marie Curie (Paris VI), the science campus of what became Sorbonne Université, with research carried out at Inria under Francis Bach and Guillaume Obozinski. The thesis, "A Markovian approach to distributional semantics," was defended in 2014 and developed probabilistic models for learning word and phrase representations from large unlabeled corpora using hidden-Markov-tree structure over syntactic dependency parses. The thesis line produced a 2011 NIPS paper on the Trace Lasso regularizer with Bach and Obozinski and a 2013 CoNLL paper on Hidden Markov tree models for semantic class induction with the same advisors.

After defending in 2014, Grave held three postdoctoral positions in succession. He spent 2014 at the University of California, Berkeley on the STATWEB Inria-Berkeley team, working between Berkeley and Inria's SIERRA group. In 2015 he was a postdoctoral researcher at Columbia University in New York. In 2016 he joined Facebook AI Research (FAIR) as a postdoctoral researcher in the Paris and New York offices. The period produced the Improving Neural Language Models with a Continuous Cache paper at ICLR 2017, the first of the language-modeling artifacts that defined the next several years of his record.

Grave converted from postdoctoral researcher to Research Scientist at FAIR in 2017 and remained at the lab through early 2023. The FAIR period produced three of his most-cited research lines. The first was the fastText word-embeddings work with Armand Joulin, Piotr Bojanowski, and Tomas Mikolov. The two flagship papers, Enriching Word Vectors with Subword Information (TACL 2017) and Bag of Tricks for Efficient Text Classification (EACL 2017), introduced subword-aware embeddings and the fastText classifier and reached over 22,000 combined Google Scholar citations by mid-2026. The associated GitHub library became one of the most widely used pieces of open-source NLP infrastructure, with multilingual word vectors for 157 languages released in Grave et al. 2018 at LREC.

The second line was retrieval-augmented language modeling, with the 2022 Atlas paper co-authored with Gautier Izacard and others showing that an 11-billion-parameter retrieval-augmented model could outperform a 540-billion-parameter dense model on Natural Questions with only 64 training examples. The line built on earlier Grave-Izacard collaborations on Fusion-in-Decoder retrieval architectures.

The third line was the LLaMA paper, submitted to arXiv on February 27, 2023 under the title "LLaMA: Open and Efficient Foundation Language Models". The 14-author byline lists Grave alongside Hugo Touvron as first author, Timothée Lacroix, Marie-Anne Lachaux, Baptiste Rozière, and others, with Guillaume Lample as the last-listed senior author. The paper introduced the foundation-model family in 7B, 13B, 33B, and 65B parameter sizes and triggered the wave of open-weights frontier-class language models that followed through 2023 and 2024.

Grave left Meta in 2023 in the larger FAIR Paris exodus that also produced Mistral AI. He spent a brief period at Apple in 2023 before joining Kyutai at its November 2023 launch as one of the lab's six founding researchers, alongside Pérez, Hervé Jegou, Laurent Mazaré, Neil Zeghidour, and Alexandre Défossez. He carries the Chief Language Officer title and leads the language-model side of the program, with the Moshi speech-text foundation model in 2024 and the Helium compact multilingual model in 2025 as the period's most visible artifacts.

Affiliations

Notable contributions

Grave's published record runs from distributional-semantics work at Inria, through the fastText embeddings line and the retrieval-augmented language-modeling line at FAIR, into the LLaMA paper, and on to the Moshi and Helium releases at Kyutai. His Google Scholar profile lists approximately 86,000 citations and an h-index of 52 as of mid-2026.

Investments and boards

No public investor activity on record in AI, semiconductors, datacenters, software, or energy as of May 2026.

Network

Grave's research collaborations trace through three overlapping cohorts. The first is his Inria PhD lineage: Francis Bach and Guillaume Obozinski were his doctoral advisors, and the early statistics-and-NLP papers from 2011 to 2014 were co-authored with them.

The second is FAIR Paris. The fastText line ran with Armand Joulin, Piotr Bojanowski, and Tomas Mikolov. The retrieval-augmented language-modeling line ran with Gautier Izacard, Grave's PhD advisee at Sorbonne and FAIR from 2019 to 2023. The LLaMA-paper byline links Grave to Hugo Touvron as first author and to several authors who later followed each other across the FAIR-Paris-to-Mistral migration, including Timothée Lacroix, Guillaume Lample, Marie-Anne Lachaux, and Thibaut Lavril. Yann LeCun was the chief AI scientist of Meta during his FAIR period.

The third is the Kyutai founding cohort: Pérez (CEO), Hervé Jegou, Laurent Mazaré, Neil Zeghidour, and Alexandre Défossez. The Moshi paper byline collects all six. The Kyutai scientific committee includes Yann LeCun, Yejin Choi, and Bernhard Schölkopf.

Position in the field

As of May 2026, Grave occupies a role at the intersection of two research trajectories. The first is the multilingual-NLP and word-embeddings line that traces through fastText and XLM-R; the line shaped how a generation of practitioners worked with non-English language data and made fastText one of the most widely deployed open-source NLP libraries in the late 2010s. The second is the retrieval-augmented and foundation-model line that traces through Atlas, the LLaMA paper, and the language-model side of Moshi and Helium at Kyutai.

His position differs from the ex-FAIR-Paris peers who founded Mistral AI. The Mistral co-founders moved from FAIR into a venture-backed commercial company; Grave moved into a nonprofit open-research lab funded by Iliad, CMA CGM, and Schmidt Futures. His public-facing presence is also lower-profile than Arthur Mensch or Guillaume Lample, with public communications channeled through arXiv submissions, the @EXGRV X account, and conference appearances. Press coverage of Kyutai has consistently identified Grave as the senior researcher carrying the language-model thread of the lab's program, alongside Zeghidour and Défossez on the speech and audio side.

Outlook

Open questions over the next 6 to 18 months:

  • Helium 2 and successor language-model releases. Whether Kyutai's compact multilingual line continues at the cadence and capability bar set by Helium 1, and whether the lab releases a frontier-scale multilingual successor.
  • Moshi successor and the speech-language stack. The integration of language-model and speech-model research at Kyutai, including any Moshi-2 or full-duplex multilingual successor, with Grave's language-side leadership shaping the language-model component.
  • Open-weights and open-data commitment. Whether Kyutai's open-research and open-data positioning continues at the commitment level set in 2023 and 2024, particularly as the lab's compute and capability footprint scales.
  • PhD pipeline and senior recruitment. Whether Grave's PhD advisees and the broader Sorbonne and Inria pipeline continue feeding researchers into Kyutai, and whether the lab can attract additional senior FAIR alumni.
  • Public-facing posture. Whether the lower-profile communications style of the Kyutai senior research group shifts as the lab's public-launch artifacts (Moshi, Helium, Mimi) attract more press and policy attention.

Sources

About the author
Nextomoro

Nextomoro

nextomoro tracks progress for AI research labs, models, and what's next.

Could not load content

AI Research Lab Intelligence

nextomoro tracks progress for AI research labs, models, and what's next.

AI Research Lab Intelligence

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to AI Research Lab Intelligence.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.