Yuhuai Wu

Yuhuai (Tony) Wu is a Chinese-Canadian machine-learning researcher known for neural theorem proving, autoformalization, and AlphaGeometry, and a former founding team member of xAI from 2023 to 2026.
Yuhuai Wu

Yuhuai Wu

Yuhuai (Tony) Wu is a Chinese-Canadian machine-learning researcher known for work on neural theorem proving, autoformalization, and machine reasoning systems for mathematics. He is a co-author of "Solving Olympiad Geometry Without Human Demonstrations" (the AlphaGeometry paper, Nature, January 2024), the Minerva quantitative-reasoning system, the Memorizing Transformer, and the Autoformalization with Large Language Models paper, and he was one of the eleven publicly named founding team members of xAI from March 2023 until his departure on February 10, 2026. As of May 2026, his post-xAI plans have not been publicly disclosed.

At a glance

Origins

Public biographical material on Wu is comparatively thin. He has no Wikipedia entry as of May 2026, and the available record runs through his personal site at yuhuaiwu.github.io, his X account, his Google Scholar profile, his University of Toronto faculty pages, the TSpace deposit of his PhD thesis, the published papers, and press coverage of the February 2026 xAI departure. He goes publicly by both "Yuhuai Wu" and "Tony Wu"; the parenthetical "(Tony)" appears on his personal site, his published papers, and his GitHub identity.

Wu was born in 1995 in Jiande, Zhejiang Province, China. He attended primary school in Zhejiang and then Hangzhou before relocating to Canada for high school. His undergraduate years at the University of New Brunswick were in mathematics; he graduated with a BSc in 2015. The transition into machine learning followed an unsolicited 2014 email he sent as a pure-mathematics student to Yoshua Bengio, who responded by offering him a summer internship at Mila in 2015. Wu has cited that internship as the start of his ML career.

Career

Wu joined the University of Toronto Department of Computer Science as a PhD student in 2015, under the joint supervision of Roger Grosse and Jimmy Ba. The thesis acknowledgements identify Christian Szegedy as his mentor in the field of mathematical reasoning, dating the relationship to a Google internship; Szegedy and Wu would later overlap at xAI. A summer 2018 visit to DeepMind put Wu in contact with James McClelland, whose talk on mathematical reasoning convinced Wu to focus the thesis on the topic and who later helped recruit him to a Stanford postdoc with Percy Liang. The thesis "Neural Networks for Mathematical Reasoning: Evaluations, Capabilities, and Techniques" was deposited at TSpace in March 2024, with the degree formally awarded in 2024.

The PhD period produced the neural-theorem-proving line that has run through Wu's subsequent career. "INT: An Inequality Benchmark for Evaluating Generalization in Theorem Proving" (Wu, Jiang, Ba, Grosse, ICLR 2021) introduced an inequality benchmark for generalization in neural theorem proving. "LIME: Learning Inductive Bias for Primitives of Mathematical Reasoning" (Wu, Rabe, Li, Ba, Grosse, Szegedy, ICML 2021) examined synthetic pretraining tasks that injected useful inductive bias into transformer models targeting mathematical reasoning.

Wu joined Google Research as a Research Scientist in 2021 on the N2Formal team led by Christian Szegedy. "Memorizing Transformers" (Wu, Rabe, Hutchins, Szegedy, ICLR 2022 Spotlight) introduced an external memory architecture extending transformer context to handle large mathematical libraries. "Autoformalization with Large Language Models" (Wu, Jiang, Li, Rabe, Staats, Jamnik, Szegedy, NeurIPS 2022) reported that LLMs could correctly translate roughly 25 percent of competition-math problems into machine-checkable Isabelle/HOL specifications without supervised training, framing autoformalization as a viable program for AI-driven formal mathematics. "Draft, Sketch, and Prove" (Jiang et al., ICLR 2023 Oral) extended the program by training models to produce informal proof sketches that guide formal-prover search. Wu was also a co-author on "Solving Quantitative Reasoning Problems with Language Models" (Lewkowycz et al., NeurIPS 2022), introducing Minerva, the language model trained for technical and quantitative reasoning, and on the PaLM 2 technical report (Anil et al., 2023).

In March 2023, Wu left Google to join xAI as one of the eleven publicly named founding team members. The team was assembled from senior researchers at Google DeepMind, OpenAI, and Google Brain, and the launch was publicly announced on July 12, 2023. Press coverage of his subsequent departure characterized Wu as a leader on the Grok 3 reasoning effort. Specific functional titles beyond "founding team member" are not publicly stated.

In January 2024, "Solving Olympiad Geometry Without Human Demonstrations" (Trinh, Wu, Le, He, Luong, Nature 2024) reported that the AlphaGeometry system solved 25 of 30 International Mathematical Olympiad geometry problems within standard contest time limits, a gold-medalist-level performance. The work was carried out at Google DeepMind, with Wu listed among the principal authors.

On February 10, 2026, Wu posted on X that he had resigned from xAI. The post described the company and founding team as a "family", thanked colleagues for shared work, and framed the next chapter as one in which "a small team armed with AIs can move mountains and redefine what's possible." Bloomberg, Silicon Republic, and TechCrunch covered the announcement, characterizing the departure as the sixth among the original twelve named co-founders. Jimmy Ba announced his own departure roughly seventeen hours later. The departures came shortly after the SpaceX acquisition of xAI on February 2, 2026, in an all-stock transaction valuing the combined entity at $1.25 trillion. Press coverage cited tensions between xAI's research orientation and the operational intensity following the SpaceX integration. As of May 2026, Wu's post-xAI plans have not been publicly disclosed beyond the framing in his X post.

Affiliations

  • University of New Brunswick: BSc in Mathematics, 2011 to 2015.
  • Mila: Research intern under Yoshua Bengio, summer 2015.
  • University of Toronto, Department of Computer Science: PhD candidate, 2015 to 2024 (degree formally awarded 2024), advised by Roger Grosse and co-advised by Jimmy Ba.
  • Google DeepMind: Research intern, summer 2018.
  • Stanford University: Postdoctoral researcher with Percy Liang and James McClelland, approximately 2021.
  • Google Research: Research Scientist, N2Formal team, 2021 to 2023.
  • xAI: Founding team member, March 2023 to February 10, 2026.

Notable contributions

Wu's published record concentrates in neural theorem proving, autoformalization, mathematical-reasoning benchmarks, and large-language-model methods for formal and informal mathematics. His Google Scholar profile lists the foundation-models survey, AlphaStar, HELM, PaLM 2, and Minerva among the most-cited entries he has co-authored, with the autoformalization and theorem-proving papers as his principal first- and joint-first-authored contributions.

Investments and boards

No public investor activity on record in AI, semiconductors, datacenters, software, or energy as of May 2026.

Network

Wu's longest-running professional relationships fall in three cohorts. The first is the University of Toronto machine-learning group, including his thesis advisors Roger Grosse and Jimmy Ba and frequent co-author Albert Q. Jiang on the formal-mathematics line. The second is the Google Research N2Formal team and the formal-mathematics community, including Christian Szegedy (Wu's mentor in the field, later xAI co-founder, now founder of Math Inc), Markus Rabe, Wenda Li, DeLesley Hutchins, and the AlphaGeometry team at Google DeepMind led by Trieu Trinh and Thang Luong. The third is the xAI founding team. Beyond Elon Musk, the founding cohort included Igor Babuschkin (departed August 2025 to launch Babuschkin Ventures), Greg Yang (informal advisor since January 2026), Christian Szegedy (departed February 2025 to join Morph Labs and later found Math Inc), Jimmy Ba (departed February 10, 2026), Manuel Kroiss, Toby Pohlen, Ross Nordeen, Kyle Kosic, Guodong Zhang, and Zihang Dai. Yoshua Bengio remains a long-standing mentor since the 2015 Mila internship.

Position in the field

As of May 2026, Wu occupies a structurally distinctive position in the mathematical-reasoning research line. The combination of co-authorship on AlphaGeometry, on Minerva, on the autoformalization paper, and on the Memorizing Transformer places him among the small group of researchers most associated with the contemporary push to make large language models capable of formal and informal mathematical reasoning. The Google Scholar citation count above 35,000 is high for the post-PhD career stage. Press coverage of his February 2026 departure from xAI consistently led with the AlphaGeometry and autoformalization attribution.

His work is the most direct intellectual lineage between the academic neural-theorem-proving line and the production reasoning capabilities now embedded in frontier large language models. The autoformalization research program he helped frame at Google has continued at xAI through the Grok reasoning effort, at Math Inc under Christian Szegedy, and at Google DeepMind through AlphaProof and AlphaGeometry 2. His public-commentary cadence is moderate: he posts intermittently to @Yuhu_ai_, with limited solo English-language video presence beyond the NeurIPS Chats episode on Minerva and bootstrapping reasoning, the December 2023 DAI keynote on autoformalization in Singapore, and academic seminar appearances.

Outlook

Open questions over the next 6 to 18 months:

  • Post-xAI role. Whether Wu starts a new venture aligned with the framing in the February 10 X post, joins another industry research lab, or returns to academic publication cadence. The departure post pointed toward a small-team-and-AIs configuration without naming a destination.
  • xAI relationship. Whether the post-departure relationship remains technically active, given the Grok 3 reasoning lineage and the prior collaboration with Christian Szegedy at Google.
  • Autoformalization trajectory. Whether Wu continues to publish on the autoformalization line that links his Google, xAI, and the broader formal-mathematics community, including the Lean and Mathlib ecosystems.
  • Math Inc collaboration. Whether the working relationship with Szegedy at Math Inc resumes commercially or publicly, given the shared autoformalization research program.
  • Public-commentary cadence. Whether Wu's posting cadence and his seminar-and-podcast schedule pick up after the xAI departure.

Sources

About the author
Nextomoro

Nextomoro

nextomoro tracks progress for AI research labs, models, and what's next.

AI Research Lab Intelligence

nextomoro tracks progress for AI research labs, models, and what's next.

AI Research Lab Intelligence

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to AI Research Lab Intelligence.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.