Elorian

Elorian is an American AI lab founded in 2025 by former Google DeepMind and Apple research leaders Andrew Dai and Yinfei Yang, building native multimodal models with a focus on visual reasoning rather than image-to-text translation.
Elorian

Elorian

Elorian is an American artificial intelligence lab founded in 2025 by Andrew Dai, formerly a senior researcher and Gemini data co-lead at Google DeepMind, and Yinfei Yang, formerly chief research scientist on Apple's machine-learning team. The company is based in Palo Alto and develops native multimodal models that reason directly across images, video, audio, and text rather than converting visual inputs into text labels before processing. As of April 2026, Elorian had raised approximately $55 million in seed funding co-led by Striker Venture Partners, Menlo Ventures, and Altimeter, at a reported valuation of approximately $500 million.

At a glance

  • Founded: 2025 in Palo Alto, California. Public announcement April 9, 2026.
  • Status: Private. Reported valuation approximately $500 million.
  • Funding: Approximately $55 million seed round. Co-led by Striker Venture Partners, Menlo Ventures, and Altimeter.
  • CEO: Andrew Dai (co-founder, formerly co-lead of Gemini data at Google DeepMind, 14-year tenure).
  • Other notable leadership: Yinfei Yang (co-founder, formerly principal research scientist at Apple), Seth Neel (co-founder, formerly assistant professor at Harvard).
  • Open weights: None disclosed.
  • Flagship products: No public model release as of April 2026. Research focus on native multimodal models with visual-reasoning capability.

Origins

Elorian was founded quietly in 2025 by Andrew Dai and Yinfei Yang and announced publicly on April 9, 2026 alongside a $55 million seed round.

Dai had spent 14 years at Google and Google DeepMind, with senior roles spanning the development of foundational language-model research and, in his final tenure, co-leadership of the data program for the Gemini family of frontier models. Yang had been a principal research scientist on Apple's machine-learning team for several years, with research focus on multimodal models and vision-language systems, before departing in late 2025. Seth Neel, formerly an assistant professor at Harvard with research focus on data and AI, joined as a co-founder.

The company assembled a research team of 11 to 50 people through late 2025 and early 2026. Public reporting describes the team as drawn predominantly from Google DeepMind, Apple, and adjacent academic-research backgrounds. Named team members include Forrest Huang, Gautam Kumar, Jared Lichtarge, Jihua Huang, Le Xue, Marcella Valentine, Qiuyi "Richard" Zhang, and Seo-Jin Bang.

Mission and strategy

Elorian's stated mission is to build AI systems that "understand and reason through visual information the way humans do." The company has described visual reasoning as "the most important work in AI today," framing the challenge as a structural limitation of current frontier models that rely on converting images into text labels before reasoning. According to Elorian's public framing, this two-step pattern is "fragile, limited, and prone to hallucination," particularly for tasks that humans solve visually without translating to language first.

The technical approach combines multimodal training with novel architectures for cross-modal reasoning. Rather than treating images as static inputs that get encoded into a fixed representation, Elorian's models are designed to actively engage with visual representations and progressively reason about spatial relationships, structural constraints, and physical-world dynamics. Dai has described the company's product framing as "a native multimodal model that can simultaneously understand and process text, images, videos, and audio" rather than a stitched-together system in which a vision encoder feeds tokenized representations into a language model.

The strategy combines three threads. First, foundational research on native multimodal architectures with vision as a first-class reasoning surface. Second, a research-first commercial posture in which capability is developed before product surface is defined. Third, a deliberate positioning contrast against the dominant frontier-model architecture pattern in which language is the central reasoning modality and other modalities are encoded into the language space.

The competitive premise is that vision-as-language-token approaches have structural limits for tasks where the underlying logic is fundamentally visual (engineering design, physical-world reasoning, medical imaging) and that a native multimodal architecture will produce capability advantages in those domains.

Models and products

  • Foundational research output. Elorian's primary public output as of April 2026 consists of the founding-team announcement and the company's stated technical thesis. No model architecture or capability target has been comprehensively disclosed.
  • No shipped models, APIs, or open weights. The company has not released a model, an API, or open weights as of April 2026.

The commercial distribution strategy beyond research output has not been publicly stated. Public framing has named potential application domains including engineering design, robotics, medical research, disaster response, and agricultural monitoring, but no specific product surface has been disclosed.

Benchmarks and standing

Elorian has not released a model and is not represented on the standardized capability leaderboards as of April 2026. The company's standing rests on the founders' research credentials at Google DeepMind, Apple, and Harvard, the Striker-Menlo-Altimeter lead-investor combination, and the reported valuation of approximately $500 million.

The standardized leaderboards measure language-and-reasoning capability and increasingly vision-language capability, but do not yet have established benchmarks for the deeper visual-reasoning capability that Elorian's research thesis targets. Direct evaluation of the company's research output against frontier models will likely require new benchmarks specific to native multimodal reasoning.

Leadership

As of April 2026, Elorian's named leadership and team includes:

  • Andrew Dai, co-founder and chief executive. Fourteen-year tenure at Google and Google DeepMind, including co-leadership of the data program for the Gemini family. Public face for the company on technical claims and product framing.
  • Yinfei Yang, co-founder. Formerly principal research scientist on Apple's machine-learning team, with research focus on multimodal models and vision-language systems.
  • Seth Neel, co-founder. Formerly an assistant professor at Harvard with research focus on data and AI.
  • Named research team, including Forrest Huang, Gautam Kumar, Jared Lichtarge, Jihua Huang, Le Xue, Marcella Valentine, Qiuyi "Richard" Zhang, and Seo-Jin Bang. Roles and prior backgrounds not comprehensively disclosed.

The senior research and engineering team includes additional unnamed hires drawn predominantly from Google DeepMind, Apple, and adjacent academic-research backgrounds.

Funding and backers

Elorian's funding history through April 2026 consists of a single closed round: the approximately $55 million seed round announced in April 2026. The round was co-led by Striker Venture Partners, Menlo Ventures, and Altimeter.

The Menlo Ventures lead is consistent with the firm's pattern of investing in research-first AI Insurgents, including a parallel position in Inception Labs. The Altimeter participation places Elorian within the senior-tier venture-capital syndicate for mid-size pre-product AI labs. Specific co-investors beyond the three lead firms have not been comprehensively disclosed.

The company has not disclosed cumulative compute commitments, cloud-partner relationships, or follow-on financing plans. The reported valuation of approximately $500 million is consistent with mid-size 2025-vintage Insurgent labs founded by senior-team founder-departure cohorts, comparable to Adaption Labs and Inception Labs.

Industry position

Elorian occupies a distinctive position within the 2025-vintage Insurgent cohort. The combination of a senior Google DeepMind and Apple founder team, the visual-reasoning technical thesis, and the multi-firm lead-investor combination produces a profile that overlaps with the broader research-first Insurgent cohort but is differentiated on technical thesis.

The closest peer comparators are research-first Insurgents pursuing alternatives to the dominant scaling-and-language-centric architectural pattern. World Labs, founded by Fei-Fei Li, pursues spatial-intelligence research with a related focus on visual and physical-world reasoning. Adaption Labs pursues continuous-learning AI. Inception Labs pursues diffusion-based generation.

The strategic risks are substantial. The visual-reasoning thesis has not been validated through a shipping product or public model release. The native-multimodal architectural approach, while differentiated, has not been demonstrated at frontier-model scale. The valuation depends on team credentials and lead-investor signals rather than capability evidence.

The strategic strengths are distinctive. Andrew Dai's role co-leading the data program for Gemini provides direct context on what a frontier multimodal training program looks like inside one of the most successful frontier labs. Yinfei Yang's principal-research-scientist tenure at Apple's ML team provides depth on multimodal-model engineering at scale. The team's mix of Google DeepMind, Apple, and academic-research backgrounds gives Elorian breadth across the relevant research traditions.

Competitive landscape

Elorian competes with several Frontier and Insurgent labs:

  • Google DeepMind. Dai's previous lab. Gemini's native multimodal architecture is the closest existing comparator to Elorian's research thesis, though the two have diverged on commercial focus.
  • OpenAI and Anthropic. Dominant frontier labs whose vision-language capability is integrated into GPT-5.5 and Claude Opus 4.7. Elorian's visual-reasoning thesis is positioned as architecturally differentiated.
  • Apple. Yang's previous employer. Apple's Apple Intelligence framework integrates third-party LLMs rather than developing native multimodal frontier models in-house, leaving Elorian as a potential acquisition or partnership target.
  • World Labs. Closest peer Insurgent on visual-and-physical-world-reasoning grounds. Both founded in 2024 to 2025 by senior research-leader departures, both pursuing visual or spatial intelligence as central research themes.
  • Black Forest Labs and Midjourney. Image-generation specialists. Compete on visual capability at the generation surface rather than the reasoning surface that Elorian targets.
  • Wayve and physical-world AI labs. Adjacent in pixel-space and physical-world reasoning research, though focused on driving rather than general visual reasoning.

Outlook

Several open questions affect Elorian's trajectory in 2026 and 2027:

  • The first published research output, including the company's specific native-multimodal architecture and any empirical demonstration at scale.
  • The first model release, if any, and the capability profile relative to frontier-tier multimodal comparators.
  • The commercial strategy beyond research output, which has not been publicly stated.
  • Whether Elorian accepts follow-on capital at a higher valuation, and on what timeline.
  • Senior-talent recruitment from Google DeepMind, Apple, and the broader academic ML community.
  • Strategic-partner relationships, particularly with hardware-platform companies for which native multimodal capability could be a distribution wedge.

Sources

About the author
Nextomoro

Nextomoro

nextomoro tracks progress for AI research labs, models, and what's next.

AI Research Lab Intelligence

nextomoro tracks progress for AI research labs, models, and what's next.

AI Research Lab Intelligence

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to AI Research Lab Intelligence.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.