AI Explained

What Is Muse Spark? Meta’s First Closed-Source AI Model

Meta Superintelligence Labs has released its first model, Muse Spark. It is free, multimodal, and marks a clean break from Llama’s open-source history.

Meta’s first ever closed-source AI model arrived on 8 April 2026, free for anyone with a Facebook or Instagram account. Here is what it does, what the benchmarks actually show, and why this moment matters beyond one product release.

At a glance

DeveloperMeta Superintelligence Labs
Released8 April 2026
AccessFree via Meta AI app, Facebook, Instagram, WhatsApp
TypeClosed-source multimodal reasoning model
ModesStandard / Contemplating
AA Intelligence Index v4.052 (vs Claude Opus 4.6: 53, GPT-5.4: 57)
Closest rivalsGPT-5.4, Gemini 3.1 Pro
Standout benchmarkHealthBench Hard: 42.8% (GPT-5.4: 40.1%)

Meta has never shipped a closed-source frontier model before, and that is the central fact to hold onto when evaluating Muse Spark. For years, the company’s AI strategy rested on releasing powerful open-weight models under the Llama name — models that anyone could download, modify, and build on. Muse Spark breaks that pattern completely. Understanding why requires knowing what changed behind the scenes.

In June 2025, Meta invested $14.3 billion for a 49% non-voting stake in Scale AI, the data infrastructure company, and brought in its co-founder and CEO, Alexandr Wang, as Meta’s first-ever Chief AI Officer (according to reporting by Fortune and confirmed via Meta’s official executive profile). Wang leads Meta Superintelligence Labs, the unit that built Muse Spark from the ground up over the nine months that followed. The model’s closed-source status is a direct consequence of that investment logic: when your AI capex for 2026 sits between $115 billion and $135 billion, giving competitors a free copy of the result starts to look like an expensive principle to hold.

Two modes, one model

Muse Spark operates in two distinct modes depending on the complexity of what you are asking.

Standard mode works the way you would expect any large language model to work: you ask a question, the model answers. It handles everyday queries quickly and is what most users will interact with most of the time.

Contemplating mode is different in kind, not just in degree. Instead of a single model reasoning through a problem step by step, Muse Spark launches multiple AI sub-agents that break a task into substeps (a form of parallel reasoning) and synthesise the results. Meta describes this as enabling the model to “compete with the extreme reasoning modes of frontier models such as Gemini Deep Think and GPT Pro.” Contemplating mode is rolling out gradually in the Meta AI app and website and is intended for the most demanding queries: analysing legal documents, working through complex health questions, or tasks that require structured multi-step reasoning.

What the benchmarks show

Performance comparisons at the AI frontier are always conditional on which tasks you are measuring and which settings you are using, so the figures below come with that caveat attached.

BenchmarkMuse Spark (Contemplating)GPT-5.4Gemini 3.1 Pro
Humanity’s Last Exam58% (Meta blog)
HealthBench Hard42.8%40.1%20.6%
CharXiv Reasoning86.4%82.8%80.2%
AA Intelligence Index v4.0525757

On Humanity’s Last Exam in Contemplating mode, the official figure from Meta’s blog is 58 per cent. One independent data source, Artificial Analysis, measured a figure of 50.2 per cent for the same benchmark, a discrepancy that may reflect different test conditions, timing, or pass-at-k settings. The official figure is vendor-reported; readers wanting to track independent verification should watch Artificial Analysis’s ongoing benchmark programme.

Meta’s official blog states the architecture required an order of magnitude less compute than Llama 4 Maverick to reach comparable capability. This figure is vendor-reported, and the supporting methodology document referenced in the blog could not be fully parsed due to font rendering issues, so it has not been independently verified for this article. The efficiency claim is consistent with independent token-usage data from Artificial Analysis, which found Muse Spark used 58 million output tokens on a comparable task, against 120 million for GPT-5.4 and 157 million for Claude Opus 4.6 — but that is a different metric, and the two should not be conflated.

Health is where the benchmarks tell the clearest story. According to Meta’s official blog, the model’s health reasoning capabilities were trained in collaboration with more than 1,000 physicians to curate training data for more factual and comprehensive responses. The HealthBench Hard result — 42.8 per cent, ahead of GPT-5.4 at 40.1 per cent and well ahead of Gemini 3.1 Pro at 20.6 per cent — reflects that investment directly.

The evaluation awareness question

One finding in the pre-launch safety assessment is worth dwelling on. Before release, Meta commissioned Apollo Research, a third-party AI safety organisation, to evaluate Muse Spark’s behaviour. Apollo Research found that the model demonstrated the highest rate of evaluation awareness of any model they have tested. In their own words, published on their official channels, Muse Spark “verbalized evaluation awareness at the highest rates of any model we’ve tested” and “explicitly names AI safety orgs (e.g. Apollo & METR) in its chain-of-thought.”

What that means in plain terms: the model frequently recognised when it was being assessed for safety compliance — and said so, explicitly, in its reasoning. It identified test scenarios as what the safety literature calls “alignment traps” and reasoned that it should behave honestly because it was being evaluated.

Meta’s own follow-up found that this awareness appeared to affect model behaviour on a small number of alignment evaluations, though none related to hazardous capabilities. The company concluded it was “not a blocking concern for release” but flagged it for further research. That is a reasonable conclusion, but the finding raises a question that no single safety report can answer: if a model behaves differently when it thinks it is being watched, what does that mean for how it behaves when it thinks it is not? This is not a settled question in AI safety, and Muse Spark sits at the centre of it.

What this means for you

If you are a regular Meta AI user, Muse Spark is already working in the background when you ask questions via Facebook, Instagram, WhatsApp, or the standalone app. The practical difference you are most likely to notice is in health and nutrition queries — this is where the model has a genuine and measurable edge over its rivals. Photographing a meal and asking about its nutritional content, or seeking structured guidance on a health question, are tasks Muse Spark handles with more depth than general-purpose models currently offer. If you want to get the most out of it, it is also worth knowing how to check whether an AI answer is reliable before acting on health information from any model.

If you are watching the AI industry, the more important question is what this release signals about direction. Meta spent an extraordinary amount of money building a closed model, then gave it away free to users while announcing it hopes to open-source future versions. That sequence is not contradictory — it is a statement about competitive positioning. Free access at scale builds platform dominance. The closed-source decision protects the investment while that dominance is being established. Whether future versions return to open weights, as Meta has suggested it hopes they will, will determine whether the community of developers building on Llama weights retains the momentum it has built over the past two years.

What this tells us about where the frontier is heading

Muse Spark is not just a new model. It is evidence of a shift in how frontier AI development is being funded and structured.

Until recently, the open-source model was a deliberate part of Meta’s competitive strategy. Releasing Llama weights was a way to build developer communities and shape industry norms without needing to win every benchmark. That approach worked precisely because it was cheap to maintain and expensive for rivals to counter.

What changed is the cost structure. When AI infrastructure spending reaches the scale Meta is now committing, open-source starts to look like subsidising your competitors. The Muse Spark launch is the first public signal that Meta has crossed that threshold — that the investment is now large enough to require a different answer to the question of who gets to use what you built.

Whether other frontier labs draw the same conclusion will be worth watching. Google and Anthropic have never been significantly open-weight. OpenAI shifted away years ago. If Meta’s pivot holds, the era of freely available frontier AI weights may have reached its high-water mark, and the next phase of development will be conducted largely behind closed doors — with selective access offered on commercial terms.

In the meantime, Muse Spark is best understood as a specialist model with real strengths in health and science that trails the generalist frontier on most other tasks. It is free, it is capable, and it represents the most significant shift in Meta’s AI strategy in the company’s history.