Inception Models
Inception logo

Mercury Coder

by Inception

Mercury Coder is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like Claude 3.5 Haiku and GPT-4o Mini while matching their performance. Mercury Coder's speed means that developers can stay in the flow while coding, enjoying rapid chat-based iteration and responsive code completion suggestions. On Copilot Arena, Mercury Coder ranks 1st in speed and ties for 2nd in quality. Read more in the [blog post here](https://www.inceptionlabs.ai/blog/introducing-mercury).

Chat with Mercury Coder
Input Price$0.25/1M tokens
Output Price$0.75/1M tokens
Intelligence32.8
Coding30.6

Specifications

Technical details and pricing.

ProviderInception
Context Window128,000 tokens
Release DateFeb 20, 2026
ModalitiesText

Benchmarks

7 benchmark scores from Artificial Analysis.

GPQA77.0%
HLE15.5%
SciCode38.7%
LCR36.3%
IFBench69.8%
Tau270.8%
TerminalBench Hard26.5%

Composite Indices

Intelligence, Coding, Math

Standard Benchmarks

Academic and industry benchmarks

Frequently Asked Questions

What is Mercury Coder good for?

Use Mercury Coder for everyday tasks like writing, summarizing, brainstorming, and getting clear explanations.

How much does Mercury Coder cost?

Pricing is based on usage. Current rates are $0.25/1M tokens for input and $0.75/1M tokens for output.

Can I try Mercury Coder for free?

Yes. You can start a chat instantly and test the model before deciding on a plan.

Does Mercury Coder support images or audio?

Mercury Coder focuses on text-based tasks.

Benchmarks and pricing are sourced from Artificial Analysis where available. OpenRouter specs are used as a fallback.