Mercury Coder
by Inception
Mercury Coder is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like Claude 3.5 Haiku and GPT-4o Mini while matching their performance. Mercury Coder's speed means that developers can stay in the flow while coding, enjoying rapid chat-based iteration and responsive code completion suggestions. On Copilot Arena, Mercury Coder ranks 1st in speed and ties for 2nd in quality. Read more in the [blog post here](https://www.inceptionlabs.ai/blog/introducing-mercury).
Specifications
Technical details and pricing.
Benchmarks
7 benchmark scores from Artificial Analysis.
Composite Indices
Intelligence, Coding, Math
Standard Benchmarks
Academic and industry benchmarks
Frequently Asked Questions
What is Mercury Coder good for?
Use Mercury Coder for everyday tasks like writing, summarizing, brainstorming, and getting clear explanations.
How much does Mercury Coder cost?
Pricing is based on usage. Current rates are $0.25/1M tokens for input and $0.75/1M tokens for output.
Can I try Mercury Coder for free?
Yes. You can start a chat instantly and test the model before deciding on a plan.
Does Mercury Coder support images or audio?
Mercury Coder focuses on text-based tasks.
Similar Models
Other models you might want to explore.
Benchmarks and pricing are sourced from Artificial Analysis where available. OpenRouter specs are used as a fallback.