modelstop.top

Compare Models

Run side-by-side checks for pricing, context window, and latency.

Inception: Mercury Coder

inception

Mercury Coder is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like Claude 3.5 Haiku...

Context window
128,000 tokens
Input cost
$0.25 / 1M
Output cost
$0.75 / 1M
Latency (p50)