| ▲ | kristianp 13 hours ago | |||||||||||||||||||
How big is Mercury 2? How many tokens is it trained on? Is it's agentic accuracy good enough to operate, say, coding agents without needing a larger model to do more difficult tasks? | ||||||||||||||||||||
| ▲ | volodia 13 hours ago | parent [-] | |||||||||||||||||||
You can think of Mercury 2 as roughly in the same intelligence tier as other speed-optimized models (e.g., Haiku 4.5, Grok Fast, GPT-Mini–class systems). The main differentiator is latency — it’s ~5× faster at comparable quality. We’re not positioning it as competing with the largest models (Opus 4.5, etc.) on hardest-case reasoning. It’s more of a “fast agent” model (like Composer in Cursor, or Haiku 4.5 in some IDEs): strong on common coding and tool-use tasks, and providing very quick iteration loops. | ||||||||||||||||||||
| ||||||||||||||||||||