| ▲ | mhitza 15 hours ago | ||||||||||||||||||||||
Comment retracted. My bad, missed some details. | |||||||||||||||||||||||
| ▲ | pants2 15 hours ago | parent | next [-] | ||||||||||||||||||||||
Reading such obvious LLM-isms in the announcement just makes me cringe a bit too, ex. > We optimize for speed users actually feel: responsiveness in the moments users experience — p95 latency under high concurrency, consistent turn-to-turn behavior, and stable throughput when systems get busy. | |||||||||||||||||||||||
| ▲ | selcuka 15 hours ago | parent | prev [-] | ||||||||||||||||||||||
I think your comment is a bit unfair. > no reasoning comparison Benchmarks against reasoning models: https://www.inceptionlabs.ai/blog/introducing-mercury-2 > no demo https://chat.inceptionlabs.ai/ > no info on numbers of parameters for the model This is a closed model. Do other providers publish the number of parameters for their models? > testimonials that don't actually read like something used in production Fair point. | |||||||||||||||||||||||
| |||||||||||||||||||||||