| ▲ | dpoloncsak a day ago | |||||||||||||||||||||||||
Yeah, one of my first projects one of my buddies asked "Why aren't you using [ChatGPT 4.0] nano? It's 99% the effectiveness with 10% the price." I've been using the smaller models ever since. Nano/mini, flash, etc. | ||||||||||||||||||||||||||
| ▲ | sixtyj a day ago | parent | next [-] | |||||||||||||||||||||||||
Yup. I have found out recently that Grok-4.1-fast has similar pricing (in cents) but 10x larger context window (2M tokens instead of ~128-200k of gpt-4-1-nano). And ~4% hallucination, lowest in blind tests in LLM arena. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | phainopepla2 a day ago | parent | prev | next [-] | |||||||||||||||||||||||||
I have been benchmarking many of my use cases, and the GPT Nano models have fallen completely flat one every single except for very short summaries. I would call them 25% effectiveness at best. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | walthamstow a day ago | parent | prev [-] | |||||||||||||||||||||||||
Flash Lite 2.5 is an unbelievably good model for the price | ||||||||||||||||||||||||||