| ▲ | kcb 2 days ago | ||||||||||||||||||||||
Nemotron 3 Super was released recently. That's a direct competitor to gpt-oss-120b. https://developer.nvidia.com/blog/introducing-nemotron-3-sup... | |||||||||||||||||||||||
| ▲ | evilduck 2 days ago | parent | next [-] | ||||||||||||||||||||||
In terms of ability, maybe, in terms of speed, it's not even close. Check out the Prompt Processing speeds between them: https://kyuz0.github.io/amd-strix-halo-toolboxes/ gpt-oss-120b is over 600 tokens/s PP for all but one backend. nemotron-3-super is at best 260 tokens/s PP. Comparing token generation, it's again like 50 tokens/sec vs 15 tokens/sec That really bogs down agentic tooling. Something needs to be categorically better to justify halving output speed, not just playing in the margins. | |||||||||||||||||||||||
| |||||||||||||||||||||||
| ▲ | coder68 2 days ago | parent | prev [-] | ||||||||||||||||||||||
I gave it a whirl but was unenthused. I'll try it again, but so far have not really enjoyed any of the nvidia models, though they are best in class for execution speed. | |||||||||||||||||||||||
| |||||||||||||||||||||||