| ▲ | blcknight 4 hours ago | ||||||||||||||||||||||||||||
There is a cognitive ceiling for what you can do with smaller models. Animals with simpler neural pathways often outperform whatever think they are capable of but there's no substitute for scale. I don't think you'll ever get a 4B or 8B model equivalent to Opus 4.6. Maybe just for coding tasks but certainly not Opus' breadth. | |||||||||||||||||||||||||||||
| ▲ | zarzavat 3 hours ago | parent | next [-] | ||||||||||||||||||||||||||||
The only thing that we are sure can't be highly compressed is knowledge, because you can only fit so much information in given entropy budget without losing fidelity. The minimal size limits of reasoning abilities are not clear at all. It could be that you don't need all that many parameters. In which case the door is open for small focused models to converge to parity with larger models in reasoning ability. If that happens we may end up with people using small local models most of the time, and only calling out to large models when they actually need the extra knowledge. | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||
| ▲ | charcircuit 3 hours ago | parent | prev | next [-] | ||||||||||||||||||||||||||||
I think you are underestimating the strength a small model can get from tool use. There may be no substitute for scale, but that scale can live outside of the model and be queried using tools. In the worst case a smaller model could use a tool that involves a bigger model to do something. | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||
| ▲ | dathinab 2 hours ago | parent | prev [-] | ||||||||||||||||||||||||||||
except you don't want knowledge in the model, and most of that "size" comes from "encoded knowledge", i.e. over fitting. The goal should be to only have language handling in the model, and the knowledge in a database you can actually update, analyze etc. It's just really hard to do so. "world models" (for cars) maybe make sense for self driving, but they are also just a crude workaround to have a physics simulation to push understanding of physics. Through in difference to most topics, basic, physics tend to not change randomly and it's based on observation of reality, so it probably can work. Law, health advice, programming stuff etc. on the other hand changes all the time and is all based on what humans wrote about it. Which in some areas (e.g. law or health) is very commonly outdated, wrong or at least incomplete in a dangerous way. And for programming changes all the time. Having this separation of language processing and knowledge sources is ... hard, language is messy and often interleaves with information. But this is most likely achievable with smaller models. Actually it might even be easier with a small model. (Through if the necessary knowledge bases are achievable to fit on run on a mac is another topic...) And this should be the goal of AI companies, as it's the only long term sustainable approach as far as I can tell. I say should because it may not be, because if they solve it that way and someone manages to clone their success then they lose all their moat for specialized areas as people can create knowledge bases for those areas with know-how OpenAI simple doesn't have access to. (Which would be a preferable outcome as it means actual competition and a potential fair working market.) | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||