| ▲ | Nihilartikel 7 hours ago | ||||||||||||||||
I'm not super deep on LLM development, but with ram being a material bottleneck and from what I've read about DeepSeek's results with offloading factual knowledge with 'engrams' I think that the near future will start moving towards the dense core of LLMs focusing much more on a distillation of universal reasoning and logic while factual knowledge is pushed out into slower storage. IIRC Nvidia's Nemotron Cascade is taking MoE even further in that direction too. I don't need a coding model to be able to give me an analysis of the declaration of independence in urdu from 'memory' and the price in ram for being able to do that, impressive as it is, is an inefficiency. | |||||||||||||||||
| ▲ | axus 4 hours ago | parent [-] | ||||||||||||||||
Were he still corporeal, L. Ron would be all over this AI stuff. | |||||||||||||||||
| |||||||||||||||||