▲ | scrollop 5 days ago | |||||||
Makes you think whether llama progress is not doing too well and/or perhaps we're entering a plateau for llm architecture development. | ||||||||
▲ | butlike 5 days ago | parent [-] | |||||||
The article got me thinking that there's some sort of bottle neck that makes scaling astronomical or the value just not really there. 1. Buy up top talent from other's working in this space 2. See what they produce over say, 6mo. to a year 3. Hire a corpus of regular ICs to see what _they_ produce 4. Open source the model to see if any programmer at all can produce something novel with a pretty robust model. Observe that nothing amazing has really come out (besides a pattern-recognizing machine that placates the user to coerce them into using more tokens for more prompts), and potentially call it on hiring for a bubble. | ||||||||
|