| ▲ | make3 3 hours ago | ||||||||||||||||
LLMs in their current form rely heavily on the vast amount of human data that's available, to learn from it as a first step (the second step is RL). We're not building a language for LLMs just yet. | |||||||||||||||||
| ▲ | jaggederest 2 hours ago | parent | next [-] | ||||||||||||||||
> We're not building a language for LLMs just yet. Working on it, actually! I think it's a really interesting problem space - being efficient on tokens, readable by humans for review, strongly typed and static for reasoning purposes, and having extremely regular syntax. One of the biggest issues with symbols is that, to a human, matching parentheses is relatively easy, but the models struggle with it. I expect a language like the one I'm playing with will mature enough over the next couple years that models with a knowledge cutoff around 1/2027 will probably know how to program it well enough for it to start being more viable. One of the things I plan to do is build evals so that I can validate the performance of various models on my as yet only partially baked language. I'm also using only LLMs to build out the entire infrastructure, mostly to see if it's possible. | |||||||||||||||||
| |||||||||||||||||
| ▲ | energy123 2 hours ago | parent | prev [-] | ||||||||||||||||
It's worth asking why we haven't had the AlphaZero moment for general learning yet, where no human data is needed. | |||||||||||||||||