▲ | brokegrammer 2 days ago | |
My thoughts exactly. Inference should be dirt cheap for LLMs to truly become powerful. It's similar to how computing used to be restricted to mega corps 100 years ago, but today, a smartphone has more computing power than any old age mainframe. Today we need Elon Musk to buy 5 million GPUs to train a model. Tomorrow, we should be able to train a top of the line model using a budget RTX card. | ||
▲ | 999900000999 2 days ago | parent [-] | |
Tbh, if the model is small enough you can train locally. I don't need my code assistant to be an expert on Greek myths. The future is probably highly specialized mini llms. I might train a model to code my way. I'm not that smart enough to figure this out, but the solution can't be to just brute force training with more gpus. There is another answer. |