Remix.run Logo
maplethorpe 2 days ago

Can you summarise? I only reached your comment after scrolling past all the others and I still don't have the answer.

Is the new data that models are more useful for coding than they once were?

dwaltrip a day ago | parent | next [-]

Cost of tokens goes down over time. Like by a lot. And it will continue to do so.

Imagine being in 2003 and saying compute costs won’t go down. That’s Ed lol.

EDIT: Some quick research on this so you guys have actual numbers: https://gist.github.com/dwaltrip/a037be938d2b5ecc8b8b238736e....

There's multiple separate angles that all contribute to token-costs going down: chip improvements, engineering improvements for running inference in general, AI architecture and training advances that give similar intelligence in a smaller model, improvements in the quality of the training data, data center design / economies of scale, networking and rack-level improvements that are multiplicative with chip advancements, and so on...

If you analyze the situation for 5 minutes, it's blindingly obvious that price-per-token will continue to improve. And there's a very similar case for intelligence-per-token as well.

And don't get me wrong -- I have many concerns about how this is all unfolding and how it will impact society. But let's get our basic facts straight.

margalabargala 2 days ago | parent | prev [-]

That sounds like a reading comprehension skill issue? In which case I don't see why me summarizing would move the needle.

But if it helps, no, the data being discussed is surrounding the economics of running inference and R&D, nothing to do with the utility of models for coding.

maplethorpe a day ago | parent [-]

Yours is the first from the top to mention this. You might want to consider the physical location of your comment before telling people to read the thread. We could do without the rudeness, too.