▲ | chillee 5 days ago | ||||||||||||||||
No. In some sense, the article comes to the right conclusion haha. But it's probably >100x off on its central premise about output tokens costing more than input. | |||||||||||||||||
▲ | martinald 5 days ago | parent | next [-] | ||||||||||||||||
Thanks for the correction (author here). I'll update the article - very fair point on compute on input tokens which I messed up. Tbh I'm pleased my napkin math was only 7x off the laws of physics :). Even rerunning the math on my use cases with way higher input token cost doesn't change much though. | |||||||||||||||||
| |||||||||||||||||
▲ | doctorpangloss 5 days ago | parent | prev [-] | ||||||||||||||||
I’m pretty sure input tokens are cheap because they want to ingest the data for training later no? They want huge contexts to slice up. | |||||||||||||||||
|