▲ | fmbb 5 days ago | |||||||
The thing that is standing in their way is probably that nobody is willing to pay for this what it costs to run. | ||||||||
▲ | simonw 5 days ago | parent [-] | |||||||
Doesn't look very expensive to me. An LLM capable of this level of summarization can run in ~12GB of GPU-connected RAM, and only needs that while it's running a prompt. The cheapest small LLMs (GPT-4.1 Nano, Google Gemini 1.5 Flash 8B) cost less than 1/100th of a cent per prompt because they are cheap to run. | ||||||||
|