| ▲ | PlatoIsADisease 3 hours ago | ||||||||||||||||||||||
>The model absolutely can be run at home. There is a huge difference between "look I got it to answer the prompt: '1+1='" and actually using it for anything of value. I remember early on people bought Macs (or some marketing team was shoveling it), and proposing people could reasonably run the 70B+ models on it. They were talking about 'look it gave an answer', not 'look this is useful'. While it was a bit obvious that 'integrated GPU' is not Nvidia VRAM, we did have 1 mac laptop at work that validated this. Its cool these models are out in the open, but its going to be a decade before people are running them at a useful level locally. | |||||||||||||||||||||||
| ▲ | esafak 2 hours ago | parent [-] | ||||||||||||||||||||||
Hear, hear. Even if the model fits, a few tokens per second make no sense. Time is money too. | |||||||||||||||||||||||
| |||||||||||||||||||||||