| ▲ | r_lee 4 hours ago | ||||||||||||||||
I think that's because there's less local AI usage now since there's all kinds of image models by the big labs, so there's really no rush of people self hosting stable diffusion etc anymore the space moved from Consumer to Enterprise pretty fast due to models getting bigger | |||||||||||||||||
| ▲ | zozbot234 4 hours ago | parent [-] | ||||||||||||||||
Today's free models are not really bigger when you account for the use of MoE (with ever increasing sparsity, meaning a smaller fraction of active parameters), and better ways of managing KV caching. You can do useful things with very little RAM/VRAM, it just gets slower and slower the more you try to squeeze it where it doesn't quite belong. But that's not a problem if you're willing to wait for every answer. | |||||||||||||||||
| |||||||||||||||||