▲ | mirekrusin 5 days ago | |||||||
MoE is something different - it's a technique to activate just a small subset of parameters during inference. Whatever is good enough now, can be much better for the same cost (time, computation, actual cost). People will always choose better over worse. | ||||||||
▲ | mmmllm 5 days ago | parent [-] | |||||||
Thanks, I wasn't aware of that. Still - why isn't there a super expensive OpenAI model that uses 1,000 experts and comes up with way better answers? Technically that would be possible to build today. I imagine it just doesn't deliver dramatically better results. | ||||||||
|