| ▲ | josteink an hour ago | |
> You are greatly underestimating the current hardware requirements for productive local LLMs. Fixed that for you. Right now most models produced are based on floating point maths and probabilities, which is "expensive" to do math on. Microsoft has researched 1-bit LLMs which can run much more efficiently, and on much cheaper hardware[1]. If this research is reproducable and reusable outside their research models, this means the cost of running self-hosted LLMs will be reduced by an order of magnitude once this hits mainstream. | ||