| ▲ | armcat 6 hours ago | |||||||
This is beautifully written and visualised, well done! The KL divergence comparisons between original and different quantisation levels is on-point. I'm not sure people realize how powerful quantisation methods are and what they've done for democratising local AI. And there are some great players out there like Unsloth and Pruna. | ||||||||
| ▲ | samwho 5 hours ago | parent [-] | |||||||
Thank you! I was really surprised how robust models are to losing information. It seems wrong that they can be compressed so much and still function at all, never mind function quite closely to the original size. Think we're only going to keep seeing more progress in this area on the research side, too. | ||||||||
| ||||||||