▲ | wongarsu a day ago | |
The answer is still terrible for the model size. Maybe it's the 4 bit quantization, smaller models tend to react worse to that For reference, [1] is what stock quen3-0.6B would answer. Not a perfect answer, but much better at nearly half the number of parameters 1: https://markdownpastebin.com/?id=7ad4ad9f325d4354a858480abdc... | ||
▲ | KTibow a day ago | parent [-] | |
It's likely the quantization on "Llama-3.2-1B-Instruct-q4f16_1-MLC". inference.net generated this more coherent answer: https://hst.sh/ovilewofox.md |