| ▲ | wild_egg 2 hours ago | |||||||
Don't have a GPU so tried the CPU option and got 0.6t/s on my old 2018 laptop using their llama.cpp fork. Then found out they didn't implement AVX2 for their Q1_0_g128 CPU kernel. Added that and getting ~12t/s which isn't shabby for this old machine. Cool model. | ||||||||
| ▲ | UncleOxidant 42 minutes ago | parent | next [-] | |||||||
Are you getting anything besides gibberish out of it? I tried their recommended commandline and it's dog slow even though I built their llama.cpp fork with AVX2 enabled. This is what I get: | ||||||||
| ▲ | cubefox an hour ago | parent | prev | next [-] | |||||||
"Not shabby" is a big understatement. | ||||||||
| ||||||||
| ▲ | 2 hours ago | parent | prev [-] | |||||||
| [deleted] | ||||||||
| ||||||||