| ▲ | danielhanchen 8 hours ago | |||||||
For those interested, made some MXFP4 GGUFs at https://huggingface.co/unsloth/Qwen3.5-397B-A17B-GGUF and a guide to run them: https://unsloth.ai/docs/models/qwen3.5 | ||||||||
| ▲ | plagiarist 3 hours ago | parent [-] | |||||||
Are smaller 2/3-bit quantizations worth running vs. a more modest model at 8- or 16-bit? I don't currently have the vRAM to match my interest in this | ||||||||
| ||||||||