| ▲ | Aurornis 12 hours ago | ||||||||||||||||
> gpt-oss-120b full quant runs on my quad 3090 A 120B model cannot fit on 4 x 24GB GPUs at full quantization. Either you're confusing this with the 20B model, or you have 48GB modded 3090s. | |||||||||||||||||
| ▲ | segmondy an hour ago | parent [-] | ||||||||||||||||
Some of you folks on here love to argue, gpt-oss-120b was trained in 4 bits, so it pretty much takes up 60gb. | |||||||||||||||||
| |||||||||||||||||