▲ | flashgordon 5 days ago | ||||||||||||||||||||||||||||||||||||||||||||||
Yeah I was actually thinking about a proper rig - My gut feel is a rig wouldnt be as expensve as a mac and would actually have a higher ROI (at the expense of portability)? My other worry about the mac is how unupgradable it is. Again not sure how fruitful it is - in my (probably fantasy land) view if I can setup a rig and then keep updating components as needed - it might last me a good 5 years say for 20k over that period? Or is that too hopeful? So for 20K over 5 years or 4k per year - it comes to about 400 a month (ish). The equivalent of 2 MAX pro subscriptions. Let us be honest - right now with these limits running more than 1 in parallel is going to be forbidden. if I can run 2 claude level models (assuming the DS and Qwens are there) then I am already breaking even but without having to participating in training with all my codebases (and I assume I can actually unlock something new in the process of being free). | |||||||||||||||||||||||||||||||||||||||||||||||
▲ | lossolo 5 days ago | parent [-] | ||||||||||||||||||||||||||||||||||||||||||||||
Buy 4–8 used 3090s (providing 96–192 GB of VRAM), depending on the model and weight quantization you want to run. Used 3090 costs around $800. Add more RAM to offload layers if needed. This setup currently offers the best value for performance. https://www.reddit.com/r/LocalLLaMA/comments/1iqpzpk/8x_rtx_... You can look for more rig examples on that subreddit. | |||||||||||||||||||||||||||||||||||||||||||||||
|