| ▲ | ge96 5 hours ago |
| Nice, I recently pulled down TheBloke 7B mistral to try out I have a 4070. |
|
| ▲ | bashbjorn 5 hours ago | parent | next [-] |
| I love mistral, but that model is... not the best. Maybe try out Gemma 4 e4b, it's a similar size to Mistral 7B, and should run great on your 4070 ("E4B" is slightly misleading naming). |
| |
| ▲ | ge96 5 hours ago | parent [-] | | Thanks for the tip, what do you use Gemma 4 e4b for? | | |
| ▲ | redanddead 5 hours ago | parent [-] | | some say it’s a miniaturized gemini model it’s good at writing, coding, decently intelligent you can try it on nvidia nim |
|
|
|
| ▲ | mixtureoftakes 4 hours ago | parent | prev | next [-] |
| 7b mistral is quite outdated. On a 12gb 4070 you can run qwen 3.5 9b q4km or qwen 3.6 35b, the latter will be a lot smarter but also a lot slower due to ram offload. Try both in lm studio, they really are surprisingly capable |
| |
| ▲ | ge96 4 hours ago | parent [-] | | I have 80gb of ram but it's slow capped by i9 CPU or specific asus mobo sucks I think only 2400mhz despite being ddr4 Tried all the stuff bios, volting |
|
|
| ▲ | ganelonhb 5 hours ago | parent | prev [-] |
| I have a 2070 and can confirm it works amazingly fast. I love TheBloke I wish he still made stuff |
| |
| ▲ | bashbjorn 5 hours ago | parent | next [-] | | Yeah, TheBloke era of local LLMs were good times. TBF Unsloth are doing a fantastic job of publishing quants of the major models quickly - they just don't have nearly the volume of "weird" models as TheBloke did. | |
| ▲ | ge96 5 hours ago | parent | prev [-] | | What do you use it for? I'm still trying to use agents, I barely use copilot, only at work when I have to. I didn't want to get personal with an LLM unless it was local so that's why I was setting this up but yeah. So far just research is what I was looking at. |
|