Remix.run Logo
ge96 5 hours ago

Nice, I recently pulled down TheBloke 7B mistral to try out I have a 4070.

bashbjorn 5 hours ago | parent | next [-]

I love mistral, but that model is... not the best. Maybe try out Gemma 4 e4b, it's a similar size to Mistral 7B, and should run great on your 4070 ("E4B" is slightly misleading naming).

ge96 5 hours ago | parent [-]

Thanks for the tip, what do you use Gemma 4 e4b for?

redanddead 5 hours ago | parent [-]

some say it’s a miniaturized gemini model

it’s good at writing, coding, decently intelligent

you can try it on nvidia nim

mixtureoftakes 4 hours ago | parent | prev | next [-]

7b mistral is quite outdated. On a 12gb 4070 you can run qwen 3.5 9b q4km or qwen 3.6 35b, the latter will be a lot smarter but also a lot slower due to ram offload.

Try both in lm studio, they really are surprisingly capable

ge96 4 hours ago | parent [-]

I have 80gb of ram but it's slow capped by i9 CPU or specific asus mobo sucks I think only 2400mhz despite being ddr4

Tried all the stuff bios, volting

ganelonhb 5 hours ago | parent | prev [-]

I have a 2070 and can confirm it works amazingly fast.

I love TheBloke I wish he still made stuff

bashbjorn 5 hours ago | parent | next [-]

Yeah, TheBloke era of local LLMs were good times. TBF Unsloth are doing a fantastic job of publishing quants of the major models quickly - they just don't have nearly the volume of "weird" models as TheBloke did.

ge96 5 hours ago | parent | prev [-]

What do you use it for? I'm still trying to use agents, I barely use copilot, only at work when I have to.

I didn't want to get personal with an LLM unless it was local so that's why I was setting this up but yeah. So far just research is what I was looking at.