|
 | ▲ | embedding-shape 3 days ago | parent | next [-] | 
 | > gpt-oss-120b is amazing Yup, I agree, easily best local model you can run today on local hardware, especially when reasoning_effort is set to "high", but "medium" does very well too. I think people missed out on how great it was because a bunch of the runners botched their implementations at launch, and it wasn't until 2-3 weeks after launch that you could properly evaluate it, and once I could run the evaluations myself on my own tasks, it really became evident how much better it is. If you haven't tried it yet, or you tried it very early after the release, do yourself a favor and try it again with updated runners.  | 
|
 | ▲ | whatreason 3 days ago | parent | prev | next [-] | 
 | What do you use to run gpt-oss here? ollama, vLLM, etc  | 
 | |
  | ▲ | embedding-shape 3 days ago | parent [-] |   | Not parent, but frequent user of GPT-OSS, tried all different ways of running it. Choice goes something like this: - Need batching + highest total throughoutput? vLLM, complicated to deploy and install though, need special versions for top performance with GPT-OSS - Easiest to manage + fast enough: llama.cpp, easier to deploy as well (just a binary) and super fast, getting ~260 tok/s on a RTX Pro 6000 for the 20B version - Easiest for people not used to running shell commands or need a GUI and don't care much for performance: Ollama Then if you really wanna go fast, try to get TensorRT running on your setup, and I think that's pretty much the fastest GPT-OSS can go currently.  |  
  | 
|
 | ▲ | giorgioz 3 days ago | parent | prev | next [-] | 
 | on what hardware you manate to run gpt-oss-120b locally?  | 
|
 | ▲ |  3 days ago | parent | prev | next [-] | 
 | [deleted] | 
|
 | ▲ | rovr138 3 days ago | parent | prev | next [-] | 
 | >  I created a RAG agent to hold most of GCP documentation (separate download, parsing, chunking, etc) If you share the scripts to gather the GCP documentation this, that'd be great. Because I have had an idea to do something like this, and the part I don't want to deal with is getting the data  | 
 | |
  | ▲ | fm2606 2 days ago | parent [-] |   | I tried scripts but got blocked.  I used wget to download tthem  |  
  | 
|
 | ▲ | lacoolj 3 days ago | parent | prev | next [-] | 
 | you can run the 120b model on an 8GB GPU? or are you running this on CPU with the 64GB RAM? I'm about to try this out lol The 20b model is not great, so I'm hoping 120b is the golden ticket.  | 
 | |
  | ▲ | ThatPlayer 3 days ago | parent | next [-] |   | With MoE models like gpt-oss, you can run some layers on the CPU (and some on GPU): https://github.com/ggml-org/llama.cpp/discussions/15396 Mentions 120b is runnable on 8GB VRAM too: "Note that even with just 8GB of VRAM, we can adjust the CPU layers so that we can run the large 120B model too"  |  |
  | ▲ | gunalx 3 days ago | parent | prev | next [-] |   | I have in many cases had better results with the 20b model, over the 120b model.  
Mostly because it is faster and I can iterate prompts quicker to choerce it to follow instructions.  |   | |
  | ▲ | embedding-shape 3 days ago | parent [-] |   | > had better results with the 20b model, over the 120b model The difference of quality and accuracy of the responses between the two is vastly different though, if tok/s isn't your biggest priority, especially when using reasoning_effort "high". 20B works great for small-ish text summarization and title generation, but for even moderately difficult programming tasks, 20B fails repeatedly while 120B gets it right on the first try.  |   | |
  | ▲ | gunalx 11 hours ago | parent [-] |   | But the 120b model has just as bad if not worse formatting issues, compared to the 20b one. For simple refactorings, or chatting about possible solutions i actually feel teh 20b halucinates less than the 120b, even if it is less competent.  Migth also be because of 120b not liking being in q8, or not being properly deployed.  |  
  |  
  |  |
  | ▲ | fm2606 3 days ago | parent | prev | next [-] |   | Everything I run, even the small models, some amount goes to the GPU and the rest to RAM.  |  |
  | ▲ | fm2606 3 days ago | parent | prev [-] |   | Hmmm...now that you say that, it might have been the 20b model. And like a dumbass I accidentally deleted the directory and didn't have a back up or under version control. Either way, I do know for a fact that the gpt-oss-XXb model beat chatgpt by 1 answer and it was 46/50 at 6 minutes and 47/50 at 1+ hour.  I remember because I was blown away that I could get that type of result running locally and I had texted a friend about it. I was really impressed but disappointed at the huge disparity between time the two.  |  
  | 
|
 | ▲ | gkfasdfasdf 2 days ago | parent | prev | next [-] | 
 | What were you using for RAG? Did you build your own or some off the shelf solution (e.g. openwebui)  | 
 | |
  | ▲ | fm2606 2 days ago | parent [-] |   | I used pg vector chunking on paragraphs.  For the answers I saved in a flat text file and then parsed to what I needed. For parsing and vectorizing of the GCP docs I used a Python script.  For reading each quiz question, getting a text embedding and submitting to an LLM, I used Spring AI. It was all roll your own. But like I stated in my original post I deleted it without backup or vcs.  It was the wrong directory that I deleted.  Rookie mistake for which I know better.  |  
  | 
|
 | ▲ | adastra22 3 days ago | parent | prev [-] | 
 | What quantization settings?  |