| ▲ | gunalx 3 days ago | |||||||
I have in many cases had better results with the 20b model, over the 120b model. Mostly because it is faster and I can iterate prompts quicker to choerce it to follow instructions.  | ||||||||
| ▲ | embedding-shape 3 days ago | parent [-] | |||||||
> had better results with the 20b model, over the 120b model The difference of quality and accuracy of the responses between the two is vastly different though, if tok/s isn't your biggest priority, especially when using reasoning_effort "high". 20B works great for small-ish text summarization and title generation, but for even moderately difficult programming tasks, 20B fails repeatedly while 120B gets it right on the first try.  | ||||||||
  | ||||||||