| ▲ | vsrinivas 9 hours ago | |
Hey - I use the same, w/ both gemma4 and gpt-oss-*; some things I have to do for a good experience: 1) Pin to an earlier version of codex (sorry) - 0.55 is the best experience IME, but YMMV (see https://github.com/openai/codex/issues/11940, https://github.com/openai/codex/issues/8272). 2) Use the older completions endpoint (llama.cpp's responses support is incomplete - https://github.com/ggml-org/llama.cpp/issues/19138) | ||