| ▲ | MallocVoidstar 4 hours ago | |
> Even when the model is explicitly instructed to pause due to insufficient tokens rather than generating an incomplete response AI models can't do this. At least not with just an instruction, maybe if you're writing some kind of custom 'agentic' setup. | ||
| ▲ | maxloh 3 hours ago | parent [-] | |
Yeah, it does. It was possible with 2.5 Flash. Here's a similar result with Qwen Qwen3.5-397B-A17B: https://chat.qwen.ai/s/530becb7-e16b-41ee-8621-af83994599ce?... | ||