| ▲ | jayd16 4 hours ago | |||||||
> Even when the model is explicitly instructed to pause due to insufficient tokens Is there actually a chance it has the introspection to do anything with this request? | ||||||||
| ▲ | maxloh 3 hours ago | parent | next [-] | |||||||
Yeah, it does. It was possible with 2.5 Flash. Here's a similar result with Qwen Qwen3.5-397B-A17B: https://chat.qwen.ai/s/530becb7-e16b-41ee-8621-af83994599ce?... | ||||||||
| ||||||||
| ▲ | verdverm 3 hours ago | parent | prev | next [-] | |||||||
No, the model doesn't have purview into this afaik I'm not even sure what "pausing" means in this context and why it would help when there are insufficient tokens. They should just stop when you reach the limit, default or manually specified, but it's typically a cutoff. You can see what happens by setting output token limit much lower | ||||||||
| ▲ | otabdeveloper4 3 hours ago | parent | prev [-] | |||||||
No. | ||||||||