| ▲ | westoncb 6 hours ago | |||||||||||||
Interesting that compaction is done using an encrypted message that "preserves the model's latent understanding of the original conversation": > Since then, the Responses API has evolved to support a special /responses/compact endpoint (opens in a new window) that performs compaction more efficiently. It returns a list of items (opens in a new window) that can be used in place of the previous input to continue the conversation while freeing up the context window. This list includes a special type=compaction item with an opaque encrypted_content item that preserves the model’s latent understanding of the original conversation. Now, Codex automatically uses this endpoint to compact the conversation when the auto_compact_limit (opens in a new window) is exceeded. | ||||||||||||||
| ▲ | icelancer 5 hours ago | parent | next [-] | |||||||||||||
Their compaction endpoint is far and away the best in the industry. Claude's has to be dead last. | ||||||||||||||
| ||||||||||||||
| ▲ | swalsh 4 hours ago | parent | prev | next [-] | |||||||||||||
Is it possible to use the compactor endpoint independently? I have my own agent loop i maintain for my domain specific use case. We built a compaction system, but I imagine this is better performance. | ||||||||||||||
| ||||||||||||||
| ▲ | jswny 4 hours ago | parent | prev [-] | |||||||||||||
How does this work for other models that aren’t OpenAI models | ||||||||||||||
| ||||||||||||||