| ▲ | hu3 2 hours ago | |
Check context size. LLMs become increasingly error-prone as their memory is fills up. Just like humans. In VSCode Copilot you can keep track of how many tokens the LLM is dealing with in realtime with "Chat Debug". When it reaches 90k tokens I should expect degraded intelligence and brace for a possible forced sumarization. Sometimes I just stop LLMs and continue the work in a new session. | ||