|
| ▲ | stingraycharles an hour ago | parent | next [-] |
| But these docs are the notes, it constantly needs to be re-primed with them, an approach which doesn’t scale. How much of this knowledge can you really put in these agent docs? There's only so much you can do, and for any serious-scale projects, there's SO much knowledge that needs to be captured. Not just "do this, do that", but also context about why certain decisions were made (rationale, business context, etc). It is exactly akin to a human that has to write down everything on notes, and re-read them every time. |
|
| ▲ | falcor84 8 hours ago | parent | prev | next [-] |
| But that's the thing: Claude Plays Pokemon is an experiment in having Claude work fully independently, so there's no "you" who would improve its onboarding docs or anything else, it has to do so on its own. And as long as it cannot do so reliably, it effectively has anterograde amnesia. And just to be clear, I'm mentioning this because I think that Claude Plays Pokemon is a playground for any agentic AI doing any sort of long-term independent work; I believe that the solution needed here is going to bring us closer to a fully independent agent in coding and other domains. It reminds me of the codeclash.ai benchmark, where similar issues are seen across multiple "rounds" of an AI working on the same codebase. |
| |
| ▲ | skybrian 6 hours ago | parent [-] | | Sure, it's not close to fully independent. But I was interpreting "much, much less employable" as not very useful for programming in its current state, and I think it is quite useful. |
|
|
| ▲ | BoredPositron 20 minutes ago | parent | prev | next [-] |
| It does not learn. It has to look at the files over and over again. If or the agent forgets to look at them it has 0 knowledge about what it did. |
|
| ▲ | kaashif 8 hours ago | parent | prev [-] |
| Yeah but it feels terrible. I put as much as I can into Claude skills and CLAUDE.md but the fact that this is something I even have to think about makes me sad. The discrete points where the context gets compacted really feel bad and not like how I think AGI or whatever should work. Just continuously learn and have a super duper massive memory. Maybe I just need a bazillion GPUs to myself to get that. But no-one wants to manage context all the time, it's incidental complexity. |
| |
| ▲ | falcor84 8 hours ago | parent | next [-] | | I agree with essentially everything you said, except for the final claim that managing context is incidental complexity. From what I know of cognitive science, I would argue that context management is a central facet of intelligence, and a lot of the success of humans in society is dependent on their ability to do so. Looking at it from the other side, executive function disorders such as ADHD offer significant challenges for many humans, and they seem to be not quite entirely unlike these context issues that Claude faces. | |
| ▲ | onion2k 4 hours ago | parent | prev [-] | | no-one wants to manage context all the time Maybe we'll start needing to have daily stand-ups with our coding agents. |
|