| ▲ | smithkl42 9 hours ago | |||||||||||||||||||||||||||||||
It's all about managing context. The bitter lesson applies over the long haul - and yes, over the long haul, as context windows get larger or go away entirely with different architectures, this sort of thing won't be needed. But we've defined enough skills in the last month or two that if we were to put them all in CLAUDE.md, we wouldn't have any context left for coding. I can only imagine that this will be a temporary standard, but given the current state of the art, it's a helpful one. | ||||||||||||||||||||||||||||||||
| ▲ | OtherShrezzing 8 hours ago | parent | next [-] | |||||||||||||||||||||||||||||||
I use Claude pretty extensively on a 2.5m loc codebase, and it's pretty decent at just reading the relevant readme docs & docstrings to figure out what's what. Those docs were written for human audiences years (sometimes decades) ago. I'm very curious to know the size & state of a codebase where skills are beneficial over just having good information hierarchy for your documentation. | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
| ▲ | iainmerrick 5 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||
To clarify, when I mentioned the bitter lesson I meant putting effort into organising the "skills" documentation in a very specific way (headlines, descriptions, etc). Splitting the docs into neat modules is a good idea (for both human readers and current AIs) and will continue to be a good idea for a while at least. Getting pedantic about filenames, documentation schemas and so on is just bikeshedding. | ||||||||||||||||||||||||||||||||
| ▲ | storus 8 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||
Why not replace the context tokens on the GPU during inference when they become no longer relevant? i.e. some tool reads a 50k token document, LLM processes it, so then just flush those document tokens out of active context, rebuild QKV caches and store just some log entry in the context as "I already did this ... with this result"? | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
| ▲ | ledauphin 9 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||
how is it different or better than maintaining an index page for your docs? Or a folder full of docs and giving Claude an instruction to `ls` the folder on startup? | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
| ▲ | stingraycharles 8 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||
Not sure why you’re being downvoted so much, it’s a valid point. It’s also related to attention — invoking a skill “now” means that the model has all the relevant information fresh in context, you’ll have much better results. What I’m doing myself is write skills that invoke Python scripts that “inject” prompts. This way you can set up multi-turn workflows for eg codebase analysis, deep thinking, root cause analysis, etc. Works very well. | ||||||||||||||||||||||||||||||||