| ▲ | Incipient 2 days ago |
| Was more of a general comment - I'm surprised there is significant variation between any of the frontier models? However, vscode with various python frameworks/libraries; dash, fastapi, pandas, etc. Typically passing the 4-5 relevant files in as context. Developing via docker so I haven't found a nice way for agents to work. |
|
| ▲ | danielbln 2 days ago | parent | next [-] |
| I would suggest using an agentic system like Cline, so that the LLM can wander through the codebase by itself and do research and build a "mental model" and then set up an implementation plan. The you iterate in that and hand it off for implementation. This flow works significantly better than what you're describing. |
| |
| ▲ | otabdeveloper4 2 days ago | parent [-] | | > LLM can wander through the codebase by itself and do research and build a "mental model" It can't really do that due to context length limitations. | | |
| ▲ | johnisgood 2 days ago | parent | next [-] | | 1k LOC is perfectly fine, I did not experience issues with Claude with most (not all) projects around ~1k LOC. | | |
| ▲ | otabdeveloper4 18 hours ago | parent [-] | | Actual projects where you'd want some LLM help start with millions of lines of code, not thousands. With 1k lines of code you don't need an LLM, the entire source code can fit in one intern's head. | | |
| ▲ | simonw 9 hours ago | parent | next [-] | | Have you tried Claude Code yet? Even with it's 200,000 token limit it's still really impressive at diving through large codebases using find and grep. | |
| ▲ | johnisgood 16 hours ago | parent | prev [-] | | The OP mentioned having LLM issues with 1k LOC, so I suppose he would have problems with millions. :D |
|
| |
| ▲ | lukan 2 days ago | parent | prev | next [-] | | I guess people are talking about different kinds of projects here in terms of project size. | |
| ▲ | exe34 2 days ago | parent | prev [-] | | It doesn't need the entire codebase, it just needs the call map, the function signatures, etc. It doesn't have to include everything in a call - but having access to all of it means it can pick what seems relevant. | | |
| ▲ | danielbln 2 days ago | parent | next [-] | | Yes, that's exactly right. The LLM gets a rough overview over the project (as you said, including function signatures and such) and will then decide what to open and use to complete/implement the objective. | |
| ▲ | otabdeveloper4 18 hours ago | parent | prev [-] | | In a real project the call map and function signatures are millions of tokens themselves. | | |
|
|
|
|
| ▲ | fragmede a day ago | parent | prev [-] |
| > I'm surprised there is significant variation between any of the frontier models? This comment of mine is a bit dated, but even the same model can have significant variation if you change the prompt by just a few words. https://news.ycombinator.com/item?id=42506554 |