▲ | danenania 2 days ago | |
It can definitely be difficult and frustrating to try to use LLMs in a large codebase—no disagreement there. You have to be very selective about the tasks you give them and how they are framed. And yeah, you often need to throw away what they produced when they go in the wrong direction. None of that means they’re getting worse though. They’re getting better; they’re just not as good as you want them to be. | ||
▲ | taormina 2 days ago | parent [-] | |
I mean, this really isn't a large codebase, this is a small-medium sized codebase as judged by prior jobs/projects. 9000 lines of code? When I give them the same task I tried to give them the day before, and the output gets noticeably worse than their last model version, is that better? When the day by day performance feels like it's degrading? They are definitely not as good as I would like them to be but that's to be expected of professionals who beg for money hyping them up. |