| ▲ | NERD_ALERT 2 days ago |
| I felt this way with Github Copilot but I started using Cursor this week and it genuinely feels like a competent pair programmer. |
|
| ▲ | Retric 2 days ago | parent | next [-] |
| What work are you doing the last few days? My experience is for a very narrow range of tasks, like getting the basics of a common but new to me API working, they are moderately useful. But the overwhelming majority of the time they are useless. |
|
| ▲ | meander_water a day ago | parent | prev | next [-] |
| This has been my experience as well. Cursor Chat and autocomplete are near useless, and generate all sorts of errors, which on the whole cost more time. However, using composer, passing in the related files explicitly in the context, and prompting small changes incrementally has been a game changer for me. It also helps if you describe the intended behaviour in excruciating detail, including how you want all the edge cases/errors handled. |
|
| ▲ | jdcasale 2 days ago | parent | prev | next [-] |
| I recently tried Cursor for about a week and I was disappointed. It was useful for generating code that someone else has definitely written before (boilerplate etc), but any time I tried to do something nontrivial, it failed no matter how much poking, prodding, and thoughtful prompting I tried. Even when I tried to ask it for stuff like refactoring a relatively simple rust file to be more idiomatic or organized, it consistently generated code that did not compile and was unable to fix the compile errors on 5 or 6 repromptings. For what it's worth, a lot of SWE work technically trivial -- it makes this much quicker so there's obviously some value there, but if we're comparing it to a pair programmer, I would definitely fire a dev who had this sort of extremely limited complexity ceiling. It really feels to me (just vibes, obviously not scientific) like it is good at interpolating between things in its training set, but is not really able to do anything more than that. Presumably this will get better over time. |
| |
| ▲ | dughnut a day ago | parent [-] | | If you asked a junior developer to refactor a rust program to be more idiomatic, how long would you expect that to take? Would you expect the work to compile on the first try? I love Cline and Copilot. If you carefully specify your task, provide context for uncommon APIs, and keep the scope limited, then the results are often very good. It’s code completion for whole classes and methods or whole utility scripts for common use cases. Refactoring to taste may be under specified. | | |
| ▲ | jdcasale a day ago | parent | next [-] | | "If you asked a junior developer to refactor a rust program to be more idiomatic, how long would you expect that to take? Would you expect the work to compile on the first try?" The purpose of giving that task to a junior dev isn't to get the task done, it's to teach them -- I will almost always be at least an order order of magnitude faster than a junior for any given task. I don't expect juniors to be similarly productive to me, I expect them to learn. The parent comment also referred to a 'competent pair programmer', not a junior dev. My point was that for the tasks that I wanted to use the LLM, frequently there was no amount of specificity that could help the model solve it -- I tried for a long time, and generally if the task wasn't obvious to me, the model generally could not solve it. I'd end up in a game of trying to do nondeterministic/fuzzy programming in English instead of just writing some code to solve the problem. Again I agree that there is significant value here, because there is a ton of SWE work that is technically trivial, boring, and just eats up time. It's also super helpful as a natural-language info-lookup interface. | | |
| ▲ | dughnut 17 hours ago | parent [-] | | Personally, I think training someone on the client’s dime is pretty unethical. |
| |
| ▲ | Retric a day ago | parent | prev [-] | | What matters here is the communication overhead not how long between responses. If I’m indefinitely spending more time handholding a jr dev than they save me eventually I just fire em, same with code gen. | | |
| ▲ | djmips a day ago | parent [-] | | A big difference is that the jr. dev is learning compared to the AI who is stuck at whatever competence was baked in from the factory. You might be more patient with the jr if you saw positive signs that the handholding was paying off. | | |
| ▲ | Retric a day ago | parent | next [-] | | That was my point, though I may not have been clear. Most people do get better over time, but for those who don’t (or LLM’s) it’s just a question of if their current skills are a net benefit. I do expect future AI to improve. My expectation is it’s going to be a long slow slog just like with self driving cars etc, but novel approaches regularly turn extremely difficult problems into seemingly trivial exercises. | |
| ▲ | dughnut 17 hours ago | parent | prev [-] | | I would be more patient with an AI that only costs me a fraction of a cent an hour. |
|
|
|
|
|
| ▲ | 2 days ago | parent | prev [-] |
| [deleted] |