| ▲ | vidarh 14 hours ago | |||||||||||||||||||||||||
I think the human + agent thing absolutely will make a huge difference. I see regularly that Claude can totally off piste and eventually claw itself back with a proper agent setup but it will take a lot of time if I don't spot it and get it back on track. I have one project Claude is working on right now where I'm testing a setup to attempt to take myself more out of the loop, because that is the hard part. It's "easy" to get an agent to multiply your output. It's hard to make that scale with your willingness to spend on tokens rather than with your ability to read and review and direct. I've ended up with roughly this (it's nothing particularly special): - Runs a evaluator that evaluates the current state and assigns scores across multiple metrics. - If a given score is above a given threshold, expand the test suite automatically. - If the score is below a given threshold, spawn a "research agent" that investgates why the scores don't meet expectations. - The research agent delivers a report, that is passed to an implementation agent. - The main agent re-runs the scoring, and if it doesn't show an improvement on one or more of the metrics, the commit is discarded, and notes made of what was tried, and why it failed. It takes a bit of trial and error to get it right (e.g. "it's the test suite that is wrong" came up early, and the main agent was almost talked into revising the test suite to remove the "problematic" tests) but a division sort of like this lets Claude do more sensible stuff for me. Throwing away commits feels drastic - an option is to let it run a little cycle of commit -> evaluate -> redo a few times before the final judgement, maybe - but it so far it feels like it'll scale better. Less crap makes it into the project. And I think this will work better than to treat these agents as if they are developers whose output costs 100x as much. Code so cheap it is disposable should change the workflows. So while I agree this is a better demonstration of a good way to build a browser, it's a less interesting demonstration as well. Now that we've seen people show that something like FastRender is possible, expect people to experiment with similarly ambitious projects but with more thought put into scoring/evaluation, including on code size and dependencies. | ||||||||||||||||||||||||||
| ▲ | embedding-shape 14 hours ago | parent [-] | |||||||||||||||||||||||||
> I think the human + agent thing absolutely will make a huge difference. Just the day(s) before, I was thinking about this too, and I think what will make the biggest difference is humans who posses "Good Taste". I wrote a bunch about it here: https://emsh.cat/good-taste/ I think the ending is most apt, and where I think we're going wrong right now: > I feel like we're building the wrong things. The whole vibe right now is "replace the human part" instead of "make better tools for the human part". I don't want a machine that replaces my taste, I want tools that help me use my taste better; see the cut faster, compare directions, compare architectural choices, find where I've missed things, catch when we're going into generics, and help me make sharper intentional choices. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||