Remix.run Logo
Kim_Bruning a day ago

> Isn’t “instruction following” the most important thing you’d want out of a model in general,

No. And for the same reason that pure "instruction following" in humans is considered a form of protest/sabotage.

https://en.wikipedia.org/wiki/Work-to-rule

stingraycharles a day ago | parent | next [-]

I don’t understand the point you’re trying to make. LLMs are not humans.

From my perspective, the whole problem with LLMs (at least for writing code) is that it shouldn’t assume anything, follow the instructions faithfully, and ask the user for clarification if there is ambiguity in the request.

I find it extremely annoying when the model pushes back / disagrees, instead of asking for clarification. For this reason, I’m not a big fan of Sonnet 4.5.

IgorPartola a day ago | parent | next [-]

Full instruction following looks like monkey’s paw/malicious compliance. A good way to eliminate a bug from a codebase is to delete the codebase, that type of thing. You want the model to have enough creative freedom to solve the problem otherwise you are just coding using an imprecise language spec.

I know what you mean: a lot of my prompts include “never use em-dashes” but all models forget this sooner or later. But in other circumstances I do want it to push back on something I am asking. “I can implement what you are asking but I just want to confirm that you are ok with this feature introducing an SQL injection attack into this API endpoint”

stingraycharles a day ago | parent [-]

My point is that it’s better that the model asks questions to better understand what’s going on before pushing back.

IgorPartola a day ago | parent [-]

Agreed. With Claude Code I will often specify the feature I want to develop, then tell it to summarize the plan for me, give me its opinion on the plan, and ask questions before it does anything. This works very well. Often times it actually catches some piece I didn’t consider and this almost always results in usable code or code that is close enough that Claude can fix after I review what it did and point out problems.

Kim_Bruning a day ago | parent | prev | next [-]

I can't help you then. You can find a close analogue in the OSS/CIA Simple Sabotage Field Manual. [1]

For that reason, I don't trust Agents (human or ai, secret or overt :-P) who don't push back.

[1] https://www.cia.gov/static/5c875f3ec660e092cf893f60b4a288df/... esp. Section 5(11)(b)(14): "Apply all regulations to the last letter." - [as a form of sabotage]

stingraycharles a day ago | parent [-]

How is asking for clarification before pushing back a bad thing?

Kim_Bruning a day ago | parent [-]

Sounds like we're not too far apart then!

Sometimes pushback is appropriate, sometimes clarification. The key thing is that one doesn't just blindly follow instructions; at least that's the thrust of it.

InsideOutSanta a day ago | parent | prev | next [-]

I would assume that if the model made no assumptions, it would be unable to complete most requests given in natural language.

stingraycharles a day ago | parent [-]

Well yes, but asking the model to ask questions to resolve ambiguities is critical if you want to have any success in eg a coding assistant.

There are shitloads of ambiguities. Most of the problems people have with LLMs is the implicit assumptions being made.

Phrased differently, telling the model to ask questions before responding to resolve ambiguities is an extremely easy way to get a lot more success.

simlevesque a day ago | parent | prev | next [-]

I think the opposite. I don't want to write down everything and I like when my agents take some initiative or come up with solutions I didn't think of.

wat10000 a day ago | parent | prev | next [-]

If I tell it to fetch the information using HTPP, I want it to ask if I meant HTTP, not go off and try to find a way to fetch the info using an old printing protocol from IBM.

MangoToupe a day ago | parent | prev | next [-]

> and ask the user for clarification if there is ambiguity in the request.

You'd just be endlessly talking to the chatbots. Humans are really bad at expressing ourselves precisely, which is why we have formal languages that preclude ambiguity.

scotty79 a day ago | parent | prev [-]

> is that it shouldn’t assume anything, follow the instructions faithfully, and ask the user for clarification if there is ambiguity in the request

We already had those. They are called programming languages. And interacting with them used to be a very well paid job.

SkyeCA a day ago | parent | prev [-]

It's still insanity to me that doing your job exactly as defined and not giving away extra work is considered a form of action.

Everyone should be working-to-rule all the time.