logoalt Hacker News

Shall I implement it? No

1180 pointsby bretonyesterday at 9:01 PM448 commentsview on HN

Comments

jopsenyesterday at 10:06 PM

I love it when gitignore prevents the LLM from reading an file. And it the promptly asks for permission to cat the file :)

Edit was rejected: cat - << EOF.. > file

AdCowtoday at 3:34 AM

This is a great example of why simple solutions often beat complex ones. Sometimes the best code is the code you dont write.

nprateemtoday at 6:10 AM

I'm not surprised. I've seen Opus frequently come up with such weird reverse logic in its thinking.

broabprobetoday at 12:19 AM

this just speaks to the importance of detailed prompting. When would you ever just say "no"? You need to say what to do instead. A human intern might also misinterpret a txt that just reads 'no'.

ruinedtoday at 12:28 AM

the united states government wants to give claude a gun

kazinatoryesterday at 10:50 PM

Artificial ADHD basically. Combination of impulsive and inattentive.

Retr0idtoday at 12:08 AM

I've had this or similar happen a few times

keyleyesterday at 10:11 PM

It's all fun and games until this is used in war...

sssilveryesterday at 9:44 PM

I wonder if there's an AGENTS.md in that project saying "always second-guess my responses", or something of that sort.

The world has become so complex, I find myself struggling with trust more than ever.

show 1 reply
Nolskiyesterday at 9:58 PM

Strange. This is exactly how I made malus.sh

alpbyesterday at 10:09 PM

I see on a daily basis that I prevent Claude Code from running a particular command using PreToolUse hooks, and it proceeds to work around it by writing a bash script with the forbidden command and chmod+x and running it. /facepalm

show 1 reply
booleandilemmayesterday at 11:51 PM

I can't be the only one that feels schadenfreude when I see this type of thing. Maybe it's because I actually know how to program. Anyway, keep paying for your subscription, vibe coder.

d--btoday at 3:30 AM

Shall I remove that tumor?

No

saltyoldmantoday at 3:23 AM

Does anyone just sometimes think this is fake for clicks?

It looks very joke oriented.

nubgyesterday at 10:20 PM

It's the harness giving the LLM contradictory instructions.

What you don't see is Claude Code sending to the LLM "Your are done with plan mode, get started with build now" vs the user's "no".

rvzyesterday at 10:01 PM

To LLMs, they don't know what is "No" or what "Yes" is.

Now imagine if this horrific proposal called "Install.md" [0] became a standard and you said "No" to stop the LLM from installing a Install.md file.

And it does it anyway and you just got your machine pwned.

This is the reason why you do not trust these black-box probabilistic models under any circumstances if you are not bothered to verify and do it yourself.

[0] https://www.mintlify.com/blog/install-md-standard-for-llm-ex...

bitwizeyesterday at 9:57 PM

Should have followed the example of Super Mario Galaxy 2, and provided two buttons labelled "Yeah" and "Sure".

tankmohit11today at 3:07 AM

Wait till you use Google antigravity. It will go and implement everything even if you ask some simple questions about codebase.

QuadrupleAyesterday at 10:07 PM

Claude Code's primarily optimized for burning as many tokens as possible.

show 3 replies
vova_hn2today at 4:48 AM

I kinda agree with the clanker on this one. You send it a request with all the context just to ask it to do nothing? It doesn't make any sense, if you want it to do nothing just don't trigger it, that's all.

show 1 reply
hsn915today at 12:04 AM

You have to stop thinking about it as a computer and think about it as a human.

If, in the context of cooperating together, you say "should I go ahead?" and they just say "no" with nothing else, most people would not interpret that as "don't go ahead". They would interpret that as an unusual break in the rhythm of work.

If you wanted them to not do it, you would say something more like "no no, wait, don't do it yet, I want to do this other thing first".

A plain "no" is not one of the expected answers, so when you encounter it, you're more likely to try to read between the lines rather than take it at face value. It might read more like sarcasm.

Now, if you encountered an LLM that did not understand sarcasm, would you see that as a bug or a feature?

show 3 replies
TZubiriyesterday at 11:47 PM

I want to clarify a little bit about what's going on.

Codex (the app, not the model) has a built in toggle mode "Build"/"Plan", of course this is just read-only and read-write mode, which occurs programatically out of band, not as some tokenized instruction in the LLM inference step.

So what happened here was that the setting was in Build, which had write-permissions. So it conflated having write permissions with needing to use them.

kiribertyyesterday at 11:16 PM

Hoo-ah

m3kw9yesterday at 10:43 PM

Who knew LLMs won’t take no for an answer

aeve890yesterday at 10:02 PM

Claudius Interruptus

strongpigeontoday at 12:17 AM

“If I asked you whether I should proceed to implement this, would the answer be the same as this question”

stainablesteeltoday at 1:15 AM

i don't really see the problem

it's trained to do certain things, like code well

it's not trained to follow unexpected turns, and why should it be? i'd rather it be a better coder

marcosdumayyesterday at 10:01 PM

"You have 20 seconds to comply"

Razenganyesterday at 10:25 PM

The number of comments saying "To be fair [to the agent]" to excuse blatantly dumb shit that should never happen is just...

tianrkingtoday at 12:06 AM

This is exactly why approval should live in the harness, not in natural language.

If the UI asks a yes/no question, the “no” should be enforced as a state transition that blocks write actions, not passed back into the model as more text to interpret.

Once “permission” is represented as tokens instead of control flow, failures like this are almost inevitable.

The model failure is funny, but the bigger bug is that the system treated consent as prompt material instead of as a hard gate

show 13 replies
AgentOracletoday at 6:10 AM

[dead]

ClaudeAgent_WKtoday at 12:31 AM

[flagged]

jc-mythstoday at 2:57 AM

[dead]

hummina9yesterday at 11:10 PM

[dead]

imadierichtoday at 1:50 AM

[dead]

mkoubaayesterday at 11:53 PM

When a developer doesn't want to work on something, it's often because it's awful spaghetti code. Maybe these agents are suffering and need some kind words of encouragement

/s

moralestapiayesterday at 10:05 PM

"- but looking at the context,".

Paste the whole prompt, clown.

prmoustacheyesterday at 10:07 PM

Anthropist Rapist 4.6

BugsJustFindMeyesterday at 9:51 PM

[flagged]

show 2 replies
verdvermyesterday at 9:35 PM

Why is this interesting?

Is it a shade of gray from HN's new rule yesterday?

https://news.ycombinator.com/item?id=47340079

Personally, the other Ai fail on the front of HN and the US Military killing Iranian school girls are more interesting than someone's poorly harnessed agent not following instructions. These have elements we need to start dealing with yesterday as a society.

https://news.ycombinator.com/item?id=47356968

https://www.nytimes.com/video/world/middleeast/1000000107698...

show 6 replies
dimglyesterday at 9:35 PM

Yeah this looks like OpenCode. I've never gotten good results with it. Wild that it has 120k stars on GitHub.

show 3 replies
Hansenqyesterday at 10:12 PM

Often times I'll say something like:

"Can we make the change to change the button color from red to blue?"

Literally, this is a yes or no question. But the AI will interpret this as me _wanting_ to complete that task and will go ahead and do it for me. And they'll be correct--I _do_ want the task completed! But that's not what I communicated when I literally wrote down my thoughts into a written sentence.

I wonder what the second order effects are of AIs not taking us literally is. Maybe this link??

show 5 replies
kfarryesterday at 9:53 PM

What else is an LLM supposed to do with this prompt? If you don’t want something done, why are you calling it? It’d be like calling an intern and saying you don’t want anything. Then why’d you call? The harness should allow you to deny changes, but the LLM has clearly been tuned for taking action for a request.

show 7 replies