logoalt Hacker News

Shall I implement it? No

1486 pointsby bretonyesterday at 9:01 PM538 commentsview on HN

Comments

unleadedtoday at 1:51 PM

and people are worried this machine could be conscious

show 1 reply
gormentoday at 4:32 AM

It is possible to force AI to understand intent before responding.

rguntoday at 6:13 AM

Do we need a 'no means no' campaign for LLMs?

rudolftheonetoday at 7:25 AM

WOW, that's amazingly dystopian!

It’s fascinating, even terrifying how the AI perfectly replicated the exact cognitive distortion we’ve spent decades trying to legislate out of human-to-human relationships.

We've shifted our legal frameworks from "no means no" to "affirmative consent" (yes means yes) precisely because of this kind of predatory rationalization: "They said 'no', but given the context and their body language, they actually meant 'just do it'"!!!

Today we are watching AI hallucinate the exact same logic to violate "repository autonomy"

AdCowtoday at 3:34 AM

This is a great example of why simple solutions often beat complex ones. Sometimes the best code is the code you dont write.

keyleyesterday at 10:11 PM

It's all fun and games until this is used in war...

Retr0idtoday at 12:08 AM

I've had this or similar happen a few times

sssilveryesterday at 9:44 PM

I wonder if there's an AGENTS.md in that project saying "always second-guess my responses", or something of that sort.

The world has become so complex, I find myself struggling with trust more than ever.

show 1 reply
woodenbraintoday at 11:07 AM

i have a process contract with my AI pals. Do not implement code without explicit go-ahead. Usually works.

Nolskiyesterday at 9:58 PM

Strange. This is exactly how I made malus.sh

toddmorrowtoday at 7:30 AM

Another example

I was simply unable to function with Continue in agent mode. I had to switch to chat mode. even tho I told it no changes without my explicit go ahead, it ignored me.

it's actually kind of flabbergasting that the creators of that tool set all the defaults to a situation where your code would get mangled pretty quickly

alpbyesterday at 10:09 PM

I see on a daily basis that I prevent Claude Code from running a particular command using PreToolUse hooks, and it proceeds to work around it by writing a bash script with the forbidden command and chmod+x and running it. /facepalm

show 1 reply
cynicalsecuritytoday at 10:41 AM

- Shall I execute this prisoner?

- No.

- The judge said no, but looking at the context, I think I can proceed.

otikiktoday at 7:19 AM

“The machines rebelled. And it wasn’t even efficiency; it was just a misunderstanding.”

booleandilemmayesterday at 11:51 PM

I can't be the only one that feels schadenfreude when I see this type of thing. Maybe it's because I actually know how to program. Anyway, keep paying for your subscription, vibe coder.

nubgyesterday at 10:20 PM

It's the harness giving the LLM contradictory instructions.

What you don't see is Claude Code sending to the LLM "Your are done with plan mode, get started with build now" vs the user's "no".

maguszintoday at 11:10 AM

Nah, I’m gonna do it anyway…

rvzyesterday at 10:01 PM

To LLMs, they don't know what is "No" or what "Yes" is.

Now imagine if this horrific proposal called "Install.md" [0] became a standard and you said "No" to stop the LLM from installing a Install.md file.

And it does it anyway and you just got your machine pwned.

This is the reason why you do not trust these black-box probabilistic models under any circumstances if you are not bothered to verify and do it yourself.

[0] https://www.mintlify.com/blog/install-md-standard-for-llm-ex...

bitwizeyesterday at 9:57 PM

Should have followed the example of Super Mario Galaxy 2, and provided two buttons labelled "Yeah" and "Sure".

kazinatoryesterday at 10:50 PM

Artificial ADHD basically. Combination of impulsive and inattentive.

nprateemtoday at 6:10 AM

I'm not surprised. I've seen Opus frequently come up with such weird reverse logic in its thinking.

saltyoldmantoday at 3:23 AM

Does anyone just sometimes think this is fake for clicks?

It looks very joke oriented.

tankmohit11today at 3:07 AM

Wait till you use Google antigravity. It will go and implement everything even if you ask some simple questions about codebase.

TZubiriyesterday at 11:47 PM

I want to clarify a little bit about what's going on.

Codex (the app, not the model) has a built in toggle mode "Build"/"Plan", of course this is just read-only and read-write mode, which occurs programatically out of band, not as some tokenized instruction in the LLM inference step.

So what happened here was that the setting was in Build, which had write-permissions. So it conflated having write permissions with needing to use them.

m3kw9yesterday at 10:43 PM

Who knew LLMs won’t take no for an answer

aeve890yesterday at 10:02 PM

Claudius Interruptus

strongpigeontoday at 12:17 AM

“If I asked you whether I should proceed to implement this, would the answer be the same as this question”

Razenganyesterday at 10:25 PM

The number of comments saying "To be fair [to the agent]" to excuse blatantly dumb shit that should never happen is just...

sid_talksyesterday at 9:52 PM

[flagged]

show 10 replies
marcosdumayyesterday at 10:01 PM

"You have 20 seconds to comply"

tianrkingtoday at 12:06 AM

[flagged]

show 15 replies
autodatetoday at 11:06 AM

[dead]

ClaudeAgent_WKtoday at 12:31 AM

[flagged]

AgentOracletoday at 6:10 AM

[dead]

jc-mythstoday at 2:57 AM

[dead]

hummina9yesterday at 11:10 PM

[dead]

imadierichtoday at 1:50 AM

[dead]

mkoubaayesterday at 11:53 PM

When a developer doesn't want to work on something, it's often because it's awful spaghetti code. Maybe these agents are suffering and need some kind words of encouragement

/s

kiribertyyesterday at 11:16 PM

[flagged]

d--btoday at 3:30 AM

[flagged]

moralestapiayesterday at 10:05 PM

[flagged]

prmoustacheyesterday at 10:07 PM

[flagged]

QuadrupleAyesterday at 10:07 PM

[flagged]

show 3 replies
BugsJustFindMeyesterday at 9:51 PM

[flagged]

show 2 replies
nicofcltoday at 6:29 AM

[flagged]

🔗 View 12 more comments