I used the word "context" in a purely technical sense in relation to LLMs: the input tokens that you send to an LLM.
Every time you send what appears as a "chat message" in any of the programs that let you "chat" with an "AI", what you really do is sending the whole conversation history (all previous messages, tool calls and responses) as an input and asking model to generate an output.
There is no conceivable scenario when sending "<tons of tokens> + no" makes any sense.
Best case scenario is:
"<tons of tokens> + no" -> "Okay, I won't do it."
In this case you've just waisted a lot of input tokens, that someone (hopefully, not you) has to pay for, to generate an absolutely pointless message that says "Okay, I won't do it.". There is no value in this message. There is bo reason to waste time and computational resources to generate this message.
Worst case scenario is what happened on the screenshot.
There is no good scenario when this input produces a valuable output.
If you want your "agent" or "model" or whatever to do nothing you just don't trigger it. It won't do anything on it's own, it doesn't wait for your response, it doesn't need your response.
I don't understand why, in this thread, every time I try to point out how nonsensical is the behavior that they want is from technical perspective (from the perspective of knowing how these tools actually work) people just cling to there anthropomorphized mind model of the LLM and insist on getting angry.
"It acts like a bad human being, therefore it's bad, useless and dangerous"
I don't even know what to say to this.
P. S. If you wind this message hard to read and understand, I'm sorry about it, I don't know how to word it better. HN disallows to use LLMs to edit comments, but I think that sending a link to an LLM-edited version of the comment should be ok:
I used the word "context" in a purely technical sense in relation to LLMs: the input tokens that you send to an LLM.
Every time you send what appears as a "chat message" in any of the programs that let you "chat" with an "AI", what you really do is sending the whole conversation history (all previous messages, tool calls and responses) as an input and asking model to generate an output.
There is no conceivable scenario when sending "<tons of tokens> + no" makes any sense.
Best case scenario is:
"<tons of tokens> + no" -> "Okay, I won't do it."
In this case you've just waisted a lot of input tokens, that someone (hopefully, not you) has to pay for, to generate an absolutely pointless message that says "Okay, I won't do it.". There is no value in this message. There is bo reason to waste time and computational resources to generate this message.
Worst case scenario is what happened on the screenshot.
There is no good scenario when this input produces a valuable output.
If you want your "agent" or "model" or whatever to do nothing you just don't trigger it. It won't do anything on it's own, it doesn't wait for your response, it doesn't need your response.
I don't understand why, in this thread, every time I try to point out how nonsensical is the behavior that they want is from technical perspective (from the perspective of knowing how these tools actually work) people just cling to there anthropomorphized mind model of the LLM and insist on getting angry.
"It acts like a bad human being, therefore it's bad, useless and dangerous"
I don't even know what to say to this.
P. S. If you wind this message hard to read and understand, I'm sorry about it, I don't know how to word it better. HN disallows to use LLMs to edit comments, but I think that sending a link to an LLM-edited version of the comment should be ok:
https://chatgpt.com/s/t_69b423f52bc88191af36a56993d55aa8