logoalt Hacker News

ForHackernewstoday at 9:03 AM1 replyview on HN

This reads like it was written by AI. I don't understand how it provides any real security if the "guardrails" against prompt injection are just a system prompt telling the dumber model "don't do this"


Replies

mobilefriendlytoday at 10:13 AM

I had the same thought as well. The firewall is just assuming a dumb model can't be tricked