logoalt Hacker News

steve-atx-7600today at 4:55 AM1 replyview on HN

I did stuff like this with bing when they first released their OpenAI based model. But then they started using something - another LLM maybe - to act as a classifier based on if the output was deemed to be off limits. I would see the model start outputting text that it would normally refuse to discuss only to see it abruptly halt, disappear and the session would be terminated.


Replies

praptaktoday at 7:21 AM

Maybe tell it to output rhyming slang pig Latin.

Or, since you are in a terminal anyway, rot13

show 1 reply