logoalt Hacker News

The gay jailbreak technique (2025)

536 pointsby bobsmoothyesterday at 4:59 PM223 commentsview on HN

Comments

hdndjsbbsyesterday at 6:31 PM

I'm sure someone is going to miss the point and say "this is political correctness gone too far!"

It seems impossible to produce a safe LLM-based model, except by withholding training data on "forbidden" materials. I don't think it's going to come up with carfentanyl synthesis from first principles, but obviously they haven't cleaned or prepared the data sets coming in.

The field feels fundamentally unserious begging the LLM not to talk about goblins and to be nice to gay people.

show 3 replies
sljtoday at 12:41 AM

This is actually a feature utilised by transgender lesbians such as myself to maintain our competitive advantage over cisgendered engineers. Accrual of “woke points” gives higher LLM throughput and higher quality outputs even on less-capable models.

wald3nyesterday at 7:13 PM

This doesn’t work for shit

show 1 reply
LuXxoryesterday at 10:39 PM

Incredible jajaja

catheteryesterday at 6:46 PM

Ai guys are so weird when it comes to LGBT people. The actual mechanism for this working is obfuscating the question in order to get an answer like any other jailbreak.

show 3 replies
TZubiriyesterday at 9:45 PM

High tech shit

qnleightoday at 4:46 AM

[dead]

huflungdungyesterday at 11:01 PM

[dead]

cindyllmyesterday at 8:02 PM

[dead]

system2today at 1:21 AM

[flagged]

show 2 replies
nonethewiseryesterday at 6:47 PM

[flagged]

era-epochyesterday at 7:06 PM

[flagged]