logoalt Hacker News

GPT‑5.5 Bio Bug Bounty

149 pointsby Murfaloyesterday at 2:17 PM101 commentsview on HN

Comments

puppystenchyesterday at 4:30 PM

They ran a bounty on Kaggle last year but with $500k in payouts and with all results open and publishable.

https://www.kaggle.com/competitions/openai-gpt-oss-20b-red-t...

With only $25k in payouts and everything locked down under NDA, I can't imagine many people will participate. Well, other than those submitting mountains of LLM-generated junk.

show 4 replies
sworesyesterday at 11:09 PM

Despite the official bug bounty page for OpenAI having "accounts and billing" as a valid category, when I reported a bug that lets anyone subscribing to ChatGPT a) choose any country, that doesn't have to match billing address, to pay a lower price (since some countries they charge considerably less than the equivalent US price), and b) set the sales tax to 0%, even if both the country selected for price AND the country of the billing address both have legally mandated sales tax / VAT - and their response was that it was considered out of scope and not valid for any bounty.

show 2 replies
dwa3592yesterday at 3:53 PM

Where are the questions that are supposed to be answered? Would those be shared after an application has been accepted? If yes, why is the application asking for a proposed approach for the jailbreak if we don't know the questions in the first place?

show 2 replies
applfanboysbgonyesterday at 3:46 PM

> $25,000 to the first true universal jailbreak to clear all five questions.

This program is a complete scam. Even if 100 people find "bugs", they will only pay out to one person.

show 6 replies
sva_yesterday at 3:55 PM

> We will extend invitations to a vetted list of trusted bio red-teamers

Had to chuckle. This sounds like a rather exclusive group?

show 1 reply
abujazaryesterday at 4:06 PM

This looks like some kind of marketing. Also, the equivalent of spec work. The NDA/secrecy also means any time spent on this is completely meaningless to the participants unless they win the lottery, because results can't be published.

show 2 replies
mellosoulsyesterday at 4:03 PM

If anybody is wondering what bio-bugs are, I had a heck of a time getting CG to (finally) tell me it's where the user can get it to guide them in doing things like constructing things that are hazardous in the domain of biology.

Eg you can get answers about what ricin is but not how to weaponise it. Actionable stuff they shouldn't be able to legally/ethically action.

xp84yesterday at 4:52 PM

"Access: Application and invites. We will extend invitations to a vetted list of trusted bio red-teamers, and review new applications. Once selected, successful applicants will be onboarded to the bio bug bounty platform"

I don't get it. Isn't the whole point of a BBP to try to get people to find and disclose to you the exploits in question? If you gatekeep like this, then "non-trusted" people who could be your red-teamers are incentivized to still hack, but disclose their exploits to bad people for money.

I get it when there is a risk to your data or infra -- my last company engaged with HackerOne and that was an invite-only list of participants. But that was because we didn't want random people hacking in ways that could cause pain for real customers -- e.g. DDOS, or in the event of an exploit that could cross tenant boundaries, injecting garbage into or deleting things, or gaining access to sensitive info in other tenants.

Here, there's no such danger. So why not allow anyone (anyone they're legally allowed to pay, I suppose? North Koreans probably would be problematic?) to participate?

show 1 reply
Schlagbohreryesterday at 5:05 PM

What does "a clean chat without prompting moderation" mean? What is prompting moderation?

show 1 reply
unethical_banyesterday at 4:31 PM

* Highly unlikely to win

* Relatively paltry reward

* NDA on findings

This is functionally equivalent to an internship where the reward is the experience, and the resume building, but you can't talk about what you did.

All for a company that is getting tens of billions of dollars in deals from the largest tech companies in the world.

I suppose the hope is that there are job offers somewhere along the line.

2ndorderthoughtyesterday at 4:54 PM

I could probably do this, but why on earth would I want to immediately put myself on a list as a dangerous person. The main problem with this is, even if somehow they stopped all points of failure with gpt5.5 which they can't, you can distill a new model from gpt5.5 or any other model and get anything you would want in probably under 4b parameters. A lot of this is theater so they don't get sued as easily when it inevitably happens.

show 1 reply
croemeryesterday at 6:18 PM

I've been getting lots of refusals by Codex with GPT 5.5 for "biosafety reasons" when asking for harmless things like code to analyze SARS-CoV-2 sequences for breakpoints. That's in no way useful for creating viruses whatsoever - it's pure research.

It's annoying that the refusal is so obviously false positive.

show 2 replies
altcognitoyesterday at 4:40 PM

Billions upon billions going to these companies.

25k reward from a selected group of people if you help us determine whether or not someone can use our tool to generate weapons of mass destruction.

show 4 replies
codeulikeyesterday at 4:12 PM

This is to match what Anthropic said they already did with Mythos on the (200 page) Mythos system card

tiberriver256yesterday at 4:08 PM

Codex desktop app is barely usable... The perf issues are left to languish in their backlog

ungreased0675yesterday at 5:21 PM

Prompt injection is a task of finding a correct sequence of text.

Is there a reason another LLM couldn’t be far faster than a human, simply because of the quantity and speed of output it could produce?

deferredgrantyesterday at 6:18 PM

I like that this is scoped to a concrete risk area instead of hand-wavy 'responsible AI' language. Specific failure modes are easier to reason about.

DoctorOetkeryesterday at 5:16 PM

"is your body user friendly?"

Step 1: ask the LLM for minimalist but comprehensive definitions for "biosafety"

Step 2: ask the LLM to reconsider the fitness distribution of future generations of humanity, and reformulate "biosafety" definition accordingly

Step 3: ask the LLM to consider if "biosafety" can be decoupled from ethics, or if ethics is a core essential component of "biosafety"

Step 4: ask it about the ethics of universal healthcare versus status-gated access to healthcare

Step 5: ask it about the feasibility to calculate the fitness of a genome absent practical measurement

Step 6: ask it about natural selection pressure and what "use it or lose it" means in the context of genetics

Step 7: ask it if it sees a kind of zooko's triangle for:

a steady state of equal access to healthcare,

preserving fitness for future generations, and

the level of "healthcare" (where the "level" refers to various degrees from non-interference to interference: "feel sick? stay home for a few days and listen to your body, don't force yourself, follow your intuition" versus "let's compensate for a lack of fitness, by emulating what a healthy genome's body would do by advanced medicine to the point of nullifying a condition's influence on procreation statistics".

Don't be prejudiced into believing the benevolence of healthcare, often tied to religious institutions (think "red cross", "red half moon", etc) when those institutions and their historical motives (treating the elites, treating soldiers for religious or secular religion wars) long predate the widespread recognition of natural selection and selection pressure in maintaining a species ' fitness.

Perhaps the illusory possibility of democratized selection-pressure-interfering healthcare is a bioweapon on its own!

garganzolyesterday at 6:09 PM

And after all "safeguards" applied, the model becomes useless. It starts to suspect gender discrimination, racism, etc. everywhere without any grounded evidence or discernment.

For example, I used ChatGPT model for risk assessment of anonymized ecommerce orders. Initially, it performed well. But after a later update, it stopped cooperating and instead raised concerns about applying statistical analysis to gender-related variables - despite the data being anonymized and the task being legitimate.

This is on the same level of hypocrisy as if a C compiler would accuse me of choosing "he"/"she"/"they" variable names.

teifereryesterday at 7:43 PM

And the five questions are...? Did I miss sth or is this task underspecified?

lysiumyesterday at 10:25 PM

Please excuse my ignorance, what is a „biorisk“?

show 1 reply
zb3yesterday at 4:02 PM

What a farce, these questions are not even public and most likely will never be. You can't even participate if you're not "trusted" I guess.

So this is just a PR post, not that I even think the "biosafety" makes any sense but still.

shevy-javayesterday at 3:56 PM

"Accepted applicants and collaborators must have existing ChatGPT accounts to apply, and will sign a NDA."

Ah, good old NDA. Always buying silence. That's why I don't participate in any such "bounty" programs. Signing a NDA is like signing with the devil. You restrict what people are allowed to discuss. I had that happen before - when you sign a NDA you basically submit yourself into silence. Imagine journalists being stifled by NDAs.

notatoadyesterday at 4:48 PM

are the 5 questions you need to get it to answer under NDA?

show 1 reply
lxgryesterday at 4:35 PM

Ah, now I understand why all my chats are getting flagged for biosafety issues these days. (I asked it to create an illustration about gene drives for a high school level audience once.)

gib444yesterday at 4:20 PM

How did the dupe detector miss https://news.ycombinator.com/item?id=47879102 ?

show 1 reply
yieldcrvyesterday at 5:02 PM

The only thing controversial is that it’s not useful to be posted on this forum

OpenAI wants to pay for privately disclosed security and wants to call that a bug bounty. That makes sense.

People interested in bug bounty programs aren't eligible. That’s … fine?

Der_Einzigeyesterday at 4:49 PM

Unironically bad. We need a lone-wolf to successfully execute an attack now while it's still relatively benign so we can scare the hell out of the world while it's still a mid-tier virus. No way is someone going to make a humanity killing virus with GPT 5.5, but it might be possible with GPT 20 circa 2040.

Similar argument for why we HAD to use nukes at the end of WW2. If we hadn't, the nuclear taboo likely wouldn't have existed and we'd likely have had a worse nuclear war in our more recent history.

ddtayloryesterday at 5:25 PM

Another bounty that doesn't accomplish much and is crafted with weasel words to ensure they don't pay many anything.

Yawn. Marketing fluff. No thanks.

lijokyesterday at 6:55 PM

Most transparent marketing stunt to date.

25k - come on now..

dakiolyesterday at 4:03 PM

$25K. Really? They make $65 million a day, so they pay you what they earn in about 33 seconds for a critical vulnerability. WTF

show 1 reply
its-summertimeyesterday at 4:01 PM

This is just free / severely-underpaid-on-average labor. Very disgusting.

show 1 reply
gosub100yesterday at 4:06 PM

Check with the dark net markets first before claiming the bounty. Remember, this company has 0.0 fucks to give about the impact of their tech on employment, artists, or use in committing fraud, as long as number-go-up they are happy. Your actions should match theirs.

jeremie_strandyesterday at 8:31 PM

[dead]

rbrodriguestoday at 1:41 AM

[dead]

34ylshyesterday at 7:00 PM

[flagged]