logoalt Hacker News

paxystoday at 7:33 PM4 repliesview on HN

The part about injecting randomness is the most intersting bit of the article.

So if you want your LLM responses to be more distributed (beyond what setting the temperature will allow), add some random english words to the start of the prompt.


Replies

FloorEggtoday at 10:12 PM

Fwiw: I didn't read the post carefully, this is just a passing by comment.

For my own use case I was trying to test consistency or an evaluation process and found that injecting a UUID into the system prompt (busting cache) made a material difference.

Without it, resubmitting the same inputs in close time intervals (e.g. 1, 5, or 30 min) would produce very consistent evaluations. Adding the UUID would decrease consistency (showing true evaluation consistency not artificially improved by catching) and highlight ambiguous evaluation criteria that was causing problems.

So I wonder how much prompt caching is a factor here. I think these LLM providers (all of them) are caching several layers beyond just tokenization.

egeozcantoday at 8:44 PM

In a way that sounds like setting the seed.

show 1 reply
BoingBoomTschaktoday at 9:41 PM

Sounds like dithering to me.

FrancoisBosuntoday at 8:52 PM

Meh, I tell it "use uuidgen and get your randomness from that". Of course, that won't work on ChatGPT web, but works well enough on the command line.