logoalt Hacker News

tedsandersyesterday at 8:09 PM7 repliesview on HN

We don't vary our model quality with time of day or load (beyond negligible non-determinism). It's the same weights all day long with no quantization or other gimmicks. They can get slower under heavy load, though.

(I'm from OpenAI.)


Replies

wasmainiacyesterday at 10:50 PM

Thanks for the response, I appreciate it. I do notice variation in quality throughout the day. I use it primarily for searching documentation since it’s faster than google in most case, often it is on point, but also it seems off at times, inaccurate or shallow maybe. In some cases I just end the session.

show 1 reply
Trufayesterday at 8:18 PM

Can you be more specific than this? does it vary in time from launch of a model to the next few months, beyond tinkering and optimization?

show 2 replies
Someone1234yesterday at 8:29 PM

Specifically including routing (i.e. which model you route to based on load/ToD)?

PS - I appreciate you coming here and commenting!

show 1 reply
zamadatixyesterday at 9:43 PM

I appreciate you taking the time to respond to these kinds of questions the last few days.

GorbachevyChasetoday at 1:35 AM

Hi Ted. I think that language models are great, and they’ve enabled me to do passion projects I never would have attempted before. I just want to say thanks.

derwikitoday at 12:07 AM

Has this always been the case?

fragmedeyesterday at 11:25 PM

I believe you when you say you're not changing the model file loaded onto the H100s or whatever, but there's something going on, beyond just being slower, when the GPUs are heavily loaded.

show 1 reply