logoalt Hacker News

throwaway314155last Thursday at 7:31 PM1 replyview on HN

> which presumably hasn't done a fresh pre-training over the web

What makes you think that?

> Did they figure out how to do more incremental knowledge updates somehow?

It's simple. You take the existing model and continue pretraining with newly collected data.


Replies

Workaccount2last Thursday at 8:45 PM

A leak reported on by semi-analyses stated that they haven't pre-trained a new model since 4o due to compute constraints.