logoalt Hacker News

DrScientistyesterday at 9:41 AM1 replyview on HN

Anybody know of good real world examples for continual learning?

Does it really work?


Replies

DoctorOetkeryesterday at 9:01 PM

In this case I think you'd want to use Source-Aware-Training [0] to associate a "timestamp" vector to each native context chunk (perhaps overlapping) of conversation, probably the weights using a kind of Gray code so that the LLM has the immediate out-of-native context history can be retrieved through the nearby gray code of 1, 2, etc steps ago compared to the current timestep gray code.

https://arxiv.org/abs/2404.01019