logoalt Hacker News

brentdyesterday at 4:57 PM1 replyview on HN

Regardless of whether the convergence is superficial or not, I am interested especially in what this could mean for future compression of weights. Quantization of models is currently very dumb (per my limited understanding). Could exploitable patterns make it smarter?


Replies

ACCount37yesterday at 5:15 PM

That's more of a "quantization-aware training" thing, really.

show 1 reply