logoalt Hacker News

konaradditoday at 12:22 PM2 repliesview on HN

> applying this compression algorithm at scale may significantly relax the memory bottleneck issue.

I don’t think they’re going to downsize though, I think the big players are just going to use the freed up memory for more workflows or larger models because the big players want to scale up. It’s a cat and mouse race for the best models.


Replies

miohtamatoday at 1:51 PM

It will also help with local inference, making AI without big players possible.

show 1 reply
Verdextoday at 12:46 PM

Known in the business as 'pulling a jevons'