logoalt Hacker News

GTPyesterday at 1:15 PM1 replyview on HN

The middle ground here would be an FPGA, but I belive you would need a very expensive one to implement an LLM on it.


Replies

dogma1138yesterday at 1:41 PM

FPGAs would be less efficient than GPUs.

FPGAs don’t scale if they did all GPUs would’ve been replaced by FPGAs for graphics a long time ago.

You use an FPGA when spinning a custom ASIC doesn’t makes financial sense and generic processor such as a CPU or GPU is overkill.

Arguably the middle ground here are TPUs, just taking the most efficient parts of a “GPU” when it comes to these workloads but still relying on memory access in every step of the computation.

show 1 reply