Hacker News new | past | comments | ask | show | jobs | submit
The point of doing local inference with huge models stored on an SSD is to do it free, even if slow.