From x.com
antirez's profile

antirez @antirez

If you want, you can't just run a 1T model from SSD, but a 10T model from manual pen and paper math. The fact is that even small LLMs like Qwen 3.5 35B are already almost too slow for serious usage in high end Mx Apple Silicon. Running big LLMs slow is nice, but useless.

Sort: