Comment by antinomicus Comment by antinomicus 6 hours ago 4 replies Copy Link View on Hacker News Isn’t the whole point to run your model locally?
Copy Link theptip 6 hours ago Next Collapse Comment - No, that’s clearly not a goal of this project.This is a learning tool. If you want a local model you are almost certainly better using something trained on far more compute. (Deepseek, Qwen, etc) Reply View | 0 replies
Copy Link simonw 5 hours ago Prev Next Collapse Comment - You can run a model locally on much less expensive hardware. It's training that requires the really big GPUs. Reply View | 0 replies
Copy Link yorwba 6 hours ago Prev Next Collapse Comment - The 80 GB are for training with a batch size of 32 times 2048 tokens each. Since the model has only about 560M parameters, you could probably run it on CPU, if a bit slow. Reply View | 0 replies
Copy Link jsight 5 hours ago Prev Collapse Comment - I'd guess that this will output faster than the average reader can read, even while using only CPU inferencing on a modern-ish CPU.The param count is small enough that even cheap (<$500) GPUs would work too. Reply View | 0 replies
No, that’s clearly not a goal of this project.
This is a learning tool. If you want a local model you are almost certainly better using something trained on far more compute. (Deepseek, Qwen, etc)