Javascript is not enabled. This site can still works but it'll be more interactive when javascript is enabled.
loading...
Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
androiddrew
9 hours ago
|
on: Tinybox – A powerful computer for deep learning
Could you share what you are using for inference and how you are running it? I have a 64G VRAM/128G system RAM setup.
reply
sosodev
7 hours ago
|
parent
Most people are using something in the llama family for inference. Llama server is my go to. Unsloth guides describe how to configure inference for your model of choice.
reply