Hacker News new | past | comments | ask | show | jobs | submit
loading story #42769222
loading story #42773052
loading story #42768827
loading story #42768789
loading story #42781567
Kind of insane how a severely limited company founded 1 year ago competes with the infinite budget of Open AI

Their parent hedge fund company isn't huge either, just 160 employees and $7b AUM according to Wikipedia. If that was a US hedge fund it would be the #180 largest in terms of AUM, so not small but nothing crazy either

loading story #42768563
loading story #42768571
loading story #42776220
loading story #42770217
loading story #42773011
loading story #42769442
loading story #42773811
loading story #42777585
loading story #42814005
loading story #42774741
loading story #42773863
loading story #42768568
loading story #42768710
loading story #42768536
loading story #42768547
I was initially enthusiastic about DS3, because of the price, but eventually I learned the following things:

- function calling is broken (responding with excessive number of duplicated FC, halucinated names and parameters)

- response quality is poor (my use case is code generation)

- support is not responding

I will give a try to the reasoning model, but my expectations are low.

ps. the positive side of this is that apparently it removed some traffic from anthropic APIs, and latency for sonnet/haikku improved significantly.

loading story #42773177
loading story #42769858
loading story #42768994
loading story #42775037
loading story #42771597
loading story #42792855
loading story #42768987
loading story #42768565
loading story #42768629
loading story #42769087
loading story #42775013
loading story #42775203
loading story #42774505
loading story #42773515
loading story #42769049
loading story #42772326
loading story #42769326
loading story #42774757
loading story #42770347
loading story #42769509
Does anyone know what kind of HW is required to run it locally? There are instructions but nothing about HW required.
loading story #42768742
Deepseek v3 required about 1tb of VRAM / RAM so 10 A100.

There are various ways to run it with lower vram if you're ok with way worse latency & throughput

Edit: sorry this is for v3, the distilled models can be ran on consumer-grade GPUs

loading story #42768625
loading story #42778829
It's just a question of having enough VRAM+RAM to fit the model into memory.
loading story #42773851
loading story #42783581
Have people tried using R1 for some real-world use cases? I attempted to use the 7b Ollama variant for my UI generation [1] and Gitlab Postgres Schema Analysis [2] tasks, but the results were not satisfactory.

- UI Generation: The generated UI failed to function due to errors in the JavaScript, and the overall user experience was poor.

- Gitlab Postgres Schema Analysis: It identified only a few design patterns.

I am not sure if these are suitable tasks for R1. I will try larger variant as well.

1. https://shekhargulati.com/2025/01/19/how-good-are-llms-at-ge... 2. https://shekhargulati.com/2025/01/14/can-openai-o1-model-ana...

loading story #42768895
loading story #42772954
loading story #42777292
loading story #42774040
loading story #42769175
loading story #42816842
loading story #42773900
loading story #42781399
loading story #42775971
loading story #42780898
loading story #42783705
loading story #42768768
loading story #42769954
loading story #42777372
loading story #42779260
loading story #42773402
loading story #42768932
loading story #42777519
loading story #42778181
loading story #42776225
loading story #42770639
loading story #42823956
loading story #42773675
loading story #42784577
loading story #42769588
loading story #42773419
loading story #42805577
loading story #42771932
loading story #42768922
loading story #42772971
loading story #42770627
loading story #42780817
loading story #42788953
loading story #42768956
loading story #42772598
loading story #42774007
loading story #42773533
loading story #42774862
loading story #42775831
loading story #42772488
loading story #42773854
loading story #42774999
loading story #42776898
loading story #42775555
loading story #42795113
loading story #42778832
loading story #42781816
loading story #42769515
loading story #42772432
loading story #42769831
loading story #42772240
loading story #42790242
loading story #42778220
loading story #42779742
loading story #42770240
loading story #42772042
loading story #42777950
loading story #42772490
loading story #42768819
loading story #42775525
loading story #42769788