Hacker News new | past | comments | ask | show | jobs | submit

Big data on the cheapest MacBook

https://duckdb.org/2026/03/11/big-data-on-the-cheapest-macbook
I’ve been tempted to buy one and do “real dev work” on it just to show people it’s not this handicapped little machine.

I built multiple iOS apps and went through two start up acquisitions with my M1 MBA as my primary computer, as a developer. And the neo is better than the M1 MBA. I edited my 30-45 min long 4k race videos in FCP on that air just fine.

loading story #47351092
loading story #47352650
loading story #47352626
loading story #47350226
loading story #47350231
loading story #47351887
loading story #47351856
I just retired my m1 air to being a server this month. They’re very capable laptops. If the neo is even comparable in spec it’s excellent for the price
loading story #47350845
loading story #47350154
loading story #47350212
loading story #47350603
loading story #47352256
loading story #47351063
loading story #47351461
loading story #47350021
loading story #47352173
loading story #47350696
loading story #47350387
loading story #47350471
loading story #47351006
loading story #47357791
loading story #47357728
This is as much an indictment of AWS compute as it is anything else.
loading story #47350145
loading story #47349992
loading story #47349960
loading story #47349952
When I teach, I use "big data" for data that won't fit in a single machine. "Small data" fits on a single machine in memory and medium data on disk.

Having said that duckDB is awesome. I recently ported a 20 year old Python app to modern Python. I made the backend swappable, polars or duckdb. Got a 40-80x speed improvement. Took 2 days.

loading story #47350637
as a broke ecologist, this little computer can do everything I need in R and word and is a phenomenal build for the price. I'm really enjoying it thus far.
loading story #47350559
loading story #47357116
I adore DuckDB.

Did a PoC on a AWS Lambda for data that was GZ'ed in a s3 bucket.

It was able to replace about 400 C# LoC with about 10 lines.

Amazing little bit of kit.

This is awesome.

I wish more companies would do showcases like this of what kind of load you can expect from commodity-ish hardware.

loading story #47350860
> The cloud instances have network-attached disks

Props for identifying the issue immediately, but armed with that knowledge, why not redo the benchmark on a different instance type that has local storage? E.g. why not try a `c8id.2xlarge` or `c8id.4xlarge` (which bracket the `c6a.4xlarge`'s cost)?

I would have benchmarked with an instance that has local nvme, like c8gd.4xlarge.
loading story #47350022
loading story #47350698
I think it’s relevant to first read [1] to see why they’re doing this. It’s basically done as a meme.

[1] https://motherduck.com/blog/big-data-is-dead/

loading story #47350183
loading story #47353990
That's not Big Data. If you "need to process Big Data on the move" - what you need is a network.
loading story #47350302
loading story #47350905
The DuckDB team benchmarked with an r7i.16xlarge which uses EBS - that's the expected bottleneck. A fairer comparison would be an i4i or c8gd with local NVMe, where you'd likely see the laptop and cloud instance much closer in practice.
loading story #47352281
Funny just yesterday I almost bought one but got cold feet and opted for a low range MacBook with M5 chip. The Apple sales rep was not convinced it would be enough when i described using it for vibecoding and deploying so kind of talked me out of getting the Neo. I normally use a mix of LLMs, then connect to Github and do a one-click deploy on CreateOS. Do you think I over-reacted? The price of the Neo is SO attractive, a clean half price compared to what I got.
loading story #47351150
loading story #47352301
loading story #47350937
loading story #47351581
Would it not also work on a raspberry.

With I/O streaming and efficient transformation I do big data on my consumer PC and good old cheap HDDs just fine.

loading story #47352068
loading story #47352426
For the TPC-DS results it would also have been nice to show how the macbook neo compares to the AWS instances.

Or am I missing something?

> compared to 3–5 GB/s

Their numbers are a bit outdated. M5 Macbook pro SSDs are literally 5x this speed. It's wild.

loading story #47350156
loading story #47350229
> TL;DR: How does the latest entry-level MacBook perform on database workloads? We benchmarked it to find out.

That's not tldr, that's just subheader.

loading story #47350494
loading story #47351048
That c8g.metal-48xl instance costs $7.63008 on demand[1], so for the price of the laptop, you could run queries on it for about ~90 hours.

:shrug: as to whether that makes the laptop or the giant instance the better place to do one's work…

[1] https://aws.amazon.com/ec2/pricing/on-demand/

If you can fit it on a thumb drive, it's not Big Data.
loading story #47350910
“Big data” doesn’t have a 5gb memory cap.

I’m guessing so many devs started out on 32gb MacBooks that the NEO seems underpowered. but it wasn’t too long ago that 8gb, 1500mb/sec IO & so many cores was an elite machine.

I did a lot of dev work on a glorified eePC Chromebook when my laptop was damaged. You don’t need a lot of ram to run a terminal.

I’m hoping NEO resets the baseline testing environment so developers get back to shipping software that doesn’t monopolize resources. “Plays nice with others” should be part of the software developer’s creed.

I'm interested by one (not for big data) but only 8 GB or RAM is kinda really sad.

My good old LG Gram (from 2017? 2015? don't even remember) already had 24 GB of RAM. That was 10 years ago.

A decade later I cannot see myself being a laptop with 1/3rd the mem.

loading story #47350165
Queue the endless blog posts about running tech on the potato macbook and being stunned it’s functional with massive trade-offs. Groundbreaking stuff.
loading story #47349705
loading story #47350939
this has a phone CPU/memory
other test:

2025-09-08 : "Big Data on the Move: DuckDB on the Framework Laptop 13"

"TL;DR: We put DuckDB through its paces on a 12-core ultrabook with 128 GB RAM, running TPC-H queries up to SF10,000."

https://duckdb.org/2025/09/08/duckdb-on-the-framework-laptop...

loading story #47353915
loading story #47353534
You could get a laptop with an Nvidia GPU, 16gb ram, 512 ssd... or a 'cheap' Macbook.

I totally understand if you need to compile for iphones. We need to make apps for the lower and middle class people that think a $40/mo cellphone is a status symbol. I get it.

But if you are not... why? I hate windows, but we have Fedora... and you get an Nvidia. Is it just a status symbol? And I have a hard time believing people who tell me stories about low power consumption, because no one had cared about that until Apple pretended people cared about it.

loading story #47350943
loading story #47351047
loading story #47351183
loading story #47350918
loading story #47352098
loading story #47350903
loading story #47350975
loading story #47351105
loading story #47350897
loading story #47351922
loading story #47351328
[flagged]
loading story #47349640
Mind blown, if you need to handle "big" data on the move - the macbook neo is not the right choice. - Who would have guessed that outcome?
loading story #47349785
That's an awesome idea to get a bricked MacBook Neo really fast because those idiots soldered the SSD inside
loading story #47349657
loading story #47349984
loading story #47349591
Seems completely unnecessary, there is probably 0 overlap between people who buy a cheap MacBook and people running DuckDB locally
loading story #47352432
loading story #47350432
loading story #47350013
loading story #47352119
loading story #47351648
Oh great, the term "big data" is back.
loading story #47349547
>Can I expect good performance from the MacBook Neo with Slack, Microsoft Office, and Google Chrome signed into Atlassian and a CRM, all running simultaneously?

No.

>Do I reject a world where all of the above is necessary to realize value from an entry-level MacBook?

In theory, yes.