Hacker News new | past | comments | ask | show | jobs | submit
IMO there are much better ways to spend 300 million in research beyond firing up a cluster for 60 days to train on internet content.
Spending $100 million one time on a GPT4-level model that is open-source would help with a lot of that research. Especially after all the 3rd party groups fine-tuned it or layered their tools on it.

I think the Copilot-equivalent tools alone would make it quickly pay itself off in productivity gains. Research summaries, PDF extraction, and OCR would add more to that.