Hacker News new | past | comments | ask | show | jobs | submit
This checks out logical speaking.

The FANG code basis are very large and date back years might not necessarily be using open source frameworks rather in house libraries and frameworks none of which are certainly available to Anthropic or OpenAI hence these models have zero visibility into them.

Therefore combined with the fact that these are not reasoning or thinking machines rather probabilistic (image/text) generators, they can't generate what they haven't seen.

That's why coding agents usually scans various files to figure out how to work in a particular codebase. I work with very large and old project, and Codex most of time manages to work with our frameworks.
No it doesn't check out. I think it's becoming abundantly clear LLMs learn in real time as they speak to you. There's a lot of denial and people claiming they don't learn that their knowledge is fixed on the training data and this is not even remotely true at all.

LLMs learn dynamically through their context window and this learning is at a rate much faster than humans and often with capabilities greater than humans and often much worse.

For a code base as complex and as closed source as google the problems an LLM faces is largely the same as a human. How much can he fit into the context window?

loading story #47403474
loading story #47397519
Huh? I have over a hundred services/repos checked out locally, ranging from 10+ years old to new. I have no problem leveraging AI to work in this large distributed codebase.

Even internal stuff is usable by the model because it’s a pattern matching machine and there should be documentation available, or it can just study the code like a human.

loading story #47403493