Hacker News new | past | comments | ask | show | jobs | submit

Kolmogorov-Arnold networks may make neural networks more understandable

https://www.quantamagazine.org/novel-architecture-makes-neural-networks-more-understandable-20240911/
loading story #41520120
loading story #41520320
loading story #41522229
loading story #41520047
loading story #41520638
loading story #41521355
loading story #41521838
Can someone explain exactly what is the "unknown" of neural networks? We built them, we know what they comprise of and how they work. Yes, we can't map out every single connection between nodes in this "multilayer perceptron" but don't we know how these connections are formed?
loading story #41520542
loading story #41520094
loading story #41525485
loading story #41520358
loading story #41521220
* Given the training data, and the architecture of the network, why does SGD with backprop find the given f? vs. any other of an infinite set.

* Why are there are a set of f each with 0-loss that work?

* Given the weight space, and an f within it, why/when is a task/skill defined as a subset of that space covered by f?

I think a major reasons why these are hard to answer is that it's assumed that NNs are operating within an inferential statistical context (ie., reversing some latent structure in the data). But they're really bad at that. In my view, they are just representation-builders that find proxy representations in a proxy "task" space (def, aprox, proxy = "shadow of some real structure, as captured in an unrelated space").

loading story #41521186
loading story #41520087
loading story #41528374
loading story #41524567
loading story #41521438
loading story #41519745
loading story #41519934
loading story #41521834
loading story #41522037