logoalt Hacker News

pashtoday at 9:23 PM0 repliesview on HN

> Even for billion-parameter theories, a small amount of vectors might dominate the behaviour.

We kinda-sorta already know this is true. The lottery-ticket hypothesis [0] says that every large network contains a randomly initialized small network that performs as well as the overall network, and over the past eight years or so researchers have indeed managed to find small networks inside large networks of many different architectures that demonstrate this phenomenon.

Nobody talks much about the lottery-ticket hypothesis these days because it isn’t practically useful at the moment. (With the pruning algorithms and hardware we have, pruning is more costly than just training a big network.) But the basic idea does suggest that there may be hope for interpretability, at least in the odd application here or there.

That is, the lottery-ticket hypothesis suggests that the training process is a search through a large parameter space for a small network that already (by random initialization) exhibit the overall desired network behavior; updating parameters during the training process is mostly about turning off the irrelevant parts of the network.

For some applications, one would think that the small sub-network hiding in there somewhere might be small enough to be interpretable. In particular, I would not be surprised if investigating neural networks does some day not too far into the future start to yield good interpretable models of phenomena of intermediate complexity (those phenomena that are too complex to be amenable to classic scientific techniques but simple enough that a neural network yields an unusually small active sub-network).

0. https://en.wikipedia.org/wiki/Lottery_ticket_hypothesis