Hacker News new | past | comments | ask | show | jobs | submit login
I have question about information theoretic view of backprop
2 points by britcruise on Oct 11, 2019 | hide | past | favorite | 2 comments
I'm trying to explain the highest level reason why (modern) back propogation works better than previous methods (such as rosenblatts back propogation in 1958 - yes it's that old). without going into the calculus, i want to just look at the information side of things.

I want to say something like the following:

back in the 1960's we tried back propogation with binary neurons. And so when we tuned the parameters backwards (from output to input) no magnitude information passed back through the net (only DIRECTION - i.e. turn this knob left or right). this is similar to how we can't reverse an operation in modulo arithmetic. so it was a very 'coarse' training process due to a lack of information. (took forever)

When we moved to non-linear units (such as relu) there was now a direct relationships (or analog) between output and input magnitude, and so when we passed backwords through the net we had MAGNITUDE and DIRECTION information (i.e. turn this knob to the right by x). that allowed us to train the net much faster because information from every neuron touched every other neuron. put simply, "we knew how much to turn them, and what direction" during training.

thoughts? what am I glossing over?

thank you.




don't under estimate the 50Mx improvement in computer performance since 1970.

Also, Relu's help with Vanishing/Exploding Gradient problem which allows the information to propagate without sending it in to la la land.

CNNs helped because they don't have to calculate across a fully connected network.


thanks for the comment.

Yes the performance boost in training is critical.

So what I'm saying is this performance boost is thanks to the switch to non-binary neuron (which allows a reversible operation to transmit a magnitude) - that's MOST important.

And separately ReLu are just better at this, because they are linear they don't have "vanishing edges" (which prevents the vanish/explode)

Separately I'm glad you brought up CNNs because CNN's are old, and go back to rosenblatt (1958), his perception had a first layer of local connections in it based on the findings in biological systems.

and of course that's because nature has found it's more efficient, and so the efficiency is huge.

but the point is CNN = fewer knobs to train.

and there are lots of simple ways to help reduce knobs - fix some connections - drop some connections




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: