Hacker News new | past | comments | ask | show | jobs | submit login
Deep Dive into Math Behind Deep Networks (towardsdatascience.com)
112 points by headalgorithm on Feb 2, 2019 | hide | past | favorite | 7 comments



That's not a "deep dive" that's the disappointingly barest minimum.


I would like to see someone explain the math behind recurrent neural networks. Feed-forward neural networks are fairly straight-forward, and there are many, many blog posts explaining them already.


I think this resource could be helpful:

http://colah.github.io/posts/2015-08-Understanding-LSTMs/

Essentially, RNNs and feed forward networks are very similar - RNNs are just "unrolled through time" and every timestep shares the same weights. The activations are slightly different as well, but the core concept is the same as feed forward networks; it's not a completely different concept or idea.


If you unroll an RNN (which is what is done) then you have many copies of a single Feed Forward Network. Nothing fancy. The gradient gets accumulated.


Backpropagation section was just a list of formulas, and "it's because of the chain rule".


isn't it just gradient descent?


yes just like a big chunk of what ML is.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: