▲ | mindcrime 3 days ago | |||||||||||||||||||||||||
Who didn't? Depending on exactly how you interpret the notion of "inventing backpropagation" it's been invented, forgotten, re-invented, forgotten again, re-re-invented, etc, about 7 or 8 times. And no, I don't have specific citations in front of me, but I will say that a lot of interesting bits about the history of the development of neural networks (including backpropagation) can be found in the book Talking Nets: An Oral History of Neural Networks[1]. [1]: https://www.amazon.com/Talking-Nets-History-Neural-Networks/... | ||||||||||||||||||||||||||
▲ | catgary 3 days ago | parent | next [-] | |||||||||||||||||||||||||
I think it’s the move towards GPU-based computing is probably more significant - the constraints put in place by GPU programming (no branching, try not to update tensors in place, etc) sync up with the constraints put in place by differentiable programming. Once people had a sufficiently compelling reason to write differentiable code, the frameworks around differentiable programming (theano, tensorflow, torch, JAX) picked up a lot of steam. | ||||||||||||||||||||||||||
▲ | convolvatron 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||
don't undergrad adaptive filters count? https://en.wikipedia.org/wiki/Adaptive_filter doesn't need a differentiation of the forward term, but if you squint it looks pretty close | ||||||||||||||||||||||||||
▲ | albertzeyer 3 days ago | parent | prev [-] | |||||||||||||||||||||||||
How do you not have the citations in front of you? They are all in the article? I don't expect that any relevant (re)invention of backprop is missing there. Or, if you really know some reinvention of backprop that is not mentioned here, tell Jürgen Schmidhuber, he is actually very curious to learn about other such instances that he is not aware of yet. | ||||||||||||||||||||||||||
|