In today’s paper I talk about one of the most important papers in Deep Learning. Dropouts were introduced to make neural networks understand the data better. What’s fascinating is that the idea of dropping out units in a neural network was inspired from Darwanian theory of evolution. It is absolutely amazing when people derive ideas from the natural world and translate those ideas into different scientific disciplines. I had a really wonderful time reading the paper and understanding its core ideas. For this episode, I have changed the format a bit. You will understand what I mean by that if you have followed my episodes thus far. Paper: https://jmlr.org/papers/volume15/srivastava14a/srivastava14a.pdf Please follow me on Spotify if you like this podcast. Suggest to your friends and colleagues if you think I am worth their time. That would mean a lot to me. Also, I would really like it if you gave me feedback or suggestions. If not, write me a mail nonetheless! I would love to get to know you. Email: paperinanutshell@gmail.com If you like my episodes, please follow me on Facebook, Instagram and Twitter. Follow me on Spotify and Apple Pocasts. Facebook: https://www.facebook.com/Paper-In-A-Nutshell-101609351791726 Instagram: https://www.instagram.com/paperinanutshell/ Twitter: https://twitter.com/NutshellPaper Streaming links: https://anchor.fm/debayan-bhattacharya
Informazioni
- Podcast
- Uscita12 dicembre 2020 alle ore 16:05 UTC
- Durata17 min
- ClassificazioneContenuti adatti a tutti