This lecture will cover an overview of developments in deep learning and then we will dive into several recent papers on developments in deep learning related to convolutional networks, model compression, time-series modeling, and distributed deep learning.
Joseph Gonzalez Overview of Deep Learning [pptx][pdf]
TensorFlow Example [direct link][view]
Symbolic Algebra [direct link][view]
Francois Belletti Principles of Neural Network Design [pdf]
Xin Wang Deep Model Compression [pdf]
Sammy Sidhu Scalable Deep Learning [pdf]
For those who want a good introduction to big ideas in deep learning checkout the following links:
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton ImageNet Classification with Deep Convolutional Neural Networks, NIPS’12.
Blog post. Recurrent Neural Networks and LSTMs
Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1, 4 (December 1989)
Szegedy et al. Going Deeper with Convolutions, CVPR’15. There is also a shorter inception blog post.
Jiquan Ngiam, Aditya Khosla, Mingyu Kim, Juhan Nam, Honglak Lee, and Andrew Y. Ng Multimodal Deep Learning, ICML’11.
Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: A Simple Way to Prevent Neural Networks from Overfitting, JMLR’14.
Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks, NIPS’15.
Felix A. Gers, Jürgen A. Schmidhuber, and Fred A. Cummins. Learning to Forget: Continual Prediction with LSTM. Neural Comput. 12, 10 (October 2000), 2451-2471.
Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, Yoshua Bengio Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling, NIPS’14 Deep Learning Workshop.
Martín Abadi et al. TensorFlow: A system for large-scale machine learning, arXiv only 2016.
Google’s Tensor Processing Unit (TPU):
IEEE Spectrum Article: Google Translate Gets a Deep-Learning Upgrade
Technical paper on the TPU Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation
Google’s Tensor Processing Unit explained: this is what the future of computing looks like
Google supercharges machine learning tasks with TPU custom chip
For those interested in learning more about TensorFlow checkout the tutorials.
Forrest N. Iandola, Khalid Ashraf, Matthew W. Moskewicz, and Kurt Keutzer FireCaffe: near-linear acceleration of deep neural network training on compute clusters CVPR 2016. [Forrest’s Slides], arXiv only
Song Han, Huizi Mao, William J. Dally Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding, ICLR’16.
Wenlin Chen, James T. Wilson, Stephen Tyree, Kilian Q. Weinberger, Yixin Chen Compressing Neural Networks with the Hashing Trick, ICML 15.
Forrest N. Iandola, Song Han, Matthew W. Moskewicz, Khalid Ashraf, William J. Dally, Kurt Keutzer SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size arXiv only