Do Deep Nets Really Need To Be Deep?
- 👤 Speaker: Rich Caruana, Microsoft Research
- 📅 Date & Time: Friday 03 October 2014, 10:00 - 11:00
- 📍 Venue: Auditorium, Microsoft Research Ltd, 21 Station Road, Cambridge, CB1 2FB
Abstract
Currently, deep neural networks are the state of the art on problems such as speech recognition and computer vision. We show that by using a method called model compression that shallow feed-forward nets can learn the complex functions previously learned by deep nets and achieve accuracies previously only achievable with deep models. Moreover, in some cases the shallow neural nets can learn these deep functions using the same number of parameters as the original deep models. On the TIMIT phoneme recognition and CIFAR -10 image recognition tasks, shallow nets can be trained that perform similarly to complex, well-engineered, deeper convolutional architectures. Our success in training shallow neural nets to mimic deeper models suggests that there may be better algorithms for training shallow nets than those currently available. I’ll also briefly discuss work we’re doing to compress extremely large deep models and ensembles of deep models to “modest-size” deep models that fit on servers, and to “small” deep models that run on mobile devices.
Series This talk is part of the Microsoft Research Cambridge, public talks series.
Included in Lists
- All Talks (aka the CURE list)
- Auditorium, Microsoft Research Ltd, 21 Station Road, Cambridge, CB1 2FB
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge talks
- Chris Davis' list
- Guy Emerson's list
- Interested Talks
- Microsoft Research Cambridge, public talks
- ndk22's list
- ob366-ai4er
- Optics for the Cloud
- personal list
- PMRFPS's
- rp587
- School of Technology
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Rich Caruana, Microsoft Research
Friday 03 October 2014, 10:00-11:00