Visualizing and Understanding Recurrent Networks
- đ¤ Speaker: Andrej Karpathy, PhD student, Stanford University
- đ Date & Time: Thursday 06 August 2015, 12:00 - 13:00
- đ Venue: Cambridge University Engineering Department, LT1
Abstract
Recurrent Neural Networks (RNNs), and specifically a variant with Long Short-Term Memory (LSTM), are enjoying renewed interest as a result of successful applications in a wide range of machine learning problems that involve sequential data. However, while LST Ms provide exceptional results in practice, the source of their performance and their limitations remain rather poorly understood. Using character-level language models as an interpretable testbed, we aim to bridge this gap by providing a comprehensive analysis of their representations, predictions and error types. In particular, our experiments reveal the existence of interpretable cells that keep track of long-range dependencies such as line lengths, quotes and brackets. Moreover, an extensive analysis with finite horizon n-gram models suggest that these dependencies are actively discovered and utilized by the networks. Finally, we provide detailed error analysis that suggests areas for further study.
Series This talk is part of the CUED Computer Vision Research Seminars series.
Included in Lists
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge talks
- Cambridge University Engineering Department, LT1
- Chris Davis' list
- CUED Computer Vision Research Seminars
- Information Engineering Division seminar list
- Interested Talks
- ndk22's list
- ob366-ai4er
- rp587
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Andrej Karpathy, PhD student, Stanford University
Thursday 06 August 2015, 12:00-13:00