Interpreting Multimodal Communication Scenes
- đ¤ Speaker: Steve Renals (Edinburgh)
- đ Date & Time: Tuesday 16 January 2007, 13:00 - 14:00
- đ Venue: LR3, Engineering Department, Baker Building
Abstract
The AMI project is about interpreting human communication using audio, video and other signals, in the context of multiparty meetings. This is an interdisciplinary problem involving signal processing and machine learning (how to make sense of multimodal communication scenes), understanding the content of the recognised signals, and developing useful applications.
In this talk I shall discuss some of the work we have been doing to automatically annotated and interpret multimodal meeting recordings. Specific issues that I’ll discuss include the use of multistream statistical models to segment meetings at different levels (dialogue acts and ``meeting actions’‘) and approaches to meeting summarisation.
Series This talk is part of the Machine Intelligence Laboratory Speech Seminars series.
Included in Lists
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge talks
- Chris Davis' list
- CUED Speech Group Seminars
- Guy Emerson's list
- Information Engineering Division seminar list
- LR3, Engineering Department, Baker Building
- Machine Intelligence Laboratory Speech Seminars
- PhD related
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Steve Renals (Edinburgh)
Tuesday 16 January 2007, 13:00-14:00