University of Cambridge > Talks.cam > Machine Intelligence Laboratory Speech Seminars > Interpreting Multimodal Communication Scenes

Interpreting Multimodal Communication Scenes

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Dr Marcus Tomalin.

The AMI project is about interpreting human communication using audio, video and other signals, in the context of multiparty meetings. This is an interdisciplinary problem involving signal processing and machine learning (how to make sense of multimodal communication scenes), understanding the content of the recognised signals, and developing useful applications.

In this talk I shall discuss some of the work we have been doing to automatically annotated and interpret multimodal meeting recordings. Specific issues that I’ll discuss include the use of multistream statistical models to segment meetings at different levels (dialogue acts and ``meeting actions’‘) and approaches to meeting summarisation.

This talk is part of the Machine Intelligence Laboratory Speech Seminars series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2021 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity