University of Cambridge > Talks.cam > Microsoft Research Cambridge, public talks > General-purpose representation learning from words to sentences

General-purpose representation learning from words to sentences

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact lecturescam.

This talk has been canceled/deleted

Real-valued vector representations of words (i.e. embeddings) that are trained on naturally occurring data by optimising general-purpose objectives are useful for a range of downstream language tasks. However, the picture is less clear for larger linguistic units such as phrases or sentences. Phrases and sentences typically encode the facts and propositions that constitute the ‘general knowledge’ missing from many NLP systems at present, so the potential benefit of making representation-learning work for these units is huge. I will present a systematic comparison of (both novel and existing) ways of inducing such representations with neural language models. The results demonstrate clear and interesting differences between the representations learned by different methods; in particular, more elaborate or computationally expensive methods are not necessarily best. I’ll also discuss a key challenge facing all research in unsupervised or representation learning for NLP - the lack of robust evaluations.

This talk is part of the Microsoft Research Cambridge, public talks series.

Tell a friend about this talk:

This talk is included in these lists:

This talk is not included in any other list

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity