|COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring.|
Sparsity: Beyond L1
If you have a question about this talk, please contact Colorado Reed.
The L1 norm is a popular choice of regularizer when learning a mapping from inputs to outputs under a frequentist framework. This norm has the useful property of being convex and promoting sparsity. Sparse solutions are desirable when you believe that only a subset of the input features are required to generate an output e.g. when the input dimension is much larger than the number of data points. A significant body of research is dedicated to studying when such regularizers work and where they do not. These ideas can be extended to group or structured sparsity, which can be used to sparse multi kernel learning and non-linear variable selection. Finally we shall discuss sparse matrices developed by generalizing the L1 norm in particular ways, including sparse PCA and dictionary learning.
This talk is part of the Machine Learning Reading Group @ CUED series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
Other listsCamTalks Martin Centre, 37th Annual Series, Architecture Stem Cell Institute Research Associates
Other talksRinocloud Inspiration on design fixation A literary history of medicine: the world's earliest history of medicine, composed in Syria by the physician and poet Ibn Abi Usaybi'ah (d. 1270) Isolated branches in the phylogeny of Platyhelminthes Low-Temperature Solution-Processed Metal-Oxide Semiconducting Heterostructures for Large-Area Electronics: A Facile Approach Towards Charge-Transport Enhancement Infant Disorganised Attachment: A reassessment of the concept and new research findings