Task Alignment
- π€ Speaker: Vihari Piratla, University of Cambridge
- π Date & Time: Wednesday 26 June 2024, 11:00 - 12:30
- π Venue: Cambridge University Engineering Department, CBL Seminar room BE4-38.
Abstract
Imagine the implications of submitting an essay or a code routine generated by an LLM without double-checking. Deploying a task-specific model without verifying or modifying is even more consequential. As with any other software, ML models cannot be perfect and need constant monitoring and patching. Yet, the problem of making targeted bug fixes to ML models received little attention and love. We will discuss representative papers of the four broad solutions and their limitations. We will conclude with a critical evaluation of the current progress and future directions. The attached image is a non-exhaustive summary of related work. The papers that we will cover (likely) are further down below.
References
Parameter editing approaches. Shibani Santurkar, Dimitris Tsipras, Mahalaxmi Elango, David Bau, Antonio Torralba, and Aleksander Madry. Editing a classifier by rewriting its prediction rules. Advances in Neural Information Processing Systems, 34:23359β23373, 2021. Transparent model approaches. Pang Wei Koh, Thao Nguyen, Yew Siang Tang, Stephen Mussmann, Emma Pierson, Been Kim, and Percy Liang. Concept bottleneck models. In International conference on machine learning, pages 5338β5348. PMLR , 2020. Bhavana Dalvi Mishra, Oyvind Tafjord, and Peter Clark. Towards teachable reasoning systems: Using a dynamic memory of user feedback for continual system improvement. arXiv preprint arXiv:2204.13074, 2022. Dense data annotation approaches Ross, Andrew Slavin, Michael C. Hughes, and Finale Doshi-Velez. “Right for the right reasons: Training differentiable models by constraining their explanations.” arXiv preprint arXiv:1703.03717 (2017). Sukrut Rao, Moritz BoΜhle, Amin Parchami-Araghi, and Bernt Schiele. Studying how to efficiently and effectively guide models with explanations. In Proceedings of the IEEE /CVF International Conference on Computer Vision, pages 1922β1933, 2023. Data augmentation approaches Shiori Sagawa, Pang Wei Koh, Tatsunori B Hashimoto, and Percy Liang. Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization. arXiv preprint arXiv:1911.08731, 2019.
Series This talk is part of the Machine Learning Reading Group @ CUED series.
Included in Lists
- All Talks (aka the CURE list)
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge talks
- Cambridge University Engineering Department, CBL Seminar room BE4-38.
- Cambridge University Engineering Department Talks
- Centre for Smart Infrastructure & Construction
- Chris Davis' list
- Computational Continuum Mechanics Group Seminars
- custom
- Featured lists
- Guy Emerson's list
- Hanchen DaDaDash
- Inference Group Journal Clubs
- Inference Group Summary
- Information Engineering Division seminar list
- Interested Talks
- Machine Learning Reading Group
- Machine Learning Reading Group @ CUED
- Machine Learning Summary
- ML
- ndk22's list
- ob366-ai4er
- Quantum Matter Journal Club
- Required lists for MLG
- rp587
- School of Technology
- Simon Baker's List
- TQS Journal Clubs
- Trust & Technology Initiative - interesting events
- yk373's list
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Vihari Piratla, University of Cambridge
Wednesday 26 June 2024, 11:00-12:30