Navigating Privacy Risks in Language Models
- đ¤ Speaker: Peter Kairouz -- Google đ Website
- đ Date & Time: Tuesday 16 July 2024, 15:00 - 16:00
- đ Venue: West 2, West Hub (https://www.westcambridgehub.uk/visit)
Abstract
The emergence of large language models (LLMs) presents significant opportunities in content generation, question answering, and information retrieval. Nonetheless, training, fine-tuning, and deploying these models entails privacy risks. This talk will address these risks, outlining privacy principles inspired by known LLM vulnerabilities when handling user data. We demonstrate how techniques like federated learning and user-level differential privacy (DP) can systematically mitigate many of these risks at the cost of increased computation. In scenarios where only moderate-to-weak user-level DP is achievable, we propose a strong (task-and-model-agnostic) membership inference attack that allows us to quantify risk by estimating the actual privacy leakage (empirical epsilon) accurately in a single training run. The talk will conclude with a few projections and compelling research directions.
Series This talk is part of the Cambridge ML Systems Seminar Series series.
Included in Lists
- All Talks (aka the CURE list)
- Artificial Intelligence Research Group Talks (Computer Laboratory)
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge talks
- Chris Davis' list
- Department of Computer Science and Technology talks and seminars
- Guy Emerson's list
- Hanchen DaDaDash
- Interested Talks
- Martin's interesting talks
- ndk22's list
- ob366-ai4er
- PhD related
- rp587
- School of Technology
- Security-related talks
- Speech Seminars
- Trust & Technology Initiative - interesting events
- West 2, West Hub (https://www.westcambridgehub.uk/visit)
- yk373's list
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Peter Kairouz -- Google 
Tuesday 16 July 2024, 15:00-16:00