Events
SoDa Symposium for Privacy Week: Two Presentations with Q&A
Event Start Date: Tuesday, January 28, 2025 - 12:00 pm
Event End Date: Tuesday, January 28, 2025 - 1:00 pm
Location: Virtual EST (Zoom)
UMD students, faculty, staff, alumni, and friends—join us for the next SoDa Symposium!
Presentation 1:
People, Language, Large Language Models, and Privacy: A Technical Problem or a Fundamental Puzzle?
Presented by Prof. Ivan Habernal
In this talk, I will most likely ask more questions than give answers. We know that people have a fundamental right to privacy, and we know that differential privacy gives us formal guarantees to protect the privacy of people in a database and to do machine learning. But humans also communicate through language, we use written text to do natural language processing, and we train large language models using the entire internet. So how does this work with privacy? From a technical perspective, it looks like it’s just about finding faster techniques or better models – I’m going to talk here about rewriting texts and generating synthetic data under differential privacy. But from a privacy perspective, things quickly get messy when we start asking fundamental questions, and we might be faced with the conundrum: What is privacy in language in the first place?
Presentation 2:
From Dialogue to Data: How Statisticians Can Safeguard Privacy and Promote Trust in LLMs
Presented by Prof. Rochelle E. Tractenberg
Generative AI and large language models are unleashing unprecedented capabilities in data processing, but they also raise serious questions about protecting the individuals and organizations whose data fuel these models. In this talk, I will bridge ethical frameworks from statistics, computing, and mathematics to illustrate how statisticians can help ensure privacy is more than a buzzword. As large language models (LLMs) blur the boundaries between data, dialogue, and personal expression, statisticians play a pivotal role in shaping responsible AI practices. This talk explores how principles from statistics, computing, and mathematics can help safeguard privacy in language-based systems and cultivate (deeper) public trust. Rather than prescribing a one-size-fits-all formula, we will reflect on the key question: How can we balance transparency, accountability, and intellectual freedom in an environment where AI may inadvertently expose sensitive information? Drawing on ethical frameworks and real-world examples, we’ll highlight the potential for statisticians to steer LLM developments in ways that honor human autonomy, encourage fairness, and uphold the spirit of confidentiality—ultimately fostering more trustworthy AI. By combining methodological rigor with ethical deliberation, statisticians have a unique role to play in defending privacy norms, cultivating transparency, and embedding responsible AI practices across research and industry.
Additional Information:
Please contact infoevents@umd.edu at least one week prior to the event to request disability accommodations. In all situations, a good faith effort (up until the time of the event) will be made to provide accommodations.
Speaker(s): Ivan Habernal, Head of Trustworthy Human Language Technologies Group; Rochelle E. Tractenberg, Director, Collaborative for Research on Outcomes and Metrics