Taylor Sorensen (he/him)
Hi! I’m Taylor Sorensen, a PhD student in Computer Science at the University of Washington, advised by Dr. Yejin Choi. You can reach me at tsor13 [AT] cs.washington.edu.
Research Interests
My research is centered around natural language processing (NLP), with a particular focus on pluralistic alignment (roadmap paper), including (post-)training large language models (LLMs) to represent broad human values and perspectives and maintain diversity. I also aim to enhance AI system’s capabilities to navigate subjective judgments, handle epistemic uncertainty, and support democracy.
News
- 📚 Dedicated NeurIPS workshop on pluralistic alignment, advised by Yoshua Bengio (Dec 2024)
- 🎤 Invited talk at the Berkeley Simons Institute (Oct 2024)
- 🇬🇧 Internship at Google DeepMind in London (July 2024)
- 🎉 Awarded research grant from the Institute for Humane Studies (June 2024)
- 📰 Our ICML position paper featured by Jack Clark’s Import AI Newsletter and Nathan Lambert’s Interconnect Newsletter (June 2024)
- 🎤 Value Kaleidoscope selected for oral presentation (top 1.5%) at AAAI (Feb 2024)
- 🗳️ Our paper on using AI for democratic discourse accepted at PNAS (Oct 2023)
- 🌧️ Internship at Allen Institute for AI (AI2) in Seattle (June 2023)
- 🎓 Started PhD at UW, advised by Yejin Choi (Sep 2022)
Publications
Publications are listed in reverse chronological order. For a list of all publications, see my google scholar profile.
Value Profiles for Encoding Human Variation
Taylor Sorensen, Pushkar Mishra, Roma Patel, Michael Henry Tessler, Michiel Bakker, Georgina Evans, Jason Gabriel, Noah Goodman, Verena Rieser
arXiv Preprint
PaperInformation-Guided Identification of Training Data Imprint in (Proprietary) Large Language Models
Abhilasha Ravichander, Jillian Fisher, Taylor Sorensen, Ximing Lu, Yuchen Lin, Maria Antoniak, Niloofar Mireshghallah, Chandra Bhagavatula, Yejin Choi
NAACL 2025
PaperInvestigating machine moral judgement through the Delphi experiment
Liwei Jiang, Jena D Hwang, Chandra Bhagavatula, Ronan Le Bras, Jenny T Liang, Sydney Levine, Jesse Dodge, Keisuke Sakaguchi, Maxwell Forbes, Jack Hessel, Jon Borchardt, Taylor Sorensen, Saadia Gabriel, Yulia Tsvetkov, Oren Etzioni, Maarten Sap, Regina Rini, Yejin Choi
Nature Machine Intelligence
PaperCan Language Models Reason about Individualistic Human Values and Preferences?
Liwei Jiang, Taylor Sorensen, Sydney Levin, Yejin Choi
Arxiv Preprint
PaperModular Pluralism: Pluralistic Alignment via Multi-LLM Collaboration
Shangbin Feng, Taylor Sorensen, Yuhan Liu, Jillian Fisher, Chan Young Park, Yejin Choi, Yulia Tsvetkov
EMNLP 2024
PaperA Roadmap to Pluralistic Alignment
Taylor Sorensen, Jared Moore, Jillian Fisher, Mitchell Gordon, Niloofar Mireshghallah, Christopher Michael Rytting, Andre Ye, Liwei Jiang, Ximing Lu, Nouha Dziri, Tim Althoff, Yejin Choi
ICML 2024 Position Paper
Paper, Featured in Jack Clark’s Import AI and Interconnects, Invited TalkLeveraging AI for democratic discourse: Chat interventions can improve online political conversations at scale
Lisa P. Argyle, Christopher A. Bail, Ethan C. Busby, Joshua R. Gubler, Thomas Howe, Christopher Rytting, Taylor Sorensen, David Wingate
Published in PNAS
Paper, Science Journal for Kids AdaptationValue Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Taylor Sorensen, Liwei Jiang, Jena Hwang, Sydney Levine, Valentina Pyatkin, Peter West, Nouha Dziri, Ximing Lu, Kavel Rao, Chandra Bhagavatula, Maarten Sap, John Tasioulas, Yejin Choi
AAAI 2024 Oral (top 3% of submissions)
Paper, Presentation, Demo, Code, Dataset, Model, Invited Talk
NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Peter West, Ronan Bras, Taylor Sorensen, Bill Lin, Liwei Jiang, Ximing Lu, Khyathi Chandu, Jack Hessel, Ashutosh Baheti, Chandra Bhagavatula, Yejin Choi
Findings of EMNLP 2023
PaperImpossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and Paraphrasing
Jaehun Jung, Peter West, Liwei Jiang, Faeze Brahman, Ximing Lu, Jillian Fisher, Taylor Sorensen, Yejin Choi
NAACL 2024
PaperTowards Coding Social Science Datasets with Language Models
Christopher Michael Rytting, Taylor Sorensen, Lisa Argyle, Ethan Busby, Nancy Fulda, Joshua Gubler, David Wingate
Arxiv Preprint
PaperPrompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models
David Wingate, Mohammad Shoeybi, Taylor Sorensen
Findings of EMNLP 2022
Paper, CodeAn Information-theoretic Approach to Prompt Engineering Without Ground Truth Labels
Taylor Sorensen, Joshua Robinson, Christopher Michael Rytting, Alexander Glenn Shaw, Kyle Jeffrey Rogers, Alexia Pauline Delorey, Mahmoud Khalil, Nancy Fulda, David Wingate
ACL 2022
Paper, Code, PresentationNl-augmenter: A framework for task-sensitive natural language augmentation
Kaustubh D Dhole, Varun Gangal, Sebastian Gehrmann, …, Taylor Sorensen et al.
Arxiv Preprint
Paper, CodeUsing first principles for deep learning and model-based control of soft robots
Curtis C Johnson, Tyler Quackenbush, Taylor Sorensen, David Wingate, Marc D Killpack
Frontiers in Robotics and AI
Paper, Code
Invited Talks
- [UPCOMING] Foundations of Responsible Computing Pre-Conference Workshop, June 2025
- [UPCOMING] Disagreement Workshop Value Profiles for Encoding Human Variation. May 2025
- National Symposium on Equitable AI Pluralistic Alignment: A Roadmap, Recent Work, and Open Problems. April 2025
- MilaNLP Pluralistic Alignment: A Roadmap, Recent Work, and Open Problems. Jan 2025
- Berkeley Simons Institute Pluralistic Alignment: A Roadmap, Recent Work, and Open Problems. Oct 2024 Recording
- Dealing with Meaning Variation in NLP, University of Utrecht - AI and Pluralistic Human Values Oct 2024
- University College London Aligning AI with Pluralistic Human Values. Sep 2024
- Vienna Alignment Workshop Pluralistic Alignment. July 2024
- IBM Research AI and Pluralistic Human Values. March 2024
- BuzzRobot AI Community Aligning AI with Pluralistic Human Values. May 2024 Recording
====== Website last updated: May 14, 2025