Maarten Sap

I am an assistant professor at CMU's LTI department with a courtesy appointment in HCII, and a part-time research scientist at the Allen Institute for AI (AI2). My research focuses on endowing NLP systems with social intelligence and social commonsense, and understanding social inequality and bias in language.

Before this, I was a Postdoc/Young Investigator at the Allen Institute for AI (AI2), working on project Mosaic. I received my PhD from the University of Washington where I was advised by Noah Smith and Yejin Choi, and have interned at AI2 working on social commonsense reasoning, and at Microsoft Research working on deep learning models for understanding human cognition.
[bio for talks]

Recent updates:

January 2025 πŸ‘¨πŸΌβ€πŸ«πŸ“’: I'm happy to be giving a talk at the First Workshop on Multilingual Counterspeech Generation at COLING 2025 (remotely)!

December 2024 πŸ‡¨πŸ‡¦β›°οΈ: Excited to be attending my very first NeurIPS conference in Vancouver BC! I'll be giving a talk at New in ML at 3pm on Tuesday!

November 2024 : I received a Google Academic Research Award for our work on participatory impact assessment of future AI use cases.

November 2024 πŸ«‚πŸ‘¨β€πŸ«: Very excited that I now have a courtesy appointment in the Human Computer Interaction Institute!

November 2024 πŸ”πŸ§‘β€πŸŽ“: As a reminder, due to my lab being quite full already, I'm not taking any students in this upcoming PhD application cycle 😟.

November 2024 πŸ–οΈπŸ“š: Excited to give a talk at the 6th Workshop on Narrative Understanding on Computational Methods of Social Causes and Effects of Stories.

November 2024 πŸ–οΈπŸŠ: Excited to attend EMNLP in Miami, where my students will be presenting their papers: Joel Mire on The Empirical Variability of Narrative Perceptions of Social Media Texts, Jocelyn Shen on HEART-felt Narratives: Tracing Empathy and Narrative Style in Personal Stories with LLMs, and Xuhui Zhou on Is This the Real Life? Is This Just Fantasy? The Misleading Success of Simulating Social Interactions With LLMs.

[older news]


My research group:

Dan Chechelnitsky

LTI PhD student
co-advised with Chrysoula Zerva

Joel Mire

LTI MLT student

Karina Halevy

LTI PhD student
co-advised with Mona Diab

Jimin Mun

LTI PhD student

Jocelyn Shen

MIT PhD student
co-advised with Cynthia Breazeal

Akhila Yerukola

LTI PhD student

Mingqian Zheng

LTI PhD student
co-advised with Carolyn RosΓ©

Xuhui Zhou

LTI PhD student


Overarching Research Themes

*Extracted by GPT-4, there may be inconsistencies.* #### *Ethics in AI Interactions* My research group explores the ethical implications of artificial intelligence in human interactions. We investigate biases present in large language models, as highlighted in the important paper [Rejected Dialects: Biases Against African American Language in Reward Models](http://maartensap.com/publications.html#mire2025rejectedDialects), which sheds light on inherent biases against specific dialects. Additionally, we emphasize the importance of user interaction through the lens of [Rel-A.I.: An Interaction-Centered Approach To Measuring Human-LM Reliance](https://arxiv.org/abs/2407.07950), focusing on how users’ reliance on AI can shape outcomes. Our work aims to enhance accountability and value alignment in AI systems, as illustrated in the paper [HAICOSYSTEM: An Ecosystem for Sandboxing Safety Risks in Human-AI Interactions](http://arxiv.org/abs/2409.16427). #### *The Dynamics of Narratives* My research group explores the complexities of narratives, particularly in how they shape and reflect social dynamics. A significant contribution to this area is the paper [Modeling Empathic Similarity in Personal Narratives](https://arxiv.org/abs/2305.14246), which examines how empathy can be quantified within storytelling. We also delve into the impact of narratives on perceptions in the digital age, as discussed in the paper [HEART-felt Narratives: Tracing Empathy and Narrative Style in Personal Stories with LLMs](https://arxiv.org/abs/2405.17633). This theme underscores the importance of understanding narrative flow, supported by findings from [Quantifying the narrative flow of imagined versus autobiographical stories](https://www.pnas.org/doi/10.1073/pnas.2211715119). #### *AI Agents and Social Intelligence* My research group explores the intricate relationships between AI agents and social intelligence. Our work focuses on evaluating the social reasoning capabilities of language models, as highlighted by the paper [SOTOPIA: Interactive Evaluation for Social Intelligence in Language Agents](https://arxiv.org/abs/2310.11667), which provides a framework for assessing AI agents' social dynamics. We also address the ethical trade-offs faced by these agents, elaborated in [AI-LieDar: Examine the Trade-off Between Utility and Truthfulness in LLM Agents](https://arxiv.org/abs/2409.09013). Additionally, our investigations into the limitations of social intelligence in AI are informed by the findings in [Neural Theory-of-Mind? On the Limits of Social Intelligence in Large LMs](https://aclanthology.org/2022.emnlp-main.248/). #### *Understanding Toxic Language* My research group explores the complexities and challenges of addressing toxic language in AI systems. Our recent study [PolygloToxicityPrompts: Multilingual Evaluation of Neural Toxic Degeneration in Large Language Models](https://arxiv.org/abs/2405.09373) surveys biases and deterioration of language models across different languages. Furthermore, we examine the perspectives of those combating online hate in our paper [Counterspeakers’ Perspectives: Unveiling Barriers and AI Needs in the Fight against Online Hate](https://arxiv.org/abs/2403.00179), which seeks to identify gaps in current approaches. Our work also addresses broader contextual factors that influence interpretations of hate speech, as detailed in [COBRA Frames: Contextual Reasoning about Effects and Harms of Offensive Statements](http://arxiv.org/abs/2306.01985).