Kristina Gligorić, Myra Cheng, Lucia Zheng, Esin Durmus, Dan Jurafsky. NLP Systems That Can’t Tell Use from Mention Censor Counterspeech, but Teaching the Distinction Helps. NAACL 2024.
Myra Cheng, Kristina Gligorić, Tiziano Piccardi, Dan Jurafsky. AnthroScore: A Computational Linguistic Measure of Anthropomorphism. EACL 2024. [website] [slides] [code]
Press coverage by Scientific American and New Scientist.
Myra Cheng, Tiziano Piccardi, Diyi Yang. CoMPosT: Characterizing and Evaluating Caricature in LLM Simulations. EMNLP 2023. [poster] [code]
Myra Cheng, Esin Durmus, Dan Jurafsky. Marked Personas: Using Natural Language Prompts to Measure Stereotypes in Language Models. ACL 2023. [slides] [code]
Social Impact Award, Nominated for Best Paper
Federico Bianchi*, Pratyusha Kalluri*, Esin Durmus*, Faisal Ladhak*, Myra Cheng*, Debora Nozza, Tatsunori Hashimoto, Dan Jurafsky, James Zou, Aylin Caliskan. Easily accessible text-to-image generation amplifies demographic stereotypes at large scale. FAccT 2023.
Press coverage by CBS News Prime Time, the Washington Post and MIT Technology Review.
Myra Cheng, Maria De-Arteaga, Lester Mackey, Adam Tauman Kalai. Social Norm Bias: Residual Harms of Fairness-Aware Algorithms. Data Mining and Knowledge Discovery 2023. [code]
Laura Weidinger, John Mellor, Maribeth Rauh, Conor Griffin, Jonathan Uesato, Po-Sen Huang, Myra Cheng et al. Ethical and Social Risks of Harm from Language Models. FAccT 2022.
Myra Cheng, Alicia DeVrio, Lisa Egede, Su Lin Blodgett, and Alexandra Olteanu. “I Am the One and Only, Your Cyber BFF”: Understanding the Impact of GenAI Requires Understanding the Impact of Anthropomorphic AI.
Pratyusha Ria Kalluri*, William Agnew*, Myra Cheng*, Kentrell Owens*, Luca Soldaini*, Abeba Birhane*. The Surveillance AI Pipeline.
Press coverage by 404 Media.