The 2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics (NAACL) is being hosted in Albuquerque, New Mexico from April 29 - May 4. We’re excited to share all the work from SAIL that’s being presented, and you’ll find links to the papers below. Feel free to reach out to the contact authors directly to learn more about the work that’s happening at Stanford!

Main Conference

Benchmarking Distributional Alignment of Large Language Models

Authors: Nicole Meister, Carlos Guestrin, Tatsunori Hashimoto
Contact: nmeist@stanford.edu
Links: Paper
Keywords: nlp tools for social analysis, language/cultural bias analysis, values and culture, benchmarking, nlp datasets, evaluation


Can Unconfident LLM Annotations Be Used for Confident Conclusions?

Authors: Kristina Gligorić*, Tijana Zrnic*, Cinoo Lee*, Emmanuel Candès, and Dan Jurafsky
Contact: gligoric@stanford.edu
Award nominations: Best paper, outstanding paper
Links: Paper
Keywords: llm annotations, statistical inference, validity, nlp tools for social analysis, human behavior analysis


REL-A.I.: An Interaction-Centered Approach To Measuring Human-LM Reliance

Authors: Kaitlyn Zhou, Jena D. Hwang, Xiang Ren, Nouha Dziri, Dan Jurafsky, Maarten Sap
Contact: katezhou@stanford.edu
Award nominations: Best Paper Runner Up
Links: Paper
Keywords: human-lm interaction, reliance, safety, expressions of uncertainty


Rethinking Word Similarity: Semantic Similarity through Classification Confusion

Authors: Kaitlyn Zhou, Haishan Gao, Sarah Chen, Dan Edelstein, Dan Jurafsky, Chen Shani
Contact: cshani@stanford.edu
Links: Paper
Keywords: semantic similarity, human-centered nlp, computational social science


Sneaking Syntax into Transformer Language Models with Tree Regularization

Authors: Ananjan Nandi, Christopher D. Manning, Shikhar Murty
Contact: tgk.ananjan@gmail.com
Award nominations: Oral
Links: Paper | Website
Keywords: multi-task approaches, constituency parsing, syntactic language models, sample efficiency, syntactic generalization


“All that Glitters”: Techniques for Evaluations with Unreliable Model and Human Annotations

Authors: Michael Hardy
Contact: hardym@stanford.edu
Links: Paper | Website
Keywords: evaluation methodologies, model bias/fairness evaluation, model bias/unfairness mitigation, educational applications, ethical considerations in nlp applications, transparency, human-subject application-grounded evaluations, human-centered evaluation


Workshop Papers

Measuring Mental Health Variables in Computational Research: Toward Validated, Dimensional, and Transdiagnostic Approaches

Authors: Chen Shani, Elizabeth C. Stade
Contact: cshani@stanford.edu
Workshop: CLPsych: The Workshop on Computational Linguistics and Clinical Psychology
Links: Paper
Keywords: nlp for cilnical psychology


What Can Large Language Models Do for Sustainable Food?

Authors: Anna T. Thomas, Adam Yee, Andrew Mayne, Maya B. Mathur, Dan Jurafsky, and Kristina Gligorić
Contact: gligoric@stanford.edu
Workshop: First Workshop on AI and Scientific Discovery: Directions and Opportunities (AISD)
Links: Paper
Keywords: large language models, sustainability, climate, food, health, optimization


We look forward to seeing you at NAACL 2025!