NeurIPS 2025 Workshop on
Socially Responsible and Trustworthy Foundation Models (ResponsibleFM)

Hilton Mexico City Reforma · Mexico City · NeurIPS 2025

About ResponsibleFM

The ResponsibleFM Workshop is an interdisciplinary forum focused on advancing ethical, inclusive, and socially responsible research in foundation models (language and multimodal). With growing societal impact, we address fairness, accountability, transparency, and safety throughout model development and deployment—proactively tackling ethical and social risks.

We bring together researchers, practitioners, ethicists, policy-makers, and affected communities to catalyze methods and best practices ensuring foundation model research serves the common good.

Where

Hilton Mexico City Reforma, Mexico City (Room: Don Alberto 1)

When

NeurIPS 2025 (Sun 30 Nov 1 p.m. — 8 p.m. CST)

Overview

Key themes and questions we will explore at ResponsibleFM.

Topics

  • Defining & Measuring Trustworthiness
    • Rigorous definitions across fairness, safety, truthfulness, privacy, explainability, robustness, and cultural awareness.
    • Standardized, reproducible evaluation protocols and best practices.
  • Techniques to Enhance Trustworthiness
    • Bias mitigation and fairness methods (pre-training & fine-tuning).
    • Knowledge editing, continual learning, and machine unlearning.
    • Watermarking and provenance tracking for accountability.
    • Defenses against adversarial attacks and jailbreaking; safety layers and red teaming.
  • Deployment & Social Good
    • Case studies in healthcare, education, public policy, social welfare, environment.
    • Managing risks in high-stakes applications and maximizing positive impact.
  • Datasets & Benchmarks
    • Diverse, inclusive, and ethically curated datasets; consent and representation.
    • Comprehensive benchmarks for fairness, robustness, privacy, and more.
    • Transparent documentation (data/model cards, datasheets).
  • Interdisciplinary Perspectives & Governance
    • Insights from social sciences, philosophy, law, and public policy.
    • Legal/ethical frameworks: compliance, auditability, regulation.
    • Participatory, community-engaged risk assessment and governance.

Call for Papers

Submission Site

OpenReview

Format & Policy

  • Format: Single PDF; up to 9 pages main text (references/appendix excluded). The main text must be self-contained.
  • Style: Use the NeurIPS 2025 LaTeX style file. Include references and supplementary in the same PDF.
  • Interdisciplinary: Cross-disciplinary submissions (non-CS) are welcome if related to foundation models.
  • Dual-submission / Non-archival: Ongoing/unpublished work and under-review manuscripts are allowed (respect venue policies). The workshop is non-archival.
  • Visibility: Submissions and reviews are not public. Only accepted papers will be made public.
  • Double-blind: Anonymize all materials (including linked code/data). No acknowledgements at submission time.

Awards: We will select one Best Paper and one Outstanding Paper.

Template: Download NeurIPS 2025 Styles

Important Dates (AoE)

  • Submission: Nov 3
  • Notification: Nov 7
  • Camera-Ready: Nov 23
  • Workshop Day: Sun 30 Nov 1 p.m. — 8 p.m. CST

Accepted Papers

We are pleased to announce the accepted papers for the ResponsibleFM workshop. You can view the full list of accepted papers on OpenReview and their posters here.

Invited Keynote Speakers

Yoshua Bengio

Yoshua Bengio

Université de Montréal, Mila
Safety and Social Impact of Frontier Foundation Models
Kush R. Varshney

Kush R. Varshney

IBM Research
Responsible and Trustworthy Foundation Models in Industry
Diyi Yang

Diyi Yang

Stanford University
Social-aware Foundation Models
Sanmi Koyejo

Sanmi Koyejo

Stanford University
Principled Understanding of Trustworthy Foundation Models
Rada Mihalcea

Rada Mihalcea

University of Michigan
Joan Nwatu

Joan Nwatu

University of Michigan (Joint with Rada)
Aylin Caliskan

Aylin Caliskan

University of Washington
Fairness of Foundation Models
Denghui Zhang

Denghui Zhang

UIUC & Stevens
Copyright Under Fire: The Ethical and Legal Perils of LL Memorization

Event Schedule (CST)

Keynote — Diyi Yang

Keynote — Yoshua Bengio

Keynote — Sanmi Koyejo

Keynote — Rada Mihalcea & Joan Nwatu

Keynote — Denghui Zhang

Oral Presentations

  • SVIP: Towards Verifiable Inference of Open-source Large Language Models
    Yifan Sun, Yuhang Li, Yue Zhang, Yuchen Jin, Huan Zhang
  • MedPAIR: Measuring Physicians and AI Relevance Alignment in Medical Question Answering
    Yuexing Hao, Kumail Alhamoud, Haoran Zhang, Hyewon Jeong, Isha Puri, Grace Yan, Philip Torr, Mike Schaekermann, Ariel Dora Stern, Marzyeh Ghassemi
  • Benchmarking Large Language Models on Safety Risks in Scientific Labs
    Yujun Zhou, Jingdong Yang, Yue Huang, Kehan Guo, Zoe Emory, Bikram Ghosh, Amita Bedar, Sujay Shekar, Zhenwen Liang, Pinyu chen, Tian Gao, Werner Geyer, Nitesh V Chawla, Xiangliang Zhang
  • Completion ≠ Collaboration: Scaling Collaborative Effort with Agents
    Shannon Zejiang Shen, Valerie Chen, Ken Gu, Alexis Ross, Zixian Ma, Jillian Ross, Alex Gu, Chenglei Si, Wayne Chi, Andi Peng, Jocelyn J Shen, Ameet Talwalkar, Tongshuang Wu, David Sontag
  • ARMs: Adaptive Red-Teaming Agent against Multimodal Models with Plug-and-Play Attacks
    Zhaorun Chen, Xun Liu, Mintong Kang, Jiawei Zhang, Minzhou Pan, Shuang Yang, Bo Li
  • Invisible Tokens, Visible Bills: The Urgent Need to Audit Hidden Operations in Opaque LLM Services
    Guoheng Sun, Ziyao Wang, Xuandong Zhao, Bowei Tian, Zheyu Shen, Yexiao He, Jinming Xing, Ang Li
  • LUMINA: Detecting Hallucinations in RAG System with Context–Knowledge Signals
    Samuel Yeh, Sharon Li, Tanwi Mallick
  • General Exploratory Bonus for Optimistic Exploration in RLHF
    Wendi Li, Changdae Oh, Sharon Li
  • Does higher interpretability imply better utility? A Pairwise Analysis on Sparse Autoencoders
    Xu Wang, Yan Hu, Benyou Wang, Difan Zou

Break & Poster Session

Keynote — Aylin Caliskan

Keynote — Kush Varshney

Organizing Committee

Canyu Chen

Canyu Chen

Northwestern University
Yue Huang

Yue Huang

University of Notre Dame
Zheyuan Liu

Zheyuan Liu

University of Notre Dame

Yilun Zhao

Yale University
Zhaorun Chen

Zhaorun Chen

University of Chicago
Haoyue Bai

Haoyue Bai

University of Wisconsin–Madison
Xuandong Zhao

Xuandong Zhao

University of California, Berkeley
Yiyou Sun

Yiyou Sun

University of California, Berkeley
Junyuan Hong

Junyuan Hong

MGH · Harvard Medical School
Xuefeng (Sean) Du

Xuefeng (Sean) Du

Nanyang Technological University
Jindong Gu

Jindong Gu

Google DeepMind & Oxford
Arman Cohan

Arman Cohan

Yale University
Xiangliang Zhang

Xiangliang Zhang

University of Notre Dame
Manling Li

Manling Li

Northwestern University

Advising Committee

Dawn Song

Dawn Song

University of California, Berkeley
Yejin Choi

Yejin Choi

Stanford University & Nvidia
Mohit Bansal

Mohit Bansal

UNC Chapel Hill
Robert Nowak

Robert Nowak

University of Wisconsin–Madison

Workshop Venue

Hilton Mexico City Reforma

Room: Don Alberto 1

Av. Juárez 70, Colonia Centro, 06010 Ciudad de México, CDMX, Mexico

Frequently Asked Questions

Are workshop papers archival?

No. ResponsibleFM is a non-archival venue; submissions may be concurrently or subsequently sent to other venues (subject to their policies).

Can I submit work under review elsewhere?

Yes, as long as you follow the other venue’s dual-submission and anonymity rules. See the CFP for details.

Where do I submit?

Submit via OpenReview.

Contact

Address

Hilton Mexico City Reforma

Join Slack

Invitation Link

Email Us

responsiblefm@googlegroups.com