Reliable and Responsible Foundation Models

Xinyu Yang, Junlin Han, Rishi Bommasani, Jinqi Luo, Wenjie Qu, Wangchunshu Zhou, Adel Bibi, Xiyao Wang, Jaehong Yoon, Elias Stengel-Eskin, Shengbang Tong, Lingfeng Shen, Rafael Rafailov, Runjia Li, Zhaoyang Wang, Yiyang Zhou, Chenhang Cui, Yu Wang, Wenhao Zheng, Huichi Zhou, Jindong Gu, Zhaorun Chen, Peng Xia, Tony Lee, Thomas P Zollo, Vikash Sehwag, Jixuan Leng, Jiuhai Chen, Yuxin Wen, Huan Zhang, Zhun Deng, Linjun Zhang, Pavel Izmailov, Pang Wei Koh, Yulia Tsvetkov, Andrew Gordon Wilson, Jiaheng Zhang, James Zou, Cihang Xie, Hao Wang, Philip Torr, Julian McAuley, David Alvarez-Melis, Florian Tramèr, Kaidi Xu, Suman Jana, Chris Callison-Burch, Rene Vidal, Filippos Kokkinos, Mohit Bansal, Beidi Chen, Huaxiu Yao
TMLR 2025
Reliable and Responsible Foundation Models - Figure

TL;DR

A comprehensive survey and position paper on building foundation models that are both reliable (robust, calibrated, safe) and responsible (fair, private, transparent)—essential reading for deploying AI systems in the real world.

Abstract

Foundation models have fundamentally transformed the artificial intelligence landscape, enabling unprecedented capabilities across diverse domains. However, deploying these powerful systems responsibly requires addressing critical challenges related to reliability and responsibility. This survey provides a comprehensive examination of the key dimensions of reliable and responsible foundation models, including robustness, uncertainty quantification, safety, fairness, privacy, and transparency. We discuss the current state of research, identify open challenges, and outline promising directions for building foundation models that are both capable and trustworthy.

Citation

@article{yang2025reliable,
  title={Reliable and Responsible Foundation Models},
  author={Yang, Xinyu and Han, Junlin and Bommasani, Rishi and others},
  journal={Transactions on Machine Learning Research},
  year={2025}
}