Philosophy Fellowship 2023 | CAIS Project
The Center for AI Safety is offering grants for philosophers to pursue research in conceptual AI safety.
Sections Program Details 2023 Fellows 2023 Speakers Fellowship News Donate Applications for the 2023 fellowship are now closed. Thanks to everyone who applied. The Program As AI capabilities continue to improve dramatically, the need for safety research has become increasingly apparent. But given the relative youth of the field, much of the conceptual groundwork has yet to be done. The CAIS Philosophy Fellowship invites philosophers from a variety of backgrounds to acquire an in-depth understanding of the current state of AI safety and contribute to novel and field-orienting research directions. Philosophy Fellowship Motivation How the work of philosophers contributes to the broader sociotechnical AI safety community. 1. Conceptual problem Identify a lack of conceptual clarity in the existing AI safety literature. 2. Conceptual Clarification Dissect the problem using rigorous conceptual analysis and relevant philosophical literature. 3. Sociotechnical Orientation Publish conceptual research to inform sociotechnical strategy. 1. Conceptual Problem Advanced AI creates unique conceptual difficulties. Artificial intelligence is reshaping many aspects of day-to-day life. As AI continues on the trajectory to outperform humans on a wide range of cognitive tasks, questions about their properties and potential harms grow increasingly urgent. Conceptual Examples: How can we build systems that are more likely to behave ethically in the face of a rapidly changing world? What processes might shape the behavior of advanced AI systems? Could advanced AI systems pose an existential risk, and if so, how? 2. Conceptual Clarification Academic philosophers are particularly well-positioned to address these conceptual difficulties Philosophers are experts at thinking hard about abstract conceptual problems with no clear answers. Their expertise in working with imprecise concepts makes them the ideal candidates to address the conceptual issues that are characteristic of AI safety. 3. Sociotechnical Orientation Conceptual clarity orients the broader sociotechnical landscape Having the frameworks to analyze which concerns are the most urgent, which are the most likely candidates for serious harm, and how to navigate these risks enables researchers and key decision-makers to reassess their strategies. Goals & Outcomes: This fellowship addresses the need for conceptual clarification through research and field-building efforts. Research: Our team of philosophers critique and build on the existing conceptual AI safety literature, producing new conceptual frameworks to guide technical research. Thus far, our fellows have collectively produced eighteen original papers, soon to be published, covering topics including interpretability, corrigibility, and multipolar scenarios, to name a few. Field-building: We aim for the influence of this fellowship to extend beyond our current cohort, promoting and incentivizing conceptual AI safety research within the broader academic philosophy community. To date, our fellows have received $50,000 in funding to run a workshop connecting technical and conceptual AI safety researchers , organized numerous workshops, and created a special issue journal publication in Philosophical Studies . 2023 Fellowship: Simon Goldstein Simon Goldstein is an Associate Professor in philosophy at Australian Catholic University. Jacqueline Harding Jacqueline Harding is a PhD Student in Symbolic Systems at Stanford University. Cameron Kirk-Giannini Cameron Domenico Kirk-Giannini is an assistant professor of philosophy at Rutgers University–Newark. Nicholas Laskowski Nick Laskowski is an Assistant Professor in the Philosophy Department at University of Maryland, College Park. Nathaniel Sharadin Nate Sharadin is an Assistant Professor of Philosophy at the University of Hong Kong. Dmitri Gallow Dmitri Gallow is a Senior Research Fellow at the Dianoia Institute of Philosophy at the Australian Catholic University. Mitchell Barrington Mitchell Barrington is a PhD student in Philosophy at the University of Michigan - Ann Arbor. Harry Lloyd Harry R. Lloyd is a PhD student in philosophy at Yale University. Frank Hong Frank Hong received his PhD in Philosophy from USC and is an incoming postdoc at Hong Kong University. Bill D'Alessandro William D’Alessandro is a Postdoctoral Fellow at the Munich Center for Mathematical Philosophy and will be a Marie Curie/UKRI Postdoctoral Fellow at the University of Oxford. Elliott Thornley Elliott Thornley is a Postdoctoral Research Fellow in Philosophy at the University of Oxford. He is working on coherence and corrigibility. Robert Long Robert Long recently completed his PhD in philsophy at New York University, during which he also worked as a Research Fellow at the Future of Humanity Institute. 2023 Guest Speakers: Peter Railton Gregory S. Kavka Distinguished Professor of Philosophy at the University of Michigan - Ann Arbor Hilary Greaves Professor of Philosophy at the University of Oxford, Former Director of the Global Priorities Institute Shelly Kagan Clark Professor of Philosophy at Yale University Vincent Müller Alexander von Humboldt Professor of Ethics and Philosophy of AI at the University of Erlangen-Nuremburg L.A. Paul Millstone Family Professor of Philosophy and Professor of Cognitive Science at Yale University Victoria Krakovna AI Research Scientist at DeepMind Jacob Steinhardt Assistant Professor of Computer Science and AI at UC Berkeley David Krueger Assistant Professor of Computer Science and AI at Cambridge University Walter Sinnott-Armstrong Chauncey Stillman Professor of Ethics at Duke University Lara Buchak Professor of Philosophy at Princeton University Johann Frick Associate Professor of Philosophy at the University of California, Berkeley Wendell Wallach Hastings Center senior advisor, ethicist, and scholar at Yale’s Center for Bioethics Rohin Shah Research Scientist at DeepMind 2023 Fellowship News Stay up to date on the latest news and research from the CAIS Philosophy Fellowship. Sign up for email alerts and announcements of future programs. Thank you! Your submission has been received! Oops! Something went wrong while submitting the form. October 24, 2023 Draft Published
Executive Summary
The article discusses the CAIS Philosophy Fellowship 2023, which aims to engage philosophers in addressing conceptual challenges in AI safety. The fellowship seeks to bridge the gap between philosophical inquiry and practical AI safety research, fostering novel conceptual frameworks to guide technical advancements. The program is structured around identifying conceptual problems, providing rigorous philosophical analysis, and orienting sociotechnical strategies. The fellowship has already produced significant research outputs, including eighteen original papers on topics such as interpretability and corrigibility, demonstrating its potential impact on the broader AI safety community.
Key Points
- ▸ The fellowship targets the need for conceptual clarity in AI safety research.
- ▸ Philosophers are uniquely positioned to address abstract and complex conceptual problems in AI.
- ▸ The fellowship aims to influence the broader academic and technical AI safety landscape through research and field-building efforts.
Merits
Interdisciplinary Approach
The fellowship effectively bridges the gap between philosophy and AI safety, leveraging philosophical expertise to address complex conceptual issues in AI.
Field-Building Impact
The program not only produces research but also aims to incentivize and promote conceptual AI safety research within the broader academic community.
Practical Outcomes
The fellowship has already yielded eighteen original papers, demonstrating tangible contributions to the field of AI safety.
Demerits
Limited Scope
The fellowship's focus on conceptual problems may overlook practical and technical aspects of AI safety that require immediate attention.
Selective Participation
The fellowship may not fully represent the diversity of philosophical perspectives, potentially limiting the breadth of conceptual analysis.
Long-Term Impact Uncertainty
While the fellowship has produced initial outputs, the long-term impact on AI safety strategies and policies remains to be seen.
Expert Commentary
The CAIS Philosophy Fellowship 2023 represents a significant step towards integrating philosophical inquiry into the field of AI safety. By addressing conceptual problems through rigorous philosophical analysis, the fellowship aims to provide a foundation for more informed and effective AI safety strategies. The interdisciplinary nature of the program is particularly noteworthy, as it highlights the importance of collaboration between different academic disciplines to tackle complex challenges. However, the fellowship's focus on conceptual issues may overlook practical and technical aspects that require immediate attention. Additionally, the long-term impact of the fellowship's research on AI safety policies and strategies remains to be seen. Overall, the fellowship's contributions to the field are promising, and its potential to influence broader academic and technical communities is substantial.
Recommendations
- ✓ Expand the fellowship to include a broader range of philosophical perspectives and interdisciplinary collaborations to enhance the diversity of conceptual analysis.
- ✓ Incorporate practical and technical aspects of AI safety into the fellowship's research agenda to ensure a more comprehensive approach to AI safety.