GCRI is a nonprofit and nonpartisan think tank that analyzes the risk of events that could significantly harm or even destroy human civilization at the global scale.
Assessing the Risk of Takeover Catastrophe from Large Language Models
For over 50 years, experts have worried about the risk of AI taking over the world and killing everyone. The concern had always been about hypothetical future AI systems—until recent LLMs emerged. This paper, published in the journal Risk Analysis, assesses how close LLMs are to having the capabilities needed to cause takeover catastrophe.
Recent Publications
Climate Change, Uncertainty, and Global Catastrophic Risk
Is climate change a global catastrophic risk? This paper, published in the journal Futures, addresses the question by examining the definition of global catastrophic risk and by comparing climate change to another severe global risk, nuclear winter. The paper concludes that yes, climate change is a global catastrophic risk, and potentially a significant one.
On the Intrinsic Value of Diversity
Diversity is a major ethics concept, but it is remarkably understudied. This paper, published in the journal Inquiry, presents a foundational study of the ethics of diversity. It adapts ideas about biodiversity and sociodiversity to the overall category of diversity. It also presents three new thought experiments, with implications for AI ethics.
Manipulating Aggregate Societal Values to Bias AI Social Choice Ethics
AI ethics concepts like value alignment propose something similar to democracy, aggregating individual values into a social choice. This paper, published in the journal AI and Ethics, explores the potential for AI systems to be manipulated in ways analogous to sham elections in authoritarian regimes.
The Origin and Implications of the COVID-19 Pandemic: An Expert Survey
Was COVID-19 caused by a natural spillover from wild animals to humans or a research-related accident? And what would each potential origin mean for future pandemics? This report addresses these questions in a survey of 168 virologists, infectious disease epidemiologists, and other scientists from 47 countries.
From the Archive
2020 Survey of Artificial General Intelligence Projects for Ethics, Risk, and Policy
An AI system with wide-ranging general intelligence could bring massive benefits or catastrophic harms, if it is built. This report surveys the global space of projects seeking to build artificial general intelligence. It finds 72 projects across 37 countries, with wide variation in parameters of relevance to AI governance.
Long-Term Trajectories of Human Civilization
What will be the fate of human civilization millions, billions, or trillions of years into the future? This paper, written by an international group of 14 scholars and published in the journal Foresight, studies several possible long-term trajectories of human civilization, including the radically good and catastrophically bad.
A Model for the Probability of Nuclear War
There has only been one nuclear war: World War II. That’s not enough data to calculate the probability. Instead, this report presents a probability model rooted in 14 nuclear war scenarios. It includes a dataset of 60 historical incidents, such as the Cuban missile crisis, that may have threatened to escalate into nuclear war.
The Far Future Argument for Confronting Catastrophic Threats to Humanity: Practical Significance and Alternatives
Global catastrophes could threaten humanity into the distant future, but many people don’t particularly care. This paper, published in the journal Futures, examines strategies for addressing global catastrophic risk that do not require concern about the far future. If the risks are addressed, it may not matter why people address them.
GCRI conducts interdisciplinary research across the global catastrophic risks, with emphasis on six topics.
GCRI Topics
Following an initial global catastrophe event, some humans may still be alive, struggling to survive, holding on to what’s left of civilization. The stakes are extremely high: the difference between humanity regaining its former glory or suffering a dark and premature death. Society today can prepare for catastrophe and help the survivors succeed. GCRI’s research on the aftermath sheds light on this dark and vitally important topic.
Concern about AI catastrophe has existed for many decades. It has recently become a lightning rod for debate alongside the growing societal prominence of AI technology. GCRI’s AI research helps to ground the discussion in the underlying character of the risk and develop solutions that help across a range of AI issues. GCRI is also active in AI ethics, providing perspective on the question of what values to build into AI systems.
Instead of thinking in terms of specific global catastrophic risks in isolation from each other, it is often better to think holistically about the entirety of global catastrophic risk. This includes interconnections between the risks and actions that can affect multiple risks. GCRI was specifically founded as an organization dedicated to studying the whole of global catastrophic risk and this remains a priority for GCRI’s ongoing work.
The world still possess about 10,000 nuclear warheads, enough to cause massive global devastation. As long as adversarial relations between nuclear-armed countries persist, there will be some chance of escalation to nuclear war. To help address the risk, GCRI contributes risk analysis of the probability and severity of nuclear war and solutions for reducing the risk, especially for global effects such as nuclear winter.
Decision analysis can help to identify decision options and their consequences, including their effect on risks. Risk analysis can characterize potential harms and quantify them in terms of their probability and severity. GCRI applies risk and decision analysis methods to global catastrophic risk. GCRI’s approach balances between the value of quantitative analysis and the pitfalls of quantifying deeply uncertain parameters.
To successfully reduce global catastrophic risk, it is necessary to formulate solutions that would reduce the risk if implemented and that will actually be implemented. This often involves engaging with people and institutions that may not have global catastrophic risk as a priority. GCRI’s research develops solutions that account for the priorities of the relevant decision-makers to achieve significant reductions in the risk.
Founded in 2011, GCRI is one of the oldest active organizations working on global catastrophic risk. We provide intellectual and moral leadership to support the broader field of global catastrophic risk.
About GCRI
About
Read about GCRI’s mission, history, organization structure, finances, and more.
People
The GCRI team is led by senior experts in the field of global catastrophic risk.
Official Statements
GCRI publishes occasional statements about important current events, issues within the field, and organizational matters.
Press
GCRI welcomes inquiries from journalists to advance public understanding of global catastrophic risk.
Join GCRI and many others around the world in helping to address humanity’s gravest threats.
How to Get Involved
Learn More
Continue reading the GCRI website to learn more about GCRI, about global catastrophic risk, and more.
Stay Informed
View GCRI publications and organization updates or subscribe to the GCRI newsletter.
Connect
GCRI’s annual Advising and Collaboration Program welcomes people interested in global catastrophic risk.
Donate
GCRI is nonprofit and relies on grants and donations to fund its activities. Visit our donate page to contribute.