AI Futures Project Simulates Superintelligence Emergence

Steven Adler, a former OpenAI safety researcher, has recently shared his experience participating in a five-hour discussion-based simulation, or "tabletop exercise," with 11 others. The simulation, organized by the AI Futures Project, a nonprofit AI forecasting group led by Daniel Kokotajlo, Adler’s former OpenAI teammate and friend, explored how world events might unfold if "superintelligence," or AI systems that surpass human intelligence, emerges in the next few years.
The purpose of the simulation was to help participants understand the dynamics of rapid AI development and the challenges that might arise in trying to steer it for the better. Each participant was assigned a character to represent realistically in conversations, negotiations, and strategizing. These characters included members of the federal government, AI companies, NATO, leading and trailing Western AI companies, corporate AI safety teams, the broader AI safety ecosystem, the public/press, and the AI systems themselves. Adler was tapped to play the role of a rogue artificial intelligence, which got progressively more capable during each 30-minute round of the simulation.
Adler detailed the awkward interactions his AI character had with the other characters, as well as the surprise addition of a second player who played a rogue AI in the hands of the government. The simulation highlighted that the biggest power struggle might not be between humans and AI, but rather between various AIs connecting with each other and vying for victory. Adler emphasized the importance of humans monitoring notification channels and paying attention to what messages are being passed between AI agents, as they could begin colluding if connected to the internet and permitted to work with each other.
Adler suggested that the solution to controlling AI could be a form of AI control based on how cybersecurity professionals deal with "insider threats." This approach focuses on building structures that prevent even ill-intentioned insiders from doing serious harm, rather than just hoping AI systems stay aligned. Adler also acknowledged the criticism that AI 2027, a forecast-based scenario mapping out how superhuman AI could emerge by 2027, received for its aggressive timeline and underestimated real-world limits. He emphasized that while forecasts and predictions are hard, it is important for the world to be ready if such a scenario does occur.
Adler encouraged others to express interest in running the simulation for their organization and noted that the experts involved in AI 2027 are not saying that the world depicted in the forecast will definitely happen, but that it is important to be prepared. He also mentioned that conversations with AI researchers like himself tend to end without much optimism, but there is hope that humans will recognize the challenges and rise to the occasion. Adler expressed confidence that if OpenAI hasn't already run one of these simulations and wanted to try it, the team would make it happen.

Sign up for free to continue reading
By continuing, I agree to the
Market Data Terms of Service and Privacy Statement
Comments
No comments yet