William Saunders, an ex-security employee of OpenAI, recently likened the AI company to White Star Line, the builder of the Titanic. Having spent three years on OpenAI’s superalignment team, Saunders revealed in a recent interview that he resigned to avoid “working for the AI Titanic.”
During his tenure at OpenAI, William Saunders debated whether the company was more akin to NASA’s Apollo program or the Titanic. His concerns centered on OpenAI’s pursuit of Artificial General Intelligence (AGI) while also launching paid products. Saunders argued that the company focuses on creating “shiny products” rather than emphasizing safety and risk assessment, making it unlike the Apollo program.
He highlighted that the Apollo program meticulously predicted and assessed risks, maintaining “sufficient redundancy” to handle serious problems, as demonstrated by Apollo 13. In contrast, the White Star Line built the Titanic with watertight compartments and marketed it as unsinkable but failed to provide enough lifeboats, leading to disaster.
Saunders fears that OpenAI overly relies on its current, inadequate security measures and suggested that the company should postpone releasing new AI models to explore potential risks. As the leader of a team focused on understanding AI language model behaviors, Saunders stressed the importance of developing techniques to evaluate whether these systems “hide dangerous capabilities or motivations.”