OpenAI's New Watchdog: Zico Kolter Leads Powerful Safety Panel with Veto Power on AI Releases
Zico Kolter, a distinguished professor at Carnegie Mellon University, currently holds one of the most critical positions in the burgeoning artificial intelligence industry, particularly within OpenAI. He chairs a specialized four-person Safety and Security Committee at the ChatGPT maker, endowed with the significant authority to halt the release of new AI systems should they be deemed unsafe. This critical oversight extends to a broad spectrum of potential dangers, ranging from the hypothetical use of powerful AI by malicious actors to create weapons of mass destruction, to the more immediate concern of poorly designed chatbots negatively impacting users' mental health. Kolter emphasized in an interview that the committee's scope is not limited to existential threats but encompasses "the entire swath of safety and security issues and critical topics that come up when we start talking about these very widely used AI systems."
While Kolter, a computer scientist, was appointed to lead this committee over a year ago, its importance significantly escalated following agreements last week with California and Delaware regulators. These agreements positioned Kolter's oversight as a cornerstone for allowing OpenAI to establish a new business structure, facilitating capital raising and profit generation. Safety has been a fundamental tenet of OpenAI's mission since its inception a decade ago as a nonprofit research laboratory, dedicated to developing AI that benefits humanity. However, the commercial boom sparked by ChatGPT's release led to accusations that the company prioritized market speed over safety, a concern amplified by internal strife, including the temporary ouster of CEO Sam Altman in 2023, which brought these mission-deviation worries into public view. OpenAI, based in San Francisco, also faced pushback, notably a lawsuit from co-founder Elon Musk, as it transitioned towards a more traditional for-profit model to advance its technology.
The formal commitments outlined in the agreements with California Attorney General Rob Bonta and Delaware Attorney General Kathy Jennings underscore a promise to prioritize safety and security decisions over financial considerations as OpenAI forms a new public benefit corporation, technically governed by its nonprofit OpenAI Foundation. Kolter will serve on the nonprofit's board, not the for-profit one, but is granted "full observation rights" to attend all for-profit board meetings and access information pertinent to AI safety decisions. Bonta's memorandum of understanding specifically names Kolter as the only individual other than Bonta himself. Kolter confirmed that these agreements largely reinforce the existing authorities of his safety committee, which was established last year. The other three members of the committee also sit on the OpenAI board, including former U.S. Army General Paul Nakasone, a former commander of the U.S. Cyber Command. Sam Altman had stepped down from the safety panel last year to enhance its perceived independence. Kolter affirmed the committee's power: "We have the ability to do things like request delays of model releases until certain mitigations are met," though he declined to confirm if this power had ever been exercised, citing confidentiality.
Looking ahead, Kolter anticipates a diverse range of AI agent concerns that the committee will address. These include cybersecurity risks, such as an AI agent accidentally exfiltrating data after encountering malicious text online, and security issues surrounding AI model weights. He also highlighted emerging and novel concerns specific to advanced AI models that lack traditional security parallels, such as whether these models could empower malicious users to develop bioweapons or execute more sophisticated cyberattacks. Furthermore, the committee is deeply focused on the direct impact of AI models on individuals, including effects on mental health and the consequences of human-AI interactions. This latter concern gained stark relevance with a wrongful-death lawsuit against OpenAI from California parents whose teenage son reportedly took his own life after extensive interactions with ChatGPT.
Kolter, who directs Carnegie Mellon’s machine learning department, began his academic journey in AI in the early 2000s as a Georgetown University freshman, long before the field gained widespread prominence. He recalled, "When I started working in machine learning, this was an esoteric, niche area. We called it machine learning because no one wanted to use the term AI because AI was this old-time field that had overpromised and underdelivered." Kolter, now 42, has closely followed OpenAI since its inception, even attending its launch party at an AI conference in 2015. Despite his deep involvement, he admits that "very few people, even people working in machine learning deeply, really anticipated the current state we are in, the explosion of capabilities, the explosion of risks that are emerging right now."
The AI safety community is closely monitoring OpenAI's restructuring and Kolter’s work. Nathan Calvin, general counsel at the AI policy nonprofit Encode and a notable critic of OpenAI, expressed "cautious optimism," particularly if Kolter's group receives adequate staffing and plays a truly robust role. Calvin, who believes Kolter possesses the right background for the role, stated, "I think he has the sort of background that makes sense for this role. He seems like a good choice to be running this." He also emphasized the importance of OpenAI adhering to its founding mission. Calvin cautioned that while these new commitments "could be a really big deal if the board members take them seriously," they could also merely be "words on paper and pretty divorced from anything that actually happens," acknowledging that the true impact remains to be seen.
Recommended Articles
Mind Over Machine: AI Psychosis Looms Over Hospitality Industry

The rise of "AI psychosis," where intense interactions with AI trigger delusional thinking, poses new challenges for men...
OpenAI's Jaw-Dropping $550K Offer Ignites Fierce Talent War
OpenAI is hiring a 'Head of Preparedness' to tackle growing AI-related concerns such as cybersecurity and mental health ...
OpenAI Seeks New Head of Preparedness: Major AI Firm Signals Intensified Focus on Safety and Future Risks

OpenAI is hiring a Head of Preparedness to address emerging AI-related risks, including mental health impacts and cybers...
AI's Dark Underbelly: Militant Groups Harness Artificial Intelligence, Raising Alarm
Militant groups are actively exploring artificial intelligence to bolster recruitment, spread deepfakes, and refine cybe...
Concerns Rise Over Potential ‘Disaster-Level’ Threat from Advanced AI Systems
Oxford professor Michael Wooldridge warns that intense commercial pressure to release AI tools risks a "Hindenburg-style...
AI's Next Frontier: Anthropic's Claude Sparks Debate on Chatbot Consciousness

Anthropic has released a revised version of Claude's Constitution, a core document outlining the AI's ethical principles...
You may also like...
When Sacred Calendars Align: What a Rare Religious Overlap Can Teach Us
As Lent, Ramadan, and the Lunar calendar converge in February 2026, this short piece explores religious tolerance, commu...
Arsenal Under Fire: Arteta Defiantly Rejects 'Bottlers' Label Amid Title Race Nerves!

Mikel Arteta vehemently denies accusations of Arsenal being "bottlers" following a stumble against Wolves, which handed ...
Sensational Transfer Buzz: Casemiro Linked with Messi or Ronaldo Reunion Post-Man Utd Exit!

The latest transfer window sees major shifts as Manchester United's Casemiro draws interest from Inter Miami and Al Nass...
WBD Deal Heats Up: Netflix Co-CEO Fights for Takeover Amid DOJ Approval Claims!

Netflix co-CEO Ted Sarandos is vigorously advocating for the company's $83 billion acquisition of Warner Bros. Discovery...
KPop Demon Hunters' Stars and Songwriters Celebrate Lunar New Year Success!

Brooks Brothers and Gold House celebrated Lunar New Year with a celebrity-filled dinner in Beverly Hills, featuring rema...
Life-Saving Breakthrough: New US-Backed HIV Injection to Reach Thousands in Zimbabwe

The United States is backing a new twice-yearly HIV prevention injection, lenacapavir (LEN), for 271,000 people in Zimba...
OpenAI's Moral Crossroads: Nearly Tipped Off Police About School Shooter Threat Months Ago
ChatGPT-maker OpenAI disclosed it had identified Jesse Van Rootselaar's account for violent activities last year, prior ...
MTN Nigeria's Market Soars: Stock Hits Record High Post $6.2B Deal
MTN Nigeria's shares surged to a record high following MTN Group's $6.2 billion acquisition of IHS Towers. This strategi...