OpenAI Unveils Risky AI Safety Plan with Board Veto Power

OpenAI Unveils Risky AI Safety Plan with Board Veto Power
Cover art/illustration via CryptoSlate. Image includes combined content which may include AI-generated content.

OpenAI has officially revealed a comprehensive safety initiative aimed at mitigating the risks associated with the development of advanced artificial intelligence (AI) models. The company’s board of directors will now possess the authority to veto decisions made by Chief Executive Sam Altman if they perceive the AI-related risks to be excessively high.

OpenAI’s Safety Initiative: Addressing Risks in Advanced AI Development

In a recent statement, OpenAI acknowledged the inadequacy of current research in addressing potential risks associated with cutting-edge AI technologies. To bridge this gap and formalize their approach to safety concerns, the organization is adopting the initial version of its Preparedness Framework. This framework outlines OpenAI’s protocols for monitoring, assessing, predicting, and safeguarding against catastrophic risks posed by increasingly powerful AI models.

While dedicated “safety systems” teams will oversee potential abuses and risks associated with existing AI models like ChatGPT, the Preparedness Team will focus on assessing frontier models. Additionally, a “superalignment” team will supervise the development of “superintelligent” models, all of which will report directly to the board.

Although the creation of AI surpassing human intelligence may seem distant, OpenAI emphasizes the importance of anticipating future developments. The company intends to push new models to their limits, providing detailed scorecards for four risk categories: cybersecurity, persuasion (involving lies and disinformation), model autonomy (independent decision-making), and CBRN (chemical, biological, radiological, and nuclear threats).

Each risk category will receive a score ranging from low to critical risk, followed by a post-mitigation assessment. If the risk is deemed medium or below, the technology can be deployed. For high risks, development can proceed with additional precautions, while critical risks will halt all further development. OpenAI also commits to accountability measures, stating that independent third parties will be engaged to audit the technology and provide feedback in case of issues.

The company emphasizes close collaboration with external parties and internal teams, such as Safety Systems and Superalignment, to monitor real-world misuse and emergent misalignment risks. OpenAI is also at the forefront of pioneering research to measure how risks evolve as models scale, drawing parallels with their earlier success in scaling laws.

Google News Icon

Add Slash Insider to your Google News Feed

Source(s): Euro News

The information above is curated from reliable sources, modified for clarity. Slash Insider is not responsible for its completeness or accuracy. Please refer to the original source for the full article. Views expressed are solely those of the original authors and not necessarily of Slash Insider. We strive to deliver reliable articles but encourage readers to verify details independently.