Prioritizing Security: A Closer Look at the Measures Being Implemented
In a bid to understand and mitigate the risks posed by powerful AI systems, several countries, including the UK, have established AI Safety Institutes (AISIs). The UK's AISI, now renamed CAISI in the US, plays a central role in coordinating technical standards, evaluation tools, and stakeholder collaboration to address AI risks.
The US's National Institute of Standards and Technology’s Center for AI Standards and Innovation (CAISI) leads the creation of inter-agency, sector-specific model evaluation tools and fosters public-academic dialogue to promote best practices in AI safety. The UK government, on the other hand, has included AI safety prominently in its 10-year National AI Strategy, recognizing long-term risks from misaligned advanced AI systems and convening global summits on AI safety to foster international coordination.
However, the governance landscape still grapples with significant challenges in effectively assessing AI model safety in real-world contexts. Real-world testing and validation gaps, dynamic and complex risk landscapes, lack of robust accountability frameworks, and voluntary compliance are some of the issues that need to be addressed.
To tackle these challenges, institutions like CAISI are expanding the use of technical standards, model evaluations, and testing environments (including regulatory sandboxes) that simulate real-world conditions to identify vulnerabilities before wide deployment. Comprehensive risk frameworks, independent safety assessment indices, and global coordination are also emerging solutions.
Despite these efforts, it's crucial to acknowledge the limitations of the voluntary regime for access to AI models. Recent reporting has highlighted that voluntary agreements are fraying, and three of the four major foundation model developers have failed to provide the requested pre-release access to the UK's AISI for their latest cutting-edge models.
To play a role in independent public interest research, the UK AISI needs to be integrated into a regulatory structure with complementary parts that can provide appropriate, context-specific assurance that AI systems are safe and effective for their intended use. The regulatory and governance regime for AI should aim to prevent and mitigate harms, and provide people with opportunities to seek redress and challenge harms when they do occur.
In summary, national AI safety institutes are pivotal in advancing AI governance by developing standards, evaluation tools, and fostering collaboration. However, the field faces ongoing challenges around real-world testing, comprehensive risk coverage, and enforcement. Prospective solutions include regulatory sandboxes, standardized risk management frameworks, independent safety audits, and global coordination to address these challenges in a rapidly evolving landscape.
Key Aspects of AI Safety Governance and Model Evaluation
| Aspect | Description | Example/Institution | |--------------------------------|-------------------------------------------------------------------------------------------------------------------------|---------------------------------------------| | National AI Safety Institutes | Lead technical evaluation, stakeholder collaboration, and development of standards | NIST CAISI (US), UK AI Safety Institute | | Regulatory Sandboxes | Controlled real-world testing environments to evaluate AI models in situ | Proposed in US and Texas programs[1][3] | | Risk Management Frameworks | National guidelines for AI risk assessment to guide development and deployment decisions | NIST AI Risk Management Framework | | Safe Harbor Provisions | Legal protection for companies that adhere to recognized risk management practices | Texas Responsible AI Governance Act[3] | | Independent Safety Assessment | Third-party indexes evaluating AI companies’ safety strategies, research, governance, and transparency | Future of Life Institute AI Safety Index[4] | | Global Coordination | International summits and ethical guidelines to harmonize AI safety approaches | UK AI Safety Summit 2023, China's ethics[2]|
Caveats and Limitations
- Current governance efforts tend to lag behind the rapid technical advancements in AI, creating a regulatory gap.
- Many promising frameworks and sandboxes are still under development or lack comprehensive scope and enforcement details.
- Effective real-world safety evaluation requires continuous updating of standards and cooperation across sectors and nations.
Technology plays a vital role in the development and evaluation of AI systems, as illustrated by the use of technical standards and model evaluations by institutions like the National Institute of Standards and Technology's Center for AI Standards and Innovation (CAISI) and the UK AI Safety Institute.
Moreover, technology also contributes to the creation of regulatory sandboxes, which serve as controlled real-world testing environments to evaluate AI models in situ. This is only one component of the broader efforts to ensure AI safety governance, which encompass the development of risk management frameworks, the provision of safe harbor provisions, independent safety assessments, and global coordination.