The Biden-Harris administration has joined forces with tech giants and other AI stakeholders to address concerns about the safety and trustworthiness of AI development.
AI Safety Institute Consortium (AISIC)
The AI Safety Institute Consortium (AISIC) was launched by the U.S. Department of Commerce. It aims to follow through on mandates laid out in President Biden’s AI executive order.
Goals of the Consortium
The consortium will focus on developing guidelines for:
- Capability evaluations
- Risk management
- Safety and security
- Watermarking synthetic content
The consortium includes over 200 participants, including:
- Tech companies: OpenAI, Google, Microsoft, Apple, Amazon, Meta, NVIDIA, Adobe, and Salesforce
- Academia: MIT, Stanford, and Cornell
- Think tanks and industry researchers: Center for AI Safety, Institute for Electrical and Electronics Engineers (IEEE), and the Responsible AI Institute
The consortium is a response to the growing concerns about the risks associated with AI development, including:
- National security
- Privacy and surveillance
- Election misinformation
- Job security
The AISIC is a significant step by the U.S. government to formally address the challenges of AI development. It is expected to help mitigate the risks and harness the potential of AI.
The AISIC brings together a diverse group of stakeholders to address the complex challenges of AI safety and trustworthiness. The consortium’s work is expected to have a significant impact on the future of AI development.