Leading AI companies publicly shared their safety policies on Friday, 27 October, in response to a request from the UK Technology Secretary last month. Google DeepMind, Anthropic, OpenAI, Microsoft, Amazon, and Meta have all published their AI policies, a move seen as a way to boost transparency and encourage the dissemination of best practices within the AI community. The move outlines the tech companies' renewed commitment to managing the risks posed by AI, including safety, security, and trust.
They agreed to information sharing across the industry and with governments, civil society, and academia on managing AI risks, including approaches to safety, attempts to bypass safeguards, and technical collaboration. For example, Microsoft and OpenAI shared how they will advance responsible AI, including by implementing prior voluntary pledges they made with other companies at the White House in July.
The tech giants are also committed to developing and deploying advanced AI systems to help address humankind's most pressing challenges, such as climate change and cancer prevention.
The UK Government's Frontier AI Taskforce has been hiring top names from all sectors of the AI ecosystem to advise on the risks and opportunities of AI. These safety policies and voluntary commitments are a significant step towards responsible AI innovation. Leading AI firms will hire independent specialists to examine their models for vulnerabilities and to disclose their findings to each other, governments, and researchers.
The news comes one day after Prime Minister Rishi Sunak confirmed the UK would establish the world's first AI Safety Institute to enhance knowledge, carefully evaluate new types of AI models, and better understand their capabilities. Sunak made the announcement on the eve of the Nov. 1