Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Do AI Companies Make Good on Voluntary Commitments to the White House?

Created by
  • Haebom

Author

Jennifer Wang, Kayla Huang, Kevin Klyman, Rishi Bommasani

Outline

This paper evaluates the compliance of major AI companies with voluntary commitments based on voluntary AI governance guidelines announced by the White House, the G7, Blechley Park, and Seoul. Based on the White House's eight voluntary commitments for 2023, we developed a detailed evaluation scale to score companies' publicly disclosed actions. The results revealed significant differences in compliance across companies. OpenAI, which received the highest score, achieved 83%, while the average score was only 53%. In particular, compliance with model weight security commitments was extremely poor, with an average of 17%, with 11 out of 16 companies achieving 0% compliance. This highlights a clear structural deficiency that must be addressed in future AI governance initiatives: proactively disclosing how companies implement their public commitments to ensure accountability and ensuring the verifiability of such disclosures. Finally, we present three policy recommendations to address issues of unclear commitments, the role of complex AI supply chains, and public transparency.

Takeaways, Limitations

Takeaways:
The study revealed that AI companies are very poorly implementing their voluntary commitments, particularly regarding model weight security.
Future AI governance initiatives should focus on ensuring transparency and verifiability of companies' commitments.
The three policy recommendations presented offer practical policy directions that can be applied to AI governance initiatives worldwide.
Limitations:
Because this study relied solely on publicly available information, it may not fully reflect the actual implementation levels of companies.
Further review of the objectivity and reliability of the assessment scale is needed.
There is a lack of concrete measures to ensure transparency and verification across the complex AI supply chain.
👍