Friday, November 22, 2024
Uncategorized

Top AI companies visit the White House to make ‘voluntary’ safety commitments

While substantive AI legislation may still be years away, the industry is moving at light speed and many — including the White House — are worried that it may get carried away. So the Biden administration has collected “voluntary commitments” from 7 of the biggest AI developers to pursue shared safety and transparency goals ahead of a planned Executive Order.

OpenAI, Anthropic, Google, Inflection, Microsoft, Meta, and Amazon are the companies taking part in this non-binding agreement, and will send representatives to the White House to meet with President Biden today.

To be clear, there is no rule or enforcement being proposed here — the practices agreed to are purely voluntary. But although no government agency will hold a company accountable if it shirks a few, it will also likely be a matter of public record.

Here’s the list of attendees at the White House gig:

  • Brad Smith, President, Microsoft
  • Kent Walker, President, Google
  • Dario Amodei, CEO, Anthropic
  • Mustafa Suleyman, CEO, Inflection AI
  • Nick Clegg, President, Meta
  • Greg Brockman, President, OpenAI
  • Adam Selipsky, CEO, Amazon Web Services

No underlings, but no billionaires, either. (And no women.)

The seven companies (and likely others that didn’t get the red carpet treatment but will want to ride along) have committed to the following:

  • Internal and external security tests of AI systems before release, including adversarial “red teaming” by experts outside the company.
  • Share information across government, academia, and “civil society” on AI risks and mitigation techniques (such as preventing “jailbreaking”).
  • Invest in cybersecurity and “insider threat safeguards” to protect private model data like weights. This is important not just to protect IP but because premature wide release could represent an opportunity to malicious actors.
  • Facilitate third-party discovery and reporting of vulnerabilities, e.g. a bug bounty program or domain expert analysis.
  • Develop robust watermarking or some other way of marking AI-generated content.
  • Report AI systems’ “capabilities, limitations, and areas of appropriate and inappropriate use.” Good luck getting a straight answer on this one.
  • Prioritize research on societal risks like systematic bias or privacy issues.
  • Develop and deploy AI “to help address society’s greatest challenges” like cancer prevention and climate change. (Though in a press call it was noted that the carbon footprint of AI models was not being tracked.)

Though the above are voluntary, one can easily imagine that the threat of an Executive Order — they are “currently developing” one — is there to encourage compliance. For instance, if some companies fail to allow external security testing of their models before release, the E.O. may develop a paragraph directing the FTC to look closely at AI products claiming robust security. (One E.O. is already in force asking agencies to watch out for bias in development and use of AI.)

The White House is plainly eager to get out ahead of this next big wave of tech, having been caught somewhat flat-footed by the disruptive capabilities of social media. The President and Vice President have both met with industry leaders and solicited advice on a national AI strategy, as well is dedicating a good deal of funding to new AI research centers and programs. Of course the national science and research apparatus is well ahead of them, as this highly comprehensive (though necessarily slightly out of date) research challenges and opportunities report from the DOE and National Labs shows.

source

Leave a Reply

Your email address will not be published. Required fields are marked *