September 3rd, 2024

OpenAI and Anthropic agree to send models to US Government for safety evaluation

OpenAI and Anthropic have partnered with the U.S. AI Safety Institute to enhance AI model safety through voluntary evaluations, though concerns about the effectiveness and clarity of safety commitments persist.

Read original articleLink Icon
ConcernSkepticismFrustration
OpenAI and Anthropic agree to send models to US Government for safety evaluation

OpenAI and Anthropic have entered into an agreement with the U.S. AI Safety Institute, part of the National Institute of Standards and Technology (NIST), to facilitate safety evaluations of their AI models. This collaboration aims to enhance AI model safety through research, testing, and evaluation, similar to practices in the U.K. The agreement allows the AI Safety Institute to access new models from both companies before and after their public release, providing feedback on potential safety improvements. OpenAI's CEO, Sam Altman, emphasized the importance of this initiative for national leadership in responsible AI development. While the agreement represents a significant step towards establishing safety protocols, it is important to note that participation in safety evaluations remains voluntary and lacks enforceable regulations. Critics have expressed concerns about the vagueness of the term "safety" and the need for AI companies to follow through on their commitments to ensure effective regulation. The agreement is seen as a positive move towards better oversight of AI development, but the effectiveness of voluntary commitments remains to be seen.

- OpenAI and Anthropic will provide AI models for safety evaluations to the U.S. AI Safety Institute.

- The collaboration aims to enhance the safety and responsible development of AI technologies.

- Participation in safety evaluations is voluntary and lacks enforceable regulations.

- Concerns exist regarding the vagueness of "safety" and the need for companies to fulfill their commitments.

- The agreement is viewed as a step forward in establishing AI safety protocols.

Related

Anthropic CEO on Being an Underdog, AI Safety, and Economic Inequality

Anthropic CEO on Being an Underdog, AI Safety, and Economic Inequality

Anthropic's CEO, Dario Amodei, emphasizes AI progress, safety, and economic equality. The company's advanced AI system, Claude 3.5 Sonnet, competes with OpenAI, focusing on public benefit and multiple safety measures. Amodei discusses government regulation and funding for AI development.

OpenAI promised to make its AI safe. Employees say it 'failed' its first test

OpenAI promised to make its AI safe. Employees say it 'failed' its first test

OpenAI faces criticism for failing safety test on GPT-4 Omni model, signaling a shift towards profit over safety. Concerns raised on self-regulation effectiveness and reliance on voluntary commitments for AI risk mitigation. Leadership changes reflect ongoing safety challenges.

AI companies promised to self-regulate one year ago. What's changed?

AI companies promised to self-regulate one year ago. What's changed?

AI companies like Amazon, Google, and Microsoft committed to safe AI development with the White House. Progress includes red-teaming practices and watermarks, but lacks transparency and accountability. Efforts like red-teaming exercises, collaboration with experts, and information sharing show improvement. Encryption and bug bounty programs enhance security, but independent verification and more actions are needed for AI safety and trust.

US lawmakers send a letter to OpenAI requesting government access

US lawmakers send a letter to OpenAI requesting government access

US lawmakers have urged OpenAI to enhance safety standards, allocate resources for AI safety research, and allow pre-deployment testing, following whistleblower allegations and concerns about AI risks and accountability.

OpenAI and Anthropic will share their models with the US government

OpenAI and Anthropic will share their models with the US government

OpenAI and Anthropic have partnered with the U.S. AI Safety Institute for pre-release testing of AI models, addressing safety and ethical concerns amid increasing commercialization and scrutiny in the AI industry.

AI: What people are saying
The comments reflect a range of concerns and opinions regarding the partnership between OpenAI, Anthropic, and the U.S. AI Safety Institute.
  • There is skepticism about the clarity and effectiveness of AI safety evaluations, with some questioning what these evaluations will entail.
  • Concerns about regulatory capture and the influence of corporations on government decisions are prevalent, suggesting that the partnership may primarily benefit the companies involved.
  • Commenters express doubt about the government's readiness to conduct comprehensive safety evaluations.
  • Some believe that the term "AI safety" is overloaded and lacks a clear definition, leading to potential misuse in shaping acceptable discourse.
  • There is a general unease about the lack of congressional input in the collaboration, despite the technical nature of the organization involved.
Link Icon 18 comments
By @bko - 6 months
My issue with AI safety is that it's an overloaded term. It could mean anything from an llm giving you instructions on how to make an atomic bomb to writing spicy jokes if you prompt it to do so. it's not clear which safety these regulatory agencies would be solving for.

But I'm worried this will be used to shape acceptable discourse as people are increasingly using LLMs as a kind of database of knowledge. It is telling that the largest players are eager to comply which suggests that they feel they're in the club and the regulations will effectively be a moat.

By @0cf8612b2e1e - 6 months
What exactly does the evaluation entail? Ask a bunch of naughty questions and see what happens? Unless the model can do nothing, I imagine all of them can be tricked into saying something unfortunate.

Naughty is in the eye is the beholder. Ask me what a Satanist is, and I would expect something about a group who challenges religious laws enshrining Christianity. Ask an evangelical and discussing the topic could be forbidden heresy.

Pretty much any religious topic is going to anger someone. Can the models safely say anything?

By @ChrisArchitect - 6 months
By @accra4rx - 6 months
Bigger question : Is US Government ready to do a comprehensive safety evaluation ? I think it it a cheap way for OpenAI and Anthropic to get a vetting that their models are safe to use and be adaptable by various Govt entity and other organization
By @squarefoot - 6 months
For some reason I mentally swapped OpenAI+Anthropic with parents and models with kids, possibly because it seemed the natural extension of a rotten corrupt mindset that can only produce disasters if given enough power.
By @josephd79 - 6 months
Great. Exactly what we need, more govt reg…
By @beefnugs - 6 months
I like to chuckel to myself that this is what happens when you try the ole "will our product be TOOO AMAZING IT MIGHT END THE WORLD??" viral marketing attempt
By @JumpCrisscross - 6 months
Not thrilled about this happening with zero input from the Congress.

That said, this is the NIST, a technical organisation. This collaboration will inform future lawmaking.

By @DSingularity - 6 months
Of course they want this. Now they get to do all the fun and profitable stuff and all responsibility is on the government which certified the models.
By @valunord - 6 months
Oh boy. Here we go.
By @earleybird - 6 months
I have the utmost respect for the standards and engineering work done by NIST. I'm left with such cognitive dissonance seeing their name juxtaposed with "AI safety". That said, if anyone can come up with a decent definition, I have faith that NIST would be the ones though I'm not holding my breath.
By @irthomasthomas - 6 months
Oh boy, I hope those models do not get leaked in the process.
By @bschmidt1 - 6 months
Because lobbying exists in this country, and because legislators receive financial support from corporations like OpenAI, any so-called concession by a major US-based company to the US Government is likely a deal that will only benefit the company.

Altman has been clear for a long time he wants the government to step in and regulate models (obvious regulatory capture move). They haven't done it, and no amount of Elon Musk or Joe Rogan influence can get people to care, or see it as anything other than regulatory capture. This is OpenAI moving forward anyway, but they can't be the only ones. Hey Anthropic, get in...

- It makes Anthropic "the other major provider", the Android to OpenAI's Apple

- It makes OpenAI not the only one calling for regulation

It reminds me of when Ted Cruz would grill Zuck on TV, yell at him, etc. - it's just a show. Zuck owns the senators, not the other way around. All the big players in our economy own a piece of the country, and they work together to make things happen - not the government. It's not a cabal with a unified agenda, there are competing interests, rivalries, and war. But we the voter aren't exposed to the real decision-making. We get the classics: Abortion, same-sex marriage, which TV actor is gonna win president - a show.

By @nutanc - 6 months
Dammit, now bureaucrats in other countries will jump on this as they have something easy to copy and get ahead in their profession.
By @plsbenice34 - 6 months
Does anyone take the term 'safety' here seriously without laughing? It is so obvious that it is a propaganda term for censorship and manipulation of cultural-political values