Anthropic aims to fund a new, more comprehensive generation of AI benchmarks

Anthropic launches a program to fund the development of new types of benchmarks to evaluate the performance and impact of AI models, including generative models such as Claude.

Anthropic’s program, unveiled Monday, will provide grants to outside organizations that, as the company put it in a blog post, “can effectively measure advanced capabilities in AI models.” Interested parties can submit applications to be reviewed on a rolling basis.

“Our investment in these assessments is intended to advance the entire field of AI safety by providing valuable tools that benefit the entire ecosystem,” Anthropic wrote on its official blog. “Developing high-quality, safety-relevant assessments remains a challenge, and demand outstrips supply.”

As we’ve highlighted before, AI has a benchmarking problem. The most commonly cited AI benchmarks today do a poor job of capturing how the average person actually uses the systems under test. There are also questions about whether some benchmarks, particularly those released before the dawn of modern generative AI, actually measure what they claim to measure, given their age.

The solution Anthropic proposes is very high-level and more difficult than it seems. It creates challenging benchmarks with a focus on AI security and societal implications through new tools, infrastructure and methods.

The company specifically calls for tests that assess a model’s ability to perform tasks like launching cyberattacks, “enhancing” weapons of mass destruction (e.g., nuclear weapons), and manipulating or misleading humans (e.g., via deepfakes or disinformation). For AI risks related to national security and defense, Anthropic says it is committed to developing some kind of “early warning system” to identify and assess risks, though it doesn’t reveal in the blog post what such a system might entail.

Anthropic also says the new program is intended to support research into benchmarks and “end-to-end” tasks that explore the potential of AI to assist with scientific research, conducting conversations in multiple languages, and reducing deep-seated biases and self-censorship around toxicity.

To accomplish all this, Anthropic is envisioning new platforms that will allow subject matter experts to develop their own evaluations and large-scale model trials involving “thousands” of users. The company says it has hired a full-time coordinator for the program and that it can buy or expand projects that it believes have the potential to scale.

“We offer a range of financing options tailored to the needs and stage of each project,” Anthropic wrote in the post, though an Anthropic spokesperson declined to provide further details on those options. “Teams will have the ability to communicate directly with Anthropic’s domain experts from the frontier red team, fine-tuning, trust and safety, and other relevant teams.”

Anthropic’s efforts to support new AI benchmarks are commendable, assuming of course that it has the money and manpower behind it. But given the company’s commercial ambitions in the AI ​​race, it may be difficult to trust it entirely.

In the blog post, Anthropic is quite transparent about the fact that it wants certain evaluations it funds to be in line with the AI Safety Classifications It developed (with some input from third parties like the nonprofit AI research organization METR). That’s well within the company’s purview. But it could also force applicants to the program to accept definitions of “safe” or “risky” AI that they might not fully agree with.

Some in the AI ​​community are also likely to object to Anthropic's references to “catastrophic” and “misleading” AI risks, such as the risks of nuclear weapons. Many experts say there is little evidence to suggest that AI as we know it will soon, if ever, gain the ability to destroy the world and outsmart humans. Claims of impending “superintelligence” only distract from today’s pressing AI regulatory issues, such as AI’s hallucinatory tendencies, these experts add.

In its post, Anthropic writes that it hopes its program will serve as “a catalyst for progress toward a future in which comprehensive AI evaluation is an industry standard.” That's a mission that the many open, company independent efforts to create better AI benchmarks can identify with. But it remains to be seen whether those efforts are willing to join forces with an AI vendor whose loyalty ultimately lies with shareholders.

Related Posts

The perfect toy for your playful cat is now $20 during Deal Days

Unless your cat is Garfield, you’re not going to entertain them with lasagna while you’re at work. If you left a tray out, you’d come home to a giant mess,…

Go straight to checkout to get this hidden camera and listening device for $40

We live in an age where privacy is truly challenged 24/7. For those who are concerned or curious about whether there are hidden cameras, audio recording listening devices or GPS…

Leave a Reply

Your email address will not be published. Required fields are marked *

You Missed

The perfect toy for your playful cat is now $20 during Deal Days

  • July 4, 2024
The perfect toy for your playful cat is now $20 during Deal Days

For first time in NHL, Jessica Campbell becomes assistant coach with Seattle Kraken : NPR

  • July 4, 2024
For first time in NHL, Jessica Campbell becomes assistant coach with Seattle Kraken : NPR

Hong Kong IPO market expected to improve over the next five years

  • July 4, 2024
Hong Kong IPO market expected to improve over the next five years

Red Sox' Hendriks slowly but surely making a comeback from Tommy John: 'I don't pamper it'

  • July 4, 2024
Red Sox' Hendriks slowly but surely making a comeback from Tommy John: 'I don't pamper it'

Hakeem Jeffries probeert voorlopig de afvalligheid van het Huis van Afgevaardigden tegen te gaan, nu Democraten in paniek raken over Biden

  • July 4, 2024
Hakeem Jeffries probeert voorlopig de afvalligheid van het Huis van Afgevaardigden tegen te gaan, nu Democraten in paniek raken over Biden

Trader Joe's is recalling candles sold nationwide because they pose a safety risk

  • July 4, 2024
Trader Joe's is recalling candles sold nationwide because they pose a safety risk

Copa America Power Rankings: United States and Mexico plummet after early eliminations; Argentina still top

  • July 4, 2024

Influencers have changed the advertising industry. What now?

  • July 4, 2024
Influencers have changed the advertising industry. What now?

Go straight to checkout to get this hidden camera and listening device for $40

  • July 4, 2024
Go straight to checkout to get this hidden camera and listening device for $40

This 'bucket strategy' can lower your taxes in retirement

  • July 4, 2024
This 'bucket strategy' can lower your taxes in retirement

Researchers capture never-before-seen image of gene transcription

  • July 4, 2024
Researchers capture never-before-seen image of gene transcription