Researchers Have Ranked AI Models Based on Risk—and Found a Wild Range

Full Article
Researchers Have Ranked AI Models Based on Risk—and Found a Wild Range

Researchers have developed a comprehensive taxonomy of AI risks, focusing on the legal, ethical, and regulatory compliance issues of various AI models. This effort, led by Bo Li and collaborators, includes a benchmark called AIR-Bench 2024, which evaluates the safety of large language models. The analysis reveals significant discrepancies between government regulations and company policies regarding AI safety.

The findings indicate that while some models, like Anthropics Claude 3 Opus, perform well in avoiding cybersecurity threats, others, such as Databricks' DBRX Instruct, scored poorly across the board. This highlights the need for companies to prioritize safety features in their AI models, especially as they deploy these technologies in sensitive areas like customer service. The ongoing evolution of AI necessitates continuous updates to risk assessments and safety measures.

• Researchers created a benchmark to assess AI model safety and compliance.

• Government regulations on AI are less comprehensive than company policies.

Key AI Terms Mentioned in this Article

AI Risk Taxonomy

This taxonomy helps in understanding the potential legal and ethical issues that AI technologies may pose.

AIR-Bench 2024

It provides insights into how different models perform in terms of specific risks.

Large Language Models (LLMs)

The performance of these models is critical in applications such as customer service and content generation.

Companies Mentioned in this Article

Anthropic

Its Claude 3 Opus model is noted for effectively avoiding cybersecurity threats.

Databricks

Its DBRX Instruct model was highlighted for scoring poorly in safety assessments.

Get Email Alerts for AI News

By creating an email alert, you agree to AIleap's Terms of Service and Privacy Policy. You can pause or unsubscribe from email alerts at any time.

Latest Articles

Alphabet's AI drug discovery platform Isomorphic Labs raises $600M from Thrive
TechCrunch 1month

Isomorphic Labs, the AI drug discovery platform that was spun out of Google's DeepMind in 2021, has raised external capital for the first time. The $600

AI In Education - Up-level Your Teaching With AI By Cloning Yourself
Forbes 1month

How to level up your teaching with AI. Discover how to use clones and GPTs in your classroom—personalized AI teaching is the future.

Trump's Third Term - How AI Can Help To Overthrow The US Government
Forbes 1month

Trump's Third Term? AI already knows how this can be done. A study shows how OpenAI, Grok, DeepSeek & Google outline ways to dismantle U.S. democracy.

Sam Altman Says OpenAI Will Release an 'Open Weight' AI Model This Summer
Wired 1month

Sam Altman today revealed that OpenAI will release an open weight artificial intelligence model in the coming months. "We are excited to release a powerful new open-weight language model with reasoning in the coming months," Altman wrote on X.

Popular Topics