Meta’s New Report Shows How to Prevent ‘Catastrophic Risks’ from AI

2 months ago 27
  • Published on February 4, 2025
  • In AI News

The framework is structured around a three-stage process: anticipating risks, evaluating and mitigating risks, and deciding whether to release, restrict, or halt the model. 

Meta LLama 4

Meta, the company behind the open-source Llama family of AI models, unveiled a new report on Monday titled ‘The Frontier AI Framework’. 

This report outlines the company’s approach to developing general-purpose AI models by evaluating and mitigating their “catastrophic risks”. It focuses on large-scale risks, such as cybersecurity threats and risks from chemical and biological weapons. 

“By prioritising these areas, we can work to protect national security while promoting innovation,” read a blog post from the company announcing the report. 

The framework is structured around a three-stage process: anticipating risks, evaluating and mitigating risks, and deciding whether the model should be released, restricted, or halted. 

The report contains risk assessments from various experts in multiple disciplines, including engineering, product management, compliance and privacy, legal and policy, and other company leaders. 

If the risks are deemed “critical”, the framework suggests stopping the development of the model and restricting access to a small number of experts with security protections. 

If the risks are deemed “high”, the model should not be released but can be accessed by a core research team with security protections. Lastly, if the risks are “moderate”, the model can be released as per the framework. 

It is recommended that you read the full report to understand each of these risk levels in detail. 

“While it’s not possible to entirely eliminate risk if we want this AI to be a net positive for society, we believe it’s important to work internally and, where appropriate, with governments and outside experts to take steps to anticipate and mitigate severe risks that it may present,” the report stated. 

Multiple companies building AI models have unveiled such frameworks over the years. For instance, Anthropic’s ‘Responsible Scaling Policy’ provides technical and organisational protocols that the company is adopting to “manage risks of developing increasingly capable AI systems”. 

Anthropic’s policy defines four safety levels. The higher the safety level, the larger the model’s capability, which increases security and safety measures. 

In October last year, the company said it was required to upgrade its security measures to an AI safety level of 3, which is the penultimate level of safety. This level suggests that the model’s capability poses a “significantly higher risk”. 

Similarly, OpenAI has a charter that outlines its mission to develop and deploy AI models safely. Furthermore, the company regularly releases a system card for all of its models, which outlines its work regarding safety and security before releasing any new variant. 

Picture of Supreeth Koundinya

Supreeth Koundinya

Supreeth is an engineering graduate who is curious about the world of artificial intelligence and loves to write stories on how it is solving problems and shaping the future of humanity.

Association of Data Scientists

GenAI Corporate Training Programs

India's Biggest Developers Summit

February 5 – 7, 2025 | Nimhans Convention Center, Bangalore

Download the easiest way to
stay informed

Ai for Bharat

The Need for Building AI for Bharat

Mohit Pandey

“India’s 5,000-year-old civilisation holds a wealth of valuable information that can be used to build better AI solutions for the population.”

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

February 5 – 7, 2025 | Nimhans Convention Center, Bangalore

Rising 2025 | DE&I in Tech & AI

Mar 20 and 21, 2025 | 📍 J N Tata Auditorium, Bengaluru

Data Engineering Summit 2025

15-16 May, 2025 | 📍 Taj Yeshwantpur, Bengaluru, India

AI Startups Conference.
April 25 / Hotel Radisson Blu / Bangalore, India

17-19 September, 2025 | 📍KTPO, Whitefield, Bangalore, India

MachineCon GCC Summit 2025

19-20th June 2025 | Bangalore

discord icon

Our Discord Community for AI Ecosystem.

Read Entire Article