Article 51: Classification of GPAI Models with Systemic Risk
Article 51 defines when a general-purpose AI model is classified as having systemic risk. A GPAI model has systemic risk when it has high-impact capabilities evaluated on the basis of appropriate technical tools and methodologies, including indicators and benchmarks—or when designated by a Commission decision (including upon qualified alert from the scientific panel). A GPAI model is presumed to have high-impact capabilities when the cumulative amount of computation used for its training exceeds 10^25 FLOPs (the threshold in Annex XIII), subject to Commission update via delegated acts. Classification triggers additional obligations under Article 55 and Section 2 of Annex XI.
Who does this apply to?
- -Providers of GPAI models that meet or may meet the systemic-risk criteria (especially the compute threshold)
- -The AI Office and the scientific panel for AI (classification and monitoring roles)
- -Downstream providers integrating GPAI models who need to know upstream systemic-risk status
- -Compliance teams tracking Annex XIII thresholds and Commission delegated acts
Scenarios
A new model is trained with cumulative compute exceeding 10^25 FLOPs.
A model is below the compute threshold but the scientific panel issues a qualified alert citing high-impact capabilities demonstrated in real-world deployments.
The Commission updates the Annex XIII threshold via delegated act after advances in training efficiency make 10^25 FLOPs less meaningful as a risk proxy.
What Article 51 does (in plain terms)
Article 51 answers the question: when is a GPAI model considered to have systemic risk? Two routes:
Route 1 — High-impact capabilities (Article 51(1)): A GPAI model is classified as systemic risk when it has high-impact capabilities evaluated using appropriate technical tools and methodologies, including indicators and benchmarks. The AI Office may designate on its own initiative or after a qualified alert from the scientific panel (Article 90).
Route 2 — Compute presumption (Article 51(2)): A GPAI model is presumed to have high-impact capabilities when the cumulative amount of computation used for its training, measured in floating-point operations (FLOPs), is greater than 10^25. The Commission may update this threshold by delegated act under Article 97 based on evolving technological benchmarks.
Providers can rebut the presumption with substantiated arguments through the Article 52 procedure—the threshold is not absolute, but the burden is on the provider.
What triggers when classification happens
Once a GPAI model is classified as systemic risk (by notification or designation):
1. Article 55 additional obligations activate: model evaluation, adversarial testing (red teaming), tracking and reporting serious incidents, ensuring adequate cybersecurity protection. 2. Annex XI Section 2 documentation becomes mandatory: detailed evaluation strategies, red-teaming measures, and system architecture. 3. The model appears on the AI Office's published list of systemic-risk GPAI models. 4. Providers may need to adhere to relevant codes of practice under Article 56 or demonstrate equivalent compliance.
The 10^25 FLOPs threshold — context and limitations
The 10^25 FLOPs presumption was calibrated to frontier models at the time of legislative negotiations. It is a rebuttable presumption, not a hard classification boundary—models below the threshold can still be designated if they demonstrate high-impact capabilities through other evidence.
As training efficiency improves (distillation, architecture advances, data quality), the compute proxy may under- or over-capture risk. The Commission has delegated-act power to adjust the threshold and the Annex XIII criteria to reflect state of the art.
How Article 51 connects to the rest of the Act
- Article 52 — Procedure for classification (notification, AI Office designation, rebuttal, publication).
- Article 53 — Baseline GPAI obligations (apply to all GPAI providers, regardless of systemic risk).
- Article 55 — Additional obligations activated by systemic-risk classification.
- Article 56 — Codes of practice as a compliance pathway for systemic-risk obligations.
- Annex XI — Section 2 documentation triggered by systemic-risk status.
- Annex XIII — Criteria and thresholds for systemic-risk classification (including the 10^25 FLOPs presumption).
- Article 90 — Scientific panel that may trigger qualified alerts for AI Office classification.
- Article 97 — Delegated acts for updating the Annex XIII threshold.
- Article 101 — Penalties for GPAI providers.
- Article 113 — Application dates.
Official wording: Article 51 (English)
1. A general-purpose AI model shall be classified as a general-purpose AI model with systemic risk if it meets any of the following conditions:
(a) it has high impact capabilities evaluated on the basis of appropriate technical tools and methodologies, including indicators and benchmarks;
(b) based on a decision of the Commission, ex officio or following a qualified alert from the scientific panel, it has capabilities or an impact equivalent to those set out in point (a) having regard to the criteria set out in Annex XIII.
2. A general-purpose AI model shall be presumed to have high impact capabilities pursuant to paragraph 1, point (a), when the cumulative amount of computation used for its training measured in floating point operations is greater than 10^25.
3. The Commission shall adopt delegated acts in accordance with Article 97 to amend the thresholds listed in paragraphs 1 and 2 of this Article, as well as to supplement benchmarks and indicators in light of evolving technological developments, such as algorithmic improvements or increased hardware efficiency, when necessary, for these thresholds to reflect the state of the art.
Recitals (preamble) on EUR-Lex
The recitals in the same consolidated AI Act on EUR-Lex contextualise systemic risk, high-impact capabilities, the 10^25 FLOPs calibration, the scientific panel role, and the Commission's delegated-act update mechanism. Use the official preamble on EUR-Lex—do not rely on unofficial recital lists without checking sequence and wording against the authentic text.
Compliance checklist
- Calculate cumulative training compute (FLOPs) for each model and compare against the 10^25 threshold.
- Monitor Commission delegated acts for threshold updates to Annex XIII.
- Set up internal triggers to notify the AI Office without delay if the threshold is met (Article 52).
- If below the threshold: document why systemic-risk classification does not apply (high-impact capabilities assessment).
- If designated: immediately operationalise Article 55 obligations and Annex XI Section 2 documentation.
- Track the AI Office's published list of systemic-risk models for upstream dependencies.
Assess your GPAI model against systemic-risk criteria—free assessment.
Start Free AssessmentRelated Articles
Article 52: Procedure for Systemic Risk Classification of GPAI Models
Article 53: Obligations for Providers of General-Purpose AI Models
Article 54: Authorised Representatives of Providers of GPAI Models
Article 55: Obligations for Providers of GPAI Models with Systemic Risk
Article 56: Codes of Practice for GPAI Models
Article 90: Penalties
Article 97: Exercise of the Delegation
Article 101: Fines for Providers of General-Purpose AI Models
Article 113: Entry into Force and Application Dates
Annex XI: Technical Documentation for Providers of General-Purpose AI Models
Annex XIII: Criteria for Classification of GPAI Models with Systemic Risk
Article 50: Transparency Obligations for Providers and Deployers of Certain AI Systems
Article 6: Classification Rules for High-Risk Systems
Article 11: Technical Documentation
Related annexes
- Annex XI — GPAI technical documentation (Section 2 triggered by systemic-risk classification)
- Annex XII — Transparency information
- Annex XIII — Criteria for systemic-risk classification (including 10^25 FLOPs threshold)
Frequently asked questions
Is the 10^25 FLOPs threshold permanent?
No. The Commission can update it via delegated act under Article 97 based on evolving technological benchmarks and state of the art.
Can a model be systemic risk without crossing the compute threshold?
Yes. Article 51(1)(b) allows the Commission to designate based on equivalent capabilities or impact, using Annex XIII criteria, even if the model is below 10^25 FLOPs.
Do open-source models avoid systemic-risk classification?
No. Open-source status does not affect Article 51 classification. If an open-weights model meets the criteria, it is systemic risk and the Article 53(2) open-source exception falls away.
What is the scientific panel?
An independent expert body under Article 90 that can issue qualified alerts to the AI Office about GPAI models potentially meeting systemic-risk criteria, triggering the Article 52 classification procedure.
Who is responsible when we fine-tune a base model?
Fine-tuning can create a new provider relationship. Whether the fine-tuned model separately triggers Article 51 depends on the cumulative compute and capabilities of the resulting model. Map your facts against the definitions.