MDR, IVDR, and the AI Act: What MDCG 2025-6 Means for Medical Device Software

Written by HATEM RABEH, MD, MSc ing

Your Clinical Evaluation Expert And Partner

Introduction

The European Union’s release of MDCG 2025-6 has reshaped the regulatory landscape for medical device software—and if your product includes any artificial intelligence, this isn’t just an update. It’s a survival checklist.

Gone are the days when medical device software (MDSW) only had to navigate MDR or IVDR. Enter MDAI: Medical Device Artificial Intelligence—a new label for machine-learning-enabled devices now potentially subject to the high-risk AI category under the AIA (Artificial Intelligence Act).

Whether you manufacture an AI-powered wearable, health-tracking mobile app, or diagnostic algorithm embedded in imaging software, this framework impacts your clinical evaluation strategy, technical documentation, risk management, and post-market responsibilities.

The transition is not optional. It’s mandated by August 2027.


What is MDCG 2025-6?

MDCG 2025-6 is a 27-page FAQ—not a guidance document—that clarifies how to reconcile compliance between the MDR/IVDR regulations and the new AI Act. While the tone is FAQ-like, its content is structured more like a checklist for implementation.

And rightly so. The regulatory weight behind this document is undeniable.

Think of it as a Rosetta Stone for navigating AI’s role in regulated medical devices in Europe. But unlike most guidance, this one comes with hard timelines, practical implementation expectations, and layers of accountability that go beyond CE marking.


MDAI: A New Class Emerges

If your software includes machine learning and is regulated under MDR or IVDR, it’s now considered Medical Device Artificial Intelligence (MDAI).

This reclassification does more than change the name. It escalates the regulatory burden, moving your device into the AIA’s “high-risk” category. And with that come requirements around:

  • Ethical risk

  • Fundamental rights

  • Bias monitoring

  • Algorithmic transparency

Even if you’ve already cleared MDR/IVDR hurdles, you’re not done yet. The AIA adds a second compliance layer that dives deep into how your AI learns, evolves, and performs over time.


Who is Affected by MDCG 2025-6?

Short answer? Almost everyone working in digital health:

  • Manufacturers of AI-based medical devices

  • Software developers creating modules or apps for diagnosis, monitoring, or therapy

  • Companies producing accessories for MDAI or Annex XVI devices

  • Deployment companies running these devices in clinical or real-world settings


High-Risk Classification under AIA

Under the AIA, any algorithm that influences health decisions—especially if it evolves after deployment—can be classified as high-risk. This includes:

  • ML-powered diagnostics

  • Triage algorithms

  • Treatment decision support tools

  • Real-time monitoring platforms

If that’s you, prepare for profound changes in compliance routines.


Clinical Evaluation Under Pressure

The MDR requires clinical evaluation. But with the AIA added, that evaluation must now also show:

  • Explainability: What the algorithm does and how it does it

  • Dataset Fairness: No bias, no blind spots

  • Human Oversight: Yes, even a “stop” button counts

  • Ongoing Data Governance: Document, validate, test, and monitor

And don’t forget: if your device continues learning post-market, you’ll need Predefined Change Control Plans (PCCPs). No exceptions.


Technical Documentation Now Has an AI Flavor

Gone are the days of static PDFs. Your technical file must now evolve alongside your AI. That includes:

  • AI-specific risk assessments

  • Human-rights impact analysis

  • Performance metrics aligned with intended purpose

  • Lifecycle control documents

  • Logging and behavioral tracking

Even better? MDCG 2025-6 allows you to embed these into your MDR/IVDR files—but only if structured correctly.


Timeline for Compliance

✅ 2025

  • Start aligning your QMS with AIA

  • Audit your clinical data for AI relevance

✅ 2026

  • Build PCCPs and QMS integrations

  • Update labeling and user roles

✅ 2027 (August)

  • All high-risk MDAI must fully comply with AIA

  • Post-market monitoring becomes law, not best practice


Key Terms: ‘User’ vs. ‘Deployer’

  • MDR: Defines a user as a professional or patient.

  • AIA: Defines a deployer as the entity using the AI system under their control.

If your documentation treats them interchangeably, auditors will notice. Define both clearly in your QMS and conformity declarations.


FAQs

What does MDCG 2025-6 require me to do?
It mandates that manufacturers embed AIA requirements into their MDR/IVDR documentation and processes.

Does the AIA apply to all medical devices?
Only those considered MDAI—i.e., medical devices using AI, especially machine learning.

Is clinical evaluation still mandatory?
Yes, and it now requires explainability, bias control, and oversight documentation.

When is the compliance deadline?
August 2027, with preparation starting immediately.

Can I reuse my MDR files for AIA compliance?
Yes, if you integrate AIA layers completely and traceably.

What happens post-2027?
Significant changes trigger a fresh conformity assessment, even if CE marking was previously granted.


Final Thoughts

MDCG 2025-6 changes the game. This isn’t about adding a page or two to your technical documentation. It’s about reinventing your compliance architecture to reflect a new reality: AI is here, and it’s regulated.

Start today. Build smart. And if your device learns, you’d better make sure your team does too.

PS:  For more information, subscribe to my newsletter and get access to exclusive content, private insights, and expert guidance on MDR compliance and CE marking: Subscribe Here

✌️ Peace,
Hatem Rabeh, MD, MSc Ing
Your Clinical Evaluation Expert & Partner
Follow me for more insights and practical advice!