Procedural Kits: When Component Evidence Fails at System Level

Hatem Rabeh

Written by HATEM RABEH, MD, MSc Ing

Your Clinical Evaluation Expert And Partner

in
S

A sterile procedural kit receives a major deficiency. The notified body accepts that every single component is clinically supported. Yet they reject the system-level evaluation. The manufacturer protests: if each part is safe and effective, why isn’t the combination? This tension appears in almost every procedural kit submission I review.

The confusion is understandable. The manufacturer invested in clinical data for each component. Each screw, each plate, each guide is covered by literature, by equivalence, or by legacy data. The dossier looks complete at component level.

But the notified body is not questioning the components. They are questioning what happens when these components work together in a defined procedure.

What MDR Actually Requires for Systems

MDR Article 61 and Annex XIV do not distinguish between single devices and systems. The requirement is the same: demonstrate safety and performance for the intended purpose.

For a procedural kit, the intended purpose is not the sum of individual component purposes. It is the execution of a specific clinical procedure with a defined workflow.

A hip replacement kit is not just a collection of implants and instruments. It is a system designed to enable a surgical workflow. That workflow creates clinical outcomes that cannot be predicted from component data alone.

Key Insight
The clinical evaluation must address whether the system enables the intended procedure safely and effectively. Component-level evidence is necessary but not sufficient.

This is not a theoretical distinction. It shapes what data you collect and how you structure your clinical evaluation report.

Where Component Evidence Stops

I review many CERs where every section is organized by component. Each instrument gets its own literature review. Each implant gets its own equivalence analysis. The structure mirrors the bill of materials.

This approach satisfies one part of the clinical evidence requirement. It shows that individual components meet safety and performance standards.

But it leaves critical questions unanswered.

What happens at the interfaces between components? Does instrument A reliably position implant B? Does guide C consistently align with drill D?

What about the procedural workflow? Does the kit enable the procedure as intended? Are there steps where user error is likely? Are there compatibility risks between components that work individually but create problems in combination?

These questions exist at system level. They require system-level evidence.

Common Deficiency
Manufacturers list component-level literature and claim it covers the system. Reviewers reject this because the literature describes individual device performance, not procedural outcomes with the specific kit.

The deficiency appears predictable once you see it from the reviewer’s perspective. They are not asking if screws work. They are asking if this particular combination of screws, plates, guides, and instruments enables the intended surgical procedure safely.

The Interface Risk Problem

Every procedural kit creates interfaces. Physical interfaces where components connect. Procedural interfaces where one component’s output becomes another component’s input.

Interface risk is where system-level problems emerge.

A drill guide that works perfectly in isolation may misalign when combined with a specific drill bit under surgical conditions. An implant that performs well individually may migrate when placed with instruments that apply forces at specific angles.

These are not theoretical concerns. I have reviewed kits where interface failures led to serious incidents despite each component meeting its individual specifications.

The clinical evaluation must address these interfaces explicitly. It cannot assume that good component data equals good system performance.

What Interface Evidence Looks Like

System-level evidence often comes from clinical data that uses the complete kit. Published case series where surgeons used all components together. Registry data that tracks outcomes for the specific procedural system.

If your kit is substantially equivalent to a predicate, the equivalence analysis must cover system-level characteristics. Not just component-to-component comparison, but system-to-system comparison.

If you introduce a new kit without a valid predicate, you need clinical investigation data at system level. This is where many manufacturers underestimate the evidence gap.

You cannot submit component equivalence analyses and expect them to bridge a system-level evidence gap. The logic does not transfer.

Key Insight
When your kit has no equivalent system on the market, component equivalence claims do not reduce your clinical investigation burden for system-level performance.

The Workflow Evaluation Gap

Many CERs describe the kit but do not evaluate the workflow. They list what components do. They do not explain how the procedure unfolds or where risks concentrate during execution.

This creates a blind spot that reviewers notice immediately.

A procedural kit enables a sequence of actions. Each action depends on prior actions. Each step creates conditions for the next step. The workflow itself becomes part of the device risk profile.

If step three requires precise positioning that depends on step two, and step two is user-dependent, the system inherits usability risk. Your clinical evaluation must address this.

If the kit is used in emergency settings where time pressure increases error probability, the system must be evaluated under those conditions. Component-level bench testing under ideal conditions does not cover this.

Usability Evidence at System Level

Usability testing for procedural kits must evaluate the complete workflow. Not individual component handling, but the entire procedure from kit opening to completion.

This data becomes part of your clinical evidence. It informs whether users can safely execute the intended procedure with your system.

I see many manufacturers treat usability as a separate compliance exercise. They complete the usability file but do not integrate findings into the clinical evaluation. This creates a disconnect that reviewers challenge.

Your CER must reference usability findings. It must explain how workflow observations inform clinical risk assessment. It must show that the system was evaluated as users will actually deploy it.

Common Deficiency
CER describes component safety profiles but does not discuss procedural failure modes. Reviewers cannot assess whether system-level risks are adequately characterized and mitigated.

How to Structure System-Level Evaluation

Your clinical evaluation report should have dedicated sections that address the system as a functional unit.

Start with a clear description of the intended procedure. Describe the workflow step by step. Identify where each component functions and where interfaces occur.

Then evaluate clinical data at system level. If published literature exists for procedures using comparable kits, analyze it for system-level outcomes. Look for complication patterns that emerge from kit use, not just component failure.

If you claim equivalence to a predicate system, your equivalence analysis must compare procedural workflows. Show that your kit enables the same procedure with the same risk profile. Demonstrate that interface characteristics are comparable.

If you cannot establish equivalence at system level, acknowledge the evidence gap. Explain how you will generate system-level clinical data. This usually means clinical investigation or structured post-market data collection.

The PMCF Angle

Post-market clinical follow-up for procedural kits must collect system-level data. You cannot monitor components in isolation and claim you are monitoring the system.

Your PMCF plan should specify outcomes that reflect procedural success. Revision rates, complication patterns, procedural time, user-reported difficulties.

These are system-level indicators. They reveal whether your kit performs as intended in real clinical settings.

Many PMCF plans I review focus on implant survival or instrument durability. These are component-level metrics. They do not answer whether the system enables safe and effective procedures.

Key Insight
Your PMCF must track procedural outcomes, not just component performance. The data should confirm that clinical use matches the intended workflow and that risks remain acceptable at system level.

What Reviewers Actually Check

When a notified body reviews a procedural kit CER, they look for evidence that the manufacturer understands the system as a functional whole.

They check whether the literature search strategy captured system-level data. Did you search for procedure names, or only component types?

They examine whether the equivalence analysis considers the procedural workflow. Did you compare how procedures unfold, or only how components compare in isolation?

They assess whether risk analysis addresses interfaces and workflow dependencies. Did you identify where system-level failures could occur?

They verify that usability findings inform the clinical evaluation. Did you integrate user feedback about procedural execution?

If any of these elements are missing or weakly developed, the CER receives a major deficiency. The reviewer cannot conclude that the system is safe and effective based only on component data.

This is not regulatory overreach. It is a logical requirement. Your intended use is procedural. Your evidence must address procedural performance.

A Practical Example

Consider a spinal fusion kit with screws, rods, and instrumentation. Each component has extensive clinical literature. Screws have decades of performance data. Rods are well-characterized. Instruments are similar to predicate devices.

The manufacturer submits a CER organized by component. Each section is strong. The notified body issues a major deficiency.

Why?

Because the kit introduces a novel rod-screw connection mechanism. This mechanism changes how forces transfer during fusion. It affects procedural workflow because it requires a different insertion sequence.

The component-level data does not address this. Screw literature assumes standard connections. Rod literature assumes standard insertion techniques.

The system creates a new clinical scenario that is not covered by legacy component data. The manufacturer needs system-level evidence that addresses the novel connection mechanism under actual surgical conditions.

This is where many submissions fail. The manufacturer assumed that strong component data would carry the evaluation. But the system-level novelty created an evidence gap that component data could not bridge.

Common Deficiency
Manufacturer claims equivalence at component level but introduces system-level modifications. Reviewers identify the gap and require clinical investigation for the novel system characteristics.

Moving Forward

If you are preparing a clinical evaluation for a procedural kit, assess your evidence at system level first. Before you compile component data, ask whether you have evidence that the complete kit enables the intended procedure safely.

If the answer is uncertain, you have an evidence gap that requires attention. Component equivalence will not close it. More instrument literature will not close it.

You need data that reflects how your system performs in clinical use. This might come from published literature on comparable procedural systems. It might come from clinical investigation. It might come from structured PMCF data if you can justify a stepwise evidence generation approach.

The key is recognizing that system-level evaluation is not optional. It is not an add-on to component evaluation. It is a foundational requirement that shapes your entire clinical evidence strategy.

Notified bodies are consistent on this point. They will accept strong component data as part of the evidence base. But they will not accept it as the complete evidence base for a procedural system.

The sooner you build system-level evidence into your development plan, the less likely you are to face major deficiencies late in the submission process.

Peace,
Hatem
Clinical Evaluation Expert for Medical Devices
Follow me for more insights and practical advice.

Frequently Asked Questions

What is a Clinical Evaluation Report (CER)?

A CER is a mandatory document under MDR 2017/745 that demonstrates the safety and performance of a medical device through systematic analysis of clinical data. It must be updated throughout the device lifecycle based on PMCF findings.

How often should the CER be updated?

The CER should be updated whenever significant new clinical data becomes available, after PMCF activities, when there are changes to the device or intended purpose, and at minimum during annual reviews as part of post-market surveillance.

What causes CER rejection by Notified Bodies?

Common reasons include inadequate equivalence demonstration, insufficient clinical data for claims, poorly structured SOTA analysis, missing gap analysis, and lack of clear benefit-risk determination. Structure and logical flow are as important as the data itself.

Which MDCG guidance documents are most relevant for clinical evaluation?

Key documents include MDCG 2020-5 (Equivalence), MDCG 2020-6 (Sufficient Clinical Evidence), MDCG 2020-13 (CEAR Template), MDCG 2020-7 (PMCF Plan), and MDCG 2020-8 (PMCF Evaluation Report).

Need Expert Help with Your Clinical Evaluation?

Get personalized guidance on MDR compliance, CER writing, and Notified Body preparation.

Peace, Hatem

Your Clinical Evaluation Partner

Follow me for more insights and practical advice.

References:
– Regulation (EU) 2017/745 (MDR), Article 61, Annex XIV
– MDCG 2020-5: Clinical Evaluation Assessment Report Template
– MDCG 2020-6: Sufficient Clinical Evidence for Legacy Devices

Deepen Your Knowledge

Read Complete Guide to Clinical Evaluation under EU MDR for a comprehensive overview of clinical evaluation under EU MDR 2017/745.