Monthly News and Updates (March 2026)

by | Apr 18, 2026 | AI News & Updates | 0 comments

March 2026 did not introduce any major or immediately actionable regulatory changes for AI in healthcare across Canada, the European Union, or the United Kingdom. Activity in these regions remained relatively stable, with no significant new guidance, enforcement shifts, or legislative developments directly impacting AI-enabled healthcare systems during this period.

In contrast, the United States saw continued regulatory engagement, particularly through stakeholder discussions and clarification efforts led by the U.S. Food and Drug Administration (FDA). These updates focused on refining existing frameworks rather than introducing entirely new regulatory structures, with implications for AI-enabled clinical decision-making tools and medical device oversight processes.

 
 
 

United States

 

1) FDA – Clinical Decision Support (CDS) Software Guidance Engagement (March 11, 2026)

Category: Regulatory clarification / implementation guidance

On March 11, 2026, the FDA held a formal town hall session to discuss its final guidance on Clinical Decision Support (CDS) software. The session focused on clarifying how the agency interprets and applies regulatory boundaries between software functions that qualify as medical devices and those that do not.

The discussion centered on the distinction between:

  • Non-device CDS software, which is excluded from FDA medical device regulation, and
  • Device CDS software, which remains subject to regulatory oversight

 

The FDA reiterated that software may be excluded from regulation if it meets specific criteria, particularly where:

  • The software is intended to support or provide recommendations to healthcare professionals (not replace them)
  • The basis for the recommendations is transparent and can be independently reviewed by the clinician
  • The clinician is able to understand how the software arrives at its output

 

Conversely, software that:

  • Produces outputs that cannot be independently evaluated by the user
  • Relies on complex or non-transparent models (including certain AI/ML approaches)
  • Influences clinical decisions without sufficient explainability

may fall within the definition of a regulated medical device.

 

The town hall also emphasized how the FDA interprets “independent review” and “transparency,” which are key criteria for determining whether CDS software remains outside regulatory scope.

 

How it applies to AI in Healthcare:

This guidance is particularly relevant for AI-enabled clinical tools because many AI systems:

  • Use complex models that are not inherently interpretable
  • Generate probabilistic or pattern-based outputs rather than rule-based recommendations

 

As a result, AI-based CDS tools are more likely to fall into the regulated category unless they are specifically designed to:

  • Provide explainable outputs
  • Allow clinicians to independently validate the reasoning behind recommendations

 

This reinforces a key regulatory principle:
Explainability and human interpretability are central to determining regulatory classification.

 

For developers, this creates a practical design implication — systems that prioritize transparency and clinician oversight may reduce regulatory burden, while more autonomous or opaque systems are likely to face full medical device requirements.

FDA – Town Hall – Clinical Decision Support Software, Final Guidance

 
 

2) FDA – MDUFA VI Reauthorization Stakeholder Meeting (March 11, 2026)

Category: Regulatory process / policy development

 

Also on March 11, 2026, the FDA conducted a stakeholder meeting as part of the reauthorization process for the Medical Device User Fee Amendments (MDUFA VI). This process involves gathering feedback from industry, regulators, and other stakeholders on how the medical device regulatory framework should evolve for the next authorization cycle.

 

The meeting addressed several aspects of the medical device regulatory system, including:

  • Review timelines and performance goals
  • Pre-market submission processes
  • Post-market oversight and monitoring
  • Resource allocation and user fee structures

 

Stakeholders provided input on how the FDA can improve efficiency, predictability, and transparency in its review processes.

While the discussion was not limited to AI, AI-enabled medical devices were implicitly included within scope, as they are regulated under the same frameworks. Therefore, any changes to:

  • Review timelines
  • Evidence expectations
  • Regulatory processes

will directly affect AI-based healthcare technologies.

 

How it applies to AI in Healthcare

For AI in healthcare, the MDUFA VI process is significant because it shapes the operational environment in which AI medical devices are reviewed and monitored.

 

Key implications include:

  • Potential changes to review timelines, which could affect time-to-market for AI tools
  • Evolving expectations for evidence, particularly around real-world performance and post-market monitoring
  • Increased emphasis on regulatory predictability, which is critical for AI developers navigating complex approval pathways

 

Although no immediate regulatory changes were enacted, the stakeholder feedback process indicates that:

  • AI-related regulatory challenges (e.g., lifecycle updates, adaptive systems, performance monitoring) are being considered within broader device policy discussions
  • Future iterations of MDUFA may indirectly shape how AI systems are evaluated and maintained over time

This reflects a broader trend where AI is not always regulated through standalone frameworks, but rather through adaptation of existing medical device systems.

FDA – Industry MDUFA VI Reauthorization Meeting

 
 

Cross-Cutting Themes

 

1) Emphasis on Clarification Over New Regulation

Regulators continue to refine and interpret existing frameworks rather than introduce entirely new AI-specific rules.

2) Central Role of Explainability

The CDS guidance reinforces that transparency and interpretability are key determinants of regulatory classification, particularly for AI systems.

3) Integration of AI into Existing Device Frameworks

AI is being governed within established medical device regulatory structures, rather than through separate, AI-specific regimes.

4) Importance of Process Efficiency and Predictability

Through MDUFA discussions, there is a clear focus on improving regulatory timelines and consistency, which is critical for emerging technologies like AI.

5) Lifecycle Oversight Considerations

Although not explicitly new, discussions around post-market monitoring and performance suggest continued attention to ongoing oversight, which is especially relevant for adaptive AI systems.

 
 

Key Considerations for Regulatory Alignment

 

For Founders & Business Owners

  • Design decisions influence regulatory classification
    AI systems that prioritize explainability and clinician interpretability may reduce regulatory complexity compared to opaque models.
  • Expect full medical device oversight for most AI-driven clinical tools
    Particularly where models are complex or outputs are not easily explainable.
  • Plan for regulatory timelines early
    Engagements like MDUFA signal that review timelines and expectations remain a critical factor in go-to-market strategy.
  • Monitor evolving FDA expectations
    Even without new laws, guidance interpretation (e.g., CDS criteria) can materially impact product classification and compliance requirements.
  • Account for post-market responsibilities
    AI systems may require ongoing monitoring, updates, and performance validation after deployment.

 

For Compliance & Regulatory Specialists

  • Carefully assess CDS classification criteria
    Particular attention should be given to:
    • Transparency of outputs
    • Ability for independent clinical review
    • Degree of reliance on AI/ML models
  • Document explainability and user interpretability
    Evidence demonstrating how clinicians can understand and validate outputs will be critical for classification decisions.
  • Prepare for evolving review expectations
    MDUFA discussions suggest continued refinement in:
    • Evidence requirements
    • Review processes
    • Performance evaluation standards
  • Strengthen post-market surveillance frameworks
    Even in the absence of new rules, regulatory focus continues to include real-world performance monitoring.
  • Align with broader device regulatory processes
    AI compliance should be integrated into existing medical device regulatory strategies rather than treated as a standalone domain.

Disclaimer: This checklist is provided for general informational purposes only and does not constitute legal, regulatory, or professional advice; organizations should consult with their legal and compliance departments to ensure adherence to specific jurisdictional requirements.

 
 

Sources

Written by Grigorii Kochetov

Cybersecurity Researcher at AI Healthcare Compliance

Read more

Practical impacts of using AI in Healthcare

Practical impacts of using AI in Healthcare

Artificial Intelligence (AI) is transforming healthcare systems globally - enhancing diagnostics, improving patient outcomes, optimizing workflows, and reducing costs. However, its adoption also brings challenges around data integrity, equity, and ethical use. Below...

read more
Monthly News and Updates (February 2026)

Monthly News and Updates (February 2026)

During February 2026, governments and regulators across Canada, the United States, and Europe advanced regulatory and governance measures directly affecting AI in healthcare. Key themes included quality system harmonisation, acceleration pathways for digital health...

read more
Compliance Overlap: HIPAA, GDPR, PIPEDA, PHIPA

Compliance Overlap: HIPAA, GDPR, PIPEDA, PHIPA

As healthcare AI systems increasingly process cross-border data, compliance is no longer about satisfying a single statute. It requires operating within overlapping regulatory frameworks that share principles but diverge in structure, scope, and enforcement. This...

read more
Monthly News and Updates (January 2026)

Monthly News and Updates (January 2026)

Editorial Update: Moving to a Monthly Schedule   To ensure we provide the most robust and actionable compliance intelligence for the healthcare AI sector, we are transitioning from weekly to monthly updates. This allows us to focus on high-impact regulatory...

read more
Weekly News and Updates (Jan 12-16, 2026)

Weekly News and Updates (Jan 12-16, 2026)

This week (January 12–16, 2026) marked a pivotal shift in AI healthcare regulation globally, characterized by the formalization of oversight and international harmonization. Key highlights include the joint FDA-EMA guiding principles for AI in drug development,...

read more
Weekly News and Updates (Jan 1-9, 2026)

Weekly News and Updates (Jan 1-9, 2026)

Between 1st and 9th January 2026, the first full week of the year marks a significant shift from theoretical frameworks to operational infrastructure in AI healthcare governance. Key developments include the UK’s closing of its “AI Growth Lab” consultation, the FDA’s...

read more
Weekly News and Updates (Dec 12 – 19, 2025)

Weekly News and Updates (Dec 12 – 19, 2025)

Between 12–19 December 2025, the regulatory landscape for AI in healthcare shifted decisively toward national-level consolidation and operational security: the U.S. White House issued a landmark Executive Order to centralize AI policy and preempt state-level...

read more
Weekly News and Updates (Nov 22 – 28, 2025)

Weekly News and Updates (Nov 22 – 28, 2025)

Between 22–28 November 2025, global regulators accelerated the shift from high-level principles to mandatory operational controls, particularly in Canada, which launched its first public AI Register detailing hundreds of government AI systems. The EU continued...

read more
Weekly News and Updates (Nov 8 – 21, 2025)

Weekly News and Updates (Nov 8 – 21, 2025)

Between 8-21 November 2025 regulators and international bodies emphasised moving from principles to practice: the EU launched COMPASS-AI to operationalise safe clinical AI; the UK (MHRA) published AI Airlock pilot outputs and announced AI drug-safety projects; the FDA...

read more
Prohibited AI Systems Under the EU AI Act

Prohibited AI Systems Under the EU AI Act

The European Union’s Artificial Intelligence Act (EU AI Act) establishes the world’s first comprehensive legal framework for governing artificial intelligence. It divides AI systems into four categories based on their potential impact on safety and fundamental rights...

read more
Weekly News and Updates (Sept 19–25, 2025)

Weekly News and Updates (Sept 19–25, 2025)

This post will begin our new weekly updates that will cover the most recent developments in AI governance and regulations, with a particular focus on how these changes affect AI in healthcare. We begin with updates from Canada — including privacy enforcement actions,...

read more