Skip to content Skip to footer

How HTA Agencies Are Evaluating AI in Evidence Generation

As AI rapidly reshapes the evidence landscape, health technology assessment (HTA) bodies are beginning to define clear expectations for its use in submissions. At CapeStart, we work closely with life sciences clients navigating these changes—especially those integrating GenAI into literature reviews, RWE generation, economic modeling, Joint Clinical Assessment (JCA) support, Clinical Evaluation Reports (CERs), Patient Reported Outcomes (PROs), and Global Value Dossiers (GVDs). Whether you’re preparing a GVD or submitting to agencies like NICE, understanding what HTA bodies expect from AI-generated content is critical. This guide offers an overview of how HTA agencies are considering AI for evidence generation.

HTA bodies are increasingly acknowledging the potential of AI—particularly generative AI—but emphasize rigorous standards around transparency, validation, and human oversight.

Understanding the Role of AI in Evidence Generation

To consider the question of how HTAs are evaluating AI in evidence generation, it’s helpful to consider the many ways AI can enhance evidence generation. Key examples include:

  • Automated Literature Reviews: Natural language processing (NLP) can rapidly sift through vast scientific literature.
  • Real-World Data Analysis: Machine learning models can detect patterns in electronic health records, claims databases, and registries.
  • Predictive Modeling: AI can estimate outcomes for subpopulations or predict long-term effects where trial data is limited.

By synthesizing mass amounts of data and information, AI has the power to advance many critical functions of life science work. For these functions, there are often high-demand, PhD-level resources needed to conduct the basic workflow of the task, providing their vital professional judgment to ensure the exercise is thoughtful and rigorous. However, these high-demand resources are also mired in tedious document gathering, mass assessment, and onerous summarization, which could be measurably more efficient with the help of AI. For HTA agencies, the question isn’t whether AI is useful, it’s whether AI-derived evidence is trustworthy enough to inform reimbursement and policy decisions.

AI in Evidence Generation

NICE Leadership: Clear Guidance with Necessary Safeguards 

In August 2024, NICE published a position statement outlining how AI should be used in evidence generation and reporting. The guidance underscores several key principles:

Augmentation, not replacement of human expertise, critical decisions and oversight remain human-led.

Transparency is essential: Submissions must declare AI use, explain methods, and justify why AI added value over traditional methods.

Rigorous validation: Any AI used for treatment effect estimation must undergo sensitivity analyses, external checks, and triangulation with existing clinical evidence. Use trusted frameworks such as PALISADE, TRIPOD+AI, LIME, and SHAP to support explainability and ethical deployment.

Early engagement recommended: Organizations are encouraged to discuss AI approaches with NICE early in the development process. 

Other HTAs: Watching, Waiting, Adopting NICE’s Principles

Agencies like CADTH (Canada), HAS (France), G‑BA/IQWiG (Germany), SMC (Scotland), and PBAC (Australia) have not issued formal AI policies—but researchers and industry reports indicate early consensus on prioritizing manual validation and dual-review processes for AI-assisted evidence.
Recommendation for buyers: Treat NICE’s guidance as a gold standard and proactively apply these principles across all jurisdictions. 

Key Applications Under Scrutiny

Systematic Literature Review (SLR) —GenAI can help automate search, screening, data extraction, and even code generation for meta-analysis. However, HTAs expect:

Real-World Evidence (RWE)—HTA bodies see potential in AI for structured extraction from unstructured sources. But they require:

  • Proven data provenance
  • Cross-validation with conventional data sources
  • Supplementary—not primary—role for AI-generated outputs

Health Economic Modeling—AI holds promise for building, optimizing, and validating economic models. Still, regulators demand:

Best-Practice Vendor Checklist for HTA-Ready AI

RequirementEssential Vendor Capabilities
TransparencyAI-use declarations, explainability, rationale for each decision
ValidationProven performance metrics, cross-validation, pilot data
Human OversightClearly defined roles and review pathways
Regulatory AlignmentAdherence to frameworks PALISADE, TRIPOD+AI, Cochrane, GIN
Security & EthicsBias mitigation, data privacy, cybersecurity risk reduction
Stakeholder EngagementEarly consultation with NICE and future HTA bodies

What’s Next? 

As comfort and adoption with AI grows, expect increasing harmonization as global HTAs align around principles initiated by NICE and emerging EU-Wide policies. This will mean:

  • Agencies will demand higher accountability around AI performance and transparency as digital evidence becomes mainstream.
  • Vendors will need to invest in explainable AI toolkits, on-premise deployment, services, and domain expertise to support HTA-grade submission.

Final Takeaway

Generative AI has tremendous potential to accelerate evidence synthesis—but successful HTA submissions require a balanced approach. Transparency, validation, human oversight, and regulatory alignment are not optional—they’re essential. For buyers and vendors alike, aligning with NICE’s position and proactively following these guidelines is your best path to future-ready, AI-supported market access.