Solutions
About Us
Insights
Careers
Contact us
Contact Us
Customer Support
Customer Support

A Preview of Cytel’s Contributions at PHUSE EU 2025

I can’t believe it has already been a year since we wrapped up PHUSE EU Connect 2024, and in two weeks we will be gathering another exciting PHUSE EU Connect conference, only a few kilometers from Heidelberg, where everything started twenty years ago with the very first PHUSE event. I was one of the couple hundred lucky attendees and now, twenty years later, I have the great honor of supporting Jennie McGuirk and Jinesh Patel as Conference Co-chair for this year’s edition.

With a promising agenda featuring about 190 presentations, 34 posters, 9 hands-on workshops, 2 panel discussions, and 3 inspiring keynote speakers, this year we are going to the city of Hamburg for the 21st PHUSE EU Connect. The agenda is full of topics looking toward the future, with about 40 talks and posters referring to AI in their titles, and once again open source will be the confirmed leitmotif.

Cytel will make a significant contribution this year, perhaps more than ever, with six presentations, one poster, active participation in both panel discussions, and co-chairing the “Scripts, Macros and Automation” and “People Leadership & Management” streams.

 

Monday topics: Agile code writing, extracting metadata from R OOP functions, and leadership

The week kicks off on Monday with Kamil Foltynski, who will present “Overcoming Challenges in Collaborative Spreadsheet Editing with Shiny, SpreadJS and JSON-Patch” in the Application Development stream at 11:30 am. Kamil will provide a technical deep dive into enabling real-time spreadsheet editing within Shiny applications, using tools such as SpreadJS, sharing key lessons learned so far. Following Kamil’s presentation, Eswara Satyanarayana Gunisetti, will present “Micro-Decisions, Macro Impact: The Role of Agile Thinking in Every Line of Code” in theCoding Tips & Tricks” stream at 12 pm. See his recent blog on the topic. Eswara will share how an agile “mindset” can positively influence the way we write code.

In the same stream, a few hours later at 2 pm, another colleague Edward Gillian, in collaboration with Sanofi, will present “Risk.assessr: Extracting OOP Function Details,” discussing strategies for extracting metadata from R Object-Oriented Programming functions. Prior to Eswara and Edward’s sessions, at 1:30 pm, Kath Wright, will moderate the Interactive People Leadership & Management session “Invisible Glue: Trust, Influence and The Architecture of Teamwork.” With this live workshop, attendees will engage in practical exercises to learn how to identify barriers to trust, evaluate influence dynamics, and apply evidence-based strategies to strengthen collaboration in both physical and virtual environments.

 

Tuesday topics: Industry trends, extracting macro usage and dependency information from SAS programs, and integrating ECA data into CDISC-compliant datasets

Tuesday also brings two presentations and one poster. Right after lunch at 1:30 pm, Cedric Marchand will join other industry leaders in the panel discussion “Reimagining Statistical Programming: AI, Standards & the Talent of Tomorrow.” The panel will explore how current industry trends, such as AI, open source, and the evolution of data standards, will influence the next generation of statistical programmers.

The afternoon continues at 4 pm with my young and talented colleague Marie Poupelin, who will present “From Zero to Programming Hero: How Internships Shape Statistical Programmers in a CRO” in the “Professional Development” stream. Marie is a great example of the success of our internship program, and she will share her journey from having “zero” statistical programming experience to becoming an industry-ready programmer. Thirty minutes later, at 4:30 pm, Guido Wendland will present “Which Macros Are Used in the Study?” in the “Scripts, Macros and Automation” stream, a stream co-led this year for the first time by my colleague Sebastià Barceló. Guido will discuss techniques to extract macro usage and dependency information from SAS programs; this is particularly useful for identifying potential issues or estimating the impact of macro updates.

Later, in the traditional Tuesday evening poster session, you can join my colleague Cyril Sombrin in discussing “Our Journey in Integrating External Control Arms (ECAs) and RWD for Rare Disease Trials.” There you can discuss real-world case studies on integrating ECA data into CDISC-compliant datasets, exploring the unique challenges and solutions when aligning real-world data with CDISC standards.

 

Wednesday topics: Real-time spreadsheet editing within Shiny applications and real-time validation and streamlined submissions

On Wednesday at 12 pm, Hugo Signol, another young talented Cytel statistical programmer and a product of our internship program, will present his talk “From XPT to Dataset-JSON: Enabling Real-Time Validation and Streamlined Submissions.” Building on Cytel’s experience from CDISC Dataset-JSON-Viewer Hackathon, Hugo will demonstrate a Shiny application that supports interactive exploration and real-time validation through API-based checks.

 

Meet us there!

Cytel will be at Booth 9 at the conference, where you can engage in discussions with our team or meet any of us throughout the week.

I hope I didn’t miss anyone, or anything! We look forward again to reuniting with colleagues and friends from around the world and meeting new acquaintances.

See you all in Hamburg!

Generative AI in Evidence Synthesis: Harnessing Potential with Responsibility

The integration of AI into the healthcare research landscape is accelerating, with one obvious area of application being evidence synthesis. From early scoping reviews to comprehensive systematic literature reviews (SLRs), AI promises to reduce manual burden and enhance efficiency by saving time. However, it is crucial to understand both the strengths and limitations of using AI in this broad context to ensure compliance, reliability, and scientific rigor.

 

Knowing where it works: A targeted approach

Artificial intelligence, including generative AI models, shines when used for targeted literature reviews (TLRs) or when generating summaries of scientific articles to support evidence-based decision-making at an early development stage. AI can synthesize large volumes of information quickly, offering valuable insights during exploratory or early-phase research.

However, it’s critical to distinguish these from regulatory-facing systematic literature reviews, especially those intended for payer or health technology assessment (HTA) submissions. In this context, SLR extractions have traditionally been completed by two independent human reviewers. This human oversight ensures objectivity and reproducibility, key elements of regulatory compliance.

 

Expertly trained models vs. generalist giants

The current landscape is filled with large generalist language models trained on diverse internet-scale data. While impressive, these models often exhibit hallucinations — the generation of plausible but incorrect or fabricated content — particularly in domain-specific applications like evidence synthesis.

This is why domain-trained expert models are preferred. These models are fine-tuned on biomedical and scientific corpora, ensuring higher reliability and reducing the risk of misinterpretation or erroneous conclusions. They understand field-specific terminology, data structures, and compliance requirements far better than their generalist counterparts.

 

The imperative of data traceability

In evidence synthesis, transparency is non-negotiable. Any AI-generated output must allow users to:

  • Highlight the exact source (i.e., sentence or section) of the original scientific article from which a conclusion or data point was extracted.
  • Compare the model’s interpretation with the source text to identify discrepancies or nuances that could affect meaning or validity.

Using structured tags to annotate key terms, qualifiers, and relationships can make these comparisons clearer and more systematic but also inform advanced search and retrieval activities. By surfacing subtle differences, tagging supports expert review, preserves contextual integrity, and strengthens the reliability and defensibility of the synthesized evidence.

 

Measuring what matters: Precision and beyond

Traditional evaluation metrics like precision, recall, and F1 score (the harmonic mean of precision and recall) remain foundational when assessing AI model performance in literature screening and data extraction.

But in generative contexts — where the task may be summarization, paraphrasing, or abstract reasoning — additional measures become valuable:

  • Answer correctness: Does the output convey a factual, verifiable point?
  • Semantic similarity: How closely does the AI output align in meaning with the ground truth?
  • BLEU, ROUGE, and BERTScore: These Natural Language Processing metrics offer quantitative insights into the quality of generated text, especially for summarization and content generation tasks.

Selecting the right mix of these metrics provides a comprehensive view of model performance and reliability.

 

Where AI makes a difference: Screening and beyond

One of the most promising applications of generative AI in evidence synthesis is in literature screening, or the ability to assess whether a publication (abstract or full text) meets the criteria for inclusion. Studies and pilot implementations suggest that AI can reduce screening time by up to 40%, making it a powerful ally for research teams.

AI tools have been leveraged to assign a probability of inclusion to a title or abstract or full text to guide the screening process but also to allow researchers to quickly understand the impact of modifying search strategies on yield. By automating this repetitive and time-consuming phase, organizations can reallocate expert human resources to higher-value tasks, such as:

  • Resolving ambiguous or context-dependent data extractions
  • Validating nuanced findings and offering insights into implications of these findings
  • Ensuring alignment with HTA submission standards

In this way, AI doesn’t replace human reviewers but augments them, driving efficiency without compromising accuracy.

 

AI with guardrails

Generative AI is reshaping the landscape of evidence synthesis, but its integration must be strategic, measured, and compliant. By combining domain-trained models, robust traceability, appropriate evaluation metrics, and human oversight, organizations can unlock the true value of AI — accelerating workflows without sacrificing quality or compliance.

When used thoughtfully, generative AI becomes more than just a tool — it becomes a partner in advancing scientific research.

 

Meet with us at ISPOR 2025!

Manuel Cossio and Nathalie Horowicz-Mehler will be in Glasgow for ISPOR Europe 2025! Click the link below to book a meeting, or stop by Booth #1024 to connect with our experts:

Breaking Barriers in Rare Disease Research with Generative AI and Synthetic Data

In healthcare innovation, one of the most pressing challenges lies in rare disease research. There are approximately 7,000 rare diseases affecting over 300 million people worldwide. With only a handful of patients dispersed globally, gathering sufficient data to power robust clinical studies or predictive models is a monumental hurdle. However, a solution is emerging at the intersection of generative AI and real-world data (RWD) — a novel approach with the potential to reshape possibilities and unlock insights to address unmet medical needs in rare diseases.

 

The rare disease data dilemma

In the U.S., rare diseases are defined as conditions affecting fewer than 200,000 people. Despite their low individual prevalence, rare diseases collectively impose a significant burden on both patients and healthcare systems.

Research and development in rare diseases often face a vicious cycle: low prevalence leads to data scarcity. Traditional clinical trials are often infeasible and/or statistically underpowered due to the limited pool of participants.

Meanwhile, RWD sources such as electronic health records (EHRs), insurance claims, registries, and patient-reported outcomes offer valuable, albeit messy and fragmented, glimpses into the patient journey. Yet even RWD struggles to paint a complete picture in rare diseases. This is where generative AI steps in.

 

Enter generative AI: Making data where there is none

Generative AI — especially models like Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and, more recently, large foundation models — has a transformative ability: it can learn patterns from limited datasets and generate synthetic yet realistic datasets.

How it works

  1. Learning from RWD: Even small datasets from rare disease patients can be used to train and fine-tune generative models. These models identify patterns, distributions, and time-dependent relationships present in the data.
  2. Synthesizing patients: Once trained, the model can create new, synthetic patient records that preserve the statistical properties and characteristics of the original data. These “digital patients” simulate disease progression, treatment responses, and comorbidities.
  3. Validating realism: Synthetic data must be validated to ensure it reflects the real-world data it was trained on. Techniques like distributional comparison, propensity scoring, and expert validation are used to ensure accuracy and utility.

 

Why synthetic data matters for rare diseases

Synthetic data can enhance rare disease clinical research in many ways, including:

 

1. Augmenting small cohorts

Synthetic data can boost sample sizes for rare disease studies, enabling:

  • Simulation of clinical trials
  • Development of more robust predictive models
  • Generation of synthetic control arms where traditional controls are ethically or logistically impractical

 

2. Enhancing privacy

In rare diseases, patient re-identification is an increased risk due to unique phenotypes or genetic markers. Synthetic data protects patient privacy, while at the same time preserves the utility of the data.

 

3. Facilitating global collaboration

As synthetic data is deidentified, it facilitates data sharing among researchers, institutions and borders, minimizing regulatory hurdles and fostering cross-collaborative discovery.

 

4. Accelerating drug development

Pharma and biotech companies can use synthetic data to:

  • Test drug targeting strategies
  • Model long-term outcomes
  • Conduct in silico trials in the earliest stages of development

 

Challenges and considerations

While promising, this approach is not without its challenges:

  • Bias amplification: Synthetic data reflects the biases of its training data. If the RWD is incomplete or skewed, so will the synthetic outputs be. Strategies to handle bias are essential.
  • Regulatory acceptance: Regulatory bodies are still evaluating how to incorporate synthetic data into approval pathways.
  • Validation standards: There is a need for consistent benchmarks and best practices for validating synthetic data — both in terms of privacy and utility, as well as broader generative AI applications in healthcare.

 

Looking ahead

The marriage of generative AI and RWD opens new doors for rare disease research. With the ability to synthesize patient data that preserves real-world complexity, we can begin to break free from the constraints of scarcity — generating insights, hypotheses, and interventions that were once out of reach.

As we move forward, interdisciplinary collaboration among clinicians, data scientists, regulatory bodies, and patient advocacy groups will be key to harnessing this potential ethically and effectively.

 

Interested in learning more?

Download our complimentary ebook, Rare Disease Clinical Trials: Design Strategies and Regulatory Considerations:

From Metadata to Submission: Rule-Based Robotic Process Automation for Statistical Programming Excellence

In the race to modernize data operations in clinical research and regulatory submissions, Robotic Process Automation (RPA) powered by rule-based systems has emerged as a dependable and high-impact solution. These systems offer clarity, control, and reproducibility — critical traits for industries like biopharma where regulatory compliance and data integrity are non-negotiable.

Here, we discuss rule-based RPA as the foundation for a scalable and auditable standards automation pipeline.

 

Rule-based automation: Transparent, trusted, and tunable

Unlike more probabilistic models, rule-based systems operate on deterministic logic. Every output is traceable back to an explicit rule, which enhances trust and simplifies troubleshooting. This transparency is particularly valuable when the processes must be easily explained to stakeholders and auditors.

Key strengths of rule-based RPA include:

Transparency

Each step in the workflow is rule-driven, making the logic easy to inspect, validate, and justify. This ensures regulatory reviewers can clearly understand how data was transformed or outputs generated — vital in submission contexts.

Consistency

Standard rules applied across studies generate consistent outputs. For example, Cytel’s ALPS system creates SDTM and ADaM code from structured specifications, producing reliable results that hold up across different projects and teams.

Customizability

Rule-based systems are modular. Teams can easily adapt existing rules to accommodate study-specific needs without overhauling the entire system. Tools like Prism allow this by applying both generic rules and study-specific layers for enriched metadata processing.

 

Cytel’s metadata-driven RPA workflow in action

Our internal automation pipeline demonstrates the power of rule-based RPA. It’s built on a modular architecture where each tool performs a specific, rules-driven task:

  • ALPS: Converts metadata specifications into ready-to-run SAS code for SDTM and ADaM datasets, reducing manual programming and minimizing error risks.
  • Lighthouse: Enables biostatisticians to build mock shells using reusable templates, ensuring consistency in table and listing structures.
  • Prism: Extracts metadata from mock shells and transforms it into XML-format ARMs (Analysis Results Metadata), enriching it through rules and generating code for up to 60% of standard safety outputs.
  • TAB Macros and CytelDocs: Automate the creation of summary tables and documentation, saving hours of effort and ensuring compliance with standardized formats.

This end-to-end pipeline reduces manual touchpoints, maintains high quality, and boosts team efficiency.

 

Where generative AI complements RPA

While rule-based systems are ideal for tasks requiring consistency and auditability, generative AI can complement these systems — particularly in areas where variability is acceptable and outputs don’t require deterministic reproducibility. For example, Gen AI can assist with:

  • Drafting exploratory narratives or documentation
  • Suggesting code for non-critical outputs
  • Enhancing user interfaces with intelligent prompts
  • Enrich the set of study specific rules to be used

However, these AI-driven capabilities are best applied where hallucinations won’t compromise integrity, and outputs don’t demand rigid consistency.

 

Business and quality benefits of rule-based RPA

By relying on rule-based RPA for core data workflows, we’ve realized several tangible gains:

  • Time efficiency: Standard code is generated automatically, freeing time for custom analysis.
  • Reduced redundancy: Developers no longer rewrite common code across projects.
  • Improved QA: Outputs are independently validated and built on rigorously tested rule sets.
  • Collaboration at scale: Uniform rules simplify onboarding and knowledge transfer.
  • Focus on what matters: Teams can concentrate on non-standard elements that require expertise.

 

Final takeaways

Rule-based RPA systems provide the transparency, structure, and adaptability required for high-stakes data environments. At Cytel, we’ve found them indispensable in our mission to expedite regulatory submissions without compromising on quality or compliance. As AI continues to evolve, generative technologies may enrich this foundation — but rule-based automation remains the core engine that ensures accuracy, accountability, and speed.

Agentic Autonomy: How Multi-Agent Systems Could Orchestrate the Future of Clinical Development

In recent years, artificial intelligence has evolved beyond basic pattern matching to become capable of autonomous reasoning, multi-step planning, and even delegation. This transition — from passive tools to goal-driven, reasoning agents — marks the rise of agentic AI.

For the life sciences sector, and especially clinical development, this evolution arrives at a critical time. Clinical trials are increasingly complex, cross-functional, and data-intensive. Agentic AI offers not just faster tools, but the possibility of autonomous collaboration — teams of agents working in harmony to reduce burden, increase efficiency, and shorten timelines.

Here we explore the evolution of agentic AI and how higher levels of autonomy could transform clinical development from reactive execution to proactive, intelligent orchestration.

 

The evolution of agentic AI

Agentic AI evolves through distinct levels of capability. Each stage unlocks new functionality — from static models to ecosystems of communicating agents. Here’s a clear breakdown of the five major levels:

 

 

Each level builds toward intelligent autonomy. The transition from Level 3 to Levels 4 and 5 introduces intentional behavior, goal-setting, and inter-agent collaboration — the foundations of autonomous operations in clinical development.

 

Agentic AI in clinical development: A new operating model

Clinical development is not just complex — it’s interdependent. Every milestone relies on the seamless handoff and integration of data, code, documents, and decisions. Agentic AI, particularly at Levels 4 and 5, promises to re-architect this model.

 

Level 4: Planning and reasoning agents

These agents can independently break down goals, design execution paths, and adapt to changing environments. Here’s how they can drive value:

  • Medical writing agents
    • What they do: Generate drafts for protocols, CSRs, and patient narratives.
    • How they help: Understand document structures, integrate real-time data, and adapt language for regulatory or clinical audiences.
    • Outcome: Faster document turnaround, reduced rework, and scalable writing support.

 

  • Statistical programming agents
    • What they do: Develop and validate analysis code in SAS, R, or Python.
    • How they help: Plan logical sequences, debug outputs, and dynamically update based on protocol amendments.
    • Outcome: Accelerated code generation with built-in quality assurance.

 

  • Information synthesis agents
    • What they do: Retrieve and synthesize information from multiple domains — scientific literature, regulatory guidelines, real-world data, health system policies, and reports on unmet medical needs.
    • How they help: Prioritize and contextualize inputs to support clinical design, indication selection, and risk-benefit assessments.
    • Outcome: Broader strategic alignment and better-informed cross-functional planning.

 

Level 5: Multi-agent systems

At this level, clinical development becomes an ecosystem of agents, each with a specialized role, working under the coordination of orchestrator agents that function like project managers.

  • Orchestrator agents
    • What they do: Assign tasks, monitor progress, and realign workflows in real time.
    • How they help: Adjust deliverables dynamically as inputs change or downstream agents complete their tasks.
    • Outcome: Continuously managed, self-optimizing trial execution.

 

  • Agent networks
    • Example: A data management agent processes raw datasets and hands outputs to a statistical agent, which triggers a writing agent to draft updated narratives — all autonomously.
    • Value: End-to-end automation with minimal human handoffs.
    • Outcome: Real-time trial updates and agility under pressure.

 

The benefits of the agent ecosystem

 

From automation to autonomy

Agentic AI reflects an evolution from “AI that assists” to “AI that takes initiative” — supporting actions, learning from experience, and extending expertise across domains. In clinical development, where complexity continues to rise and efficiency is critical, this shift offers a meaningful opportunity rather than just an advantage.

As we look toward Levels 4 and 5, we can imagine a future where trials increasingly manage themselves, where teams are supported by networks of intelligent agents, and where human professionals gain more space to focus on innovation, thoughtful oversight, and meaningful patient outcomes.

 

Meet with us at ISPOR 2025!

Manuel Cossio will be in Glasgow for ISPOR Europe 2025! Click the link below to book a meeting, or stop by Booth #1024 to connect with our experts:

Redefining Clinical Documentation in the Age of Intelligent Collaboration: The Rise of the AI-Assisted Medical Writing Strategist

The introduction of AI into medical writing workflows marks a pivotal turning point in clinical development. As life sciences companies deploy AI agents to generate clinical documents — from clinical study protocols (CSPs) together with the Statistical Analysis Plan (SAP) to clinical study reports (CSRs) — a new role is emerging: the AI-assisted medical writing strategist.

This role represents a shift in mindset and skillset. No longer is the medical writer just a document author; they are becoming a strategic orchestrator of AI tools, data-driven narratives, and regulatory precision.

 

What is an AI-assisted medical writing strategist?

An AI-assisted medical writing strategist is a clinical and regulatory expert who partners with AI systems to accelerate and optimize the development of clinical documents. They bring together deep scientific understanding, regulatory knowledge, and technical fluency to co-create documents that are not only accurate and compliant but also delivered at unprecedented speed.

They are not just reviewing AI outputs — they are shaping the way AI generates those outputs, continuously fine-tuning the interaction between human judgment and machine efficiency.

 

Core pillars of the strategist role

The AI-assisted medical writing strategist role is defined by the following five key pillars:

 

1. AI orchestration, not just review

At the heart of the strategist’s work is the ability to guide AI systems toward producing high-quality, usable first drafts. This means:

  • Designing intelligent prompts based on document type and trial context.
  • Structuring modular content frameworks that AI can populate and iterate on.
  • Embedding company-specific style guides, preferred language, and regulatory templates into AI workflows.

 

2. Scientific and regulatory oversight

Even with AI generating drafts, clinical development demands nuanced, evidence-based interpretation. The strategist ensures:

  • Scientific rigor in efficacy and safety narratives.
  • Consistency in data interpretation across documents.
  • Adherence to ICH, FDA, EMA, and country-specific requirements.

AI might know the rules, but the strategist knows the exceptions, the subtleties, and the evolving guidance that govern every submission.

 

3. Training the AI with human expertise

AI systems improve through feedback. Strategists:

  • Curate and label high-quality training datasets (e.g., past CSRs, protocols).
  • Correct and comment on AI-generated drafts to reinforce preferred structures and content styles.
  • Continuously evaluate model performance and guide retraining cycles.

They act as domain-informed teachers, helping the AI become a better writing partner over time.

 

4. Cross-functional bridge builder

Medical writing is inherently collaborative. The strategist aligns AI output with expectations from:

  • Clinical, data management, and statistical teams.
  • Regulatory affairs and quality assurance.
  • Legal, ethical, and patient advocacy groups.

In doing so, they help organizations reimagine review cycles, moving from linear drafting to agile co-creation.

 

5. Champion of ethics and transparency

AI is powerful — but it must be used responsibly. Strategists play a leading role in:

  • Ensuring AI doesn’t fabricate data or misrepresent study outcomes.
  • Clarifying where automation was used in document creation.
  • Promoting transparency, reproducibility, and compliance in every AI-assisted process.

 

Why this role matters

The volume and complexity of clinical documentation are only increasing. At the same time, timelines are shrinking, budgets are tightening, and regulatory scrutiny is rising. AI offers a way forward — but only when guided by human intelligence.

The AI-Assisted Medical Writing Strategist ensures that automation enhances human value rather than diminishing it. They unlock:

  • Faster turnaround times for key deliverables.
  • More consistent documentation across global studies.
  • Greater focus on high-value tasks like interpretation, innovation, and communication.

 

How to prepare for this role

Transitioning into this role requires new capabilities:

  • AI literacy: Understanding how large language models (LLMs) work, how they’re trained, and where they fall short.
  • Prompt engineering: Knowing how to ask the right questions and frame the right context for AI tools.
  • Regulatory acumen: Staying current with guidance on AI use in regulated document environments.
  • Change leadership: Helping others adopt AI tools confidently and responsibly.

 

Final thoughts

The AI-assisted medical writing strategist is more than a job title — it’s a vision for the future of clinical documentation. As the life sciences industry embraces digital transformation, this role becomes essential to ensure that automation is paired with accountability, speed with accuracy, and efficiency with empathy.

By stepping into this role, medical writers don’t just adapt to the AI era — they lead it.

Streamlining Data Management and Improving Statistical Accuracy in Clinical Trials with AI

As clinical trials grow increasingly complex, the need for smarter, faster, and more efficient data processes and analysis is in demand. Artificial intelligence (AI) emerges as a powerful tool, especially in programming and data management. For clinical trial professionals, AI offers the promise of streamlining workflows, improving data quality, and reducing time to database lock.

 

The evolving role of AI in clinical data programming

AI is not replacing clinical programmers; it’s augmenting them. AI should be considered a tool to use within clinical trials, just as EDC and SAS are commonly used tools. Automation tools driven by machine learning can now handle routine, rules-based programming tasks such as edit check generation, derivation logic, and data transformation. This allows programmers to focus on more strategic activities like validating statistical code or optimizing data pipelines. AI needs the expertise of our clinical trial professionals.

Natural Language Processing (NLP) is also making great progress. For instance, NLP can interpret free-text protocol documents to auto-generate specifications, electronic case report form (eCRF) templates, or even suggest initial SDTM mappings, significantly reducing manual effort.

 

AI in data cleaning and quality oversight

Traditionally, data cleaning has been labor-intensive, with data managers manually reviewing queries, data listings, and edit checks across multiple data sources and systems. AI tools can now proactively flag anomalies or data trends that human review might miss, such as unexpected patterns in lab values, inconsistencies across visits, or possible fraudulent data across participants and sites.

Predictive models can help identify study participants at high risk of dropout or noncompliance, enabling earlier intervention. This not only improves data completeness but also enhances trial efficiency and participant retention. The effort and cost of replacing clinical trial participants is significant and felt across all stakeholders. Improving the patient’s experience would be a significant way to save time, money, and accelerating progress.

 

AI in statistical programming: From code automation to advanced insights

Statistical programming is central to clinical trial analysis from producing tables, listings, and figures (TLFs) to preparing submission-ready datasets. Traditionally reliant on manual coding in SAS or R, this work is now gaining speed, consistency, and quality through AI augmentation.

 

Where AI adds value in statistical programming

  • Automated code generation: AI models trained on historical programming logic can produce initial SAS macros or R scripts for common TLFs and dataset derivations. These drafts accelerate development by up to 40–60%, freeing programmers and biostatisticians to focus on complex analyses and interpretation.
  • Code review and validation: AI-assisted tools can scan code for logic errors, inefficiencies, redundant steps, and deviations from programming standards. Acting as a “second reviewer,” they flag potential issues early and suggest optimizations.
  • Dynamic statistical modeling: AI algorithms can rapidly explore large trial datasets to detect subgroup effects, anomalies, or emerging trends. When guided by statistical oversight, these insights can refine analysis plans and support adaptive trial decisions.

The aim is not to replace human judgment, but to boost productivity, reproducibility, and the speed of insight generation, without compromising scientific rigor.

 

AI in biostatistics: Powering smarter, more adaptive clinical trials

 Biostatistics remains the foundation of evidence generation in clinical trials, providing the methodological rigor to transform raw data into reliable conclusions. In the context of AI, biostatisticians play a dual role: safeguarding scientific validity while leveraging new computational tools to enhance insight generation. This requires a careful balance between deep domain knowledge and technical proficiency in emerging AI-driven methodologies. From applying knowledge graphs (KGs) to map complex biomedical relationships, to developing predictive models that anticipate trial outcomes, biostatistics is evolving into a more dynamic and interconnected discipline.

 

Where AI adds value in biostatistics

  • Balanced expertise: Integrating statistical theory with AI/ML techniques to ensure robust, interpretable results.
  • Knowledge graph applications: Using KGs to uncover hidden relationships between biomarkers, treatments, and outcomes, supporting hypothesis generation and trial design.
  • Early prediction tools: Building predictive models for recruitment success, dropout risk, and endpoint achievement.
  • Segmentation and personalization: Identifying patient subgroups most likely to benefit from a therapy, improving trial efficiency and precision medicine strategies.
  • Support for registrational trials: Leveraging AI to optimize trial design, stratify patient populations, and run simulations that ensure the study is powered and structured for regulatory success.

 

Regulatory readiness and caution

Despite its promise, AI must be implemented thoughtfully. Regulatory agencies like the FDA are increasingly open to the use of advanced technologies but expect transparency, traceability, and validation. AI-based tools must be auditable and explainable, especially when used in clinical data workflows that feed into regulatory submissions.

 

What’s next?

As AI becomes more embedded in clinical trial ecosystems, we can expect increased integration with EDC systems, CDISC standards, and statistical programming tools. The goal isn’t to eliminate human oversight but to enhance it, allowing clinical data professionals to make faster, better-informed decisions.

 

Final takeaways

AI is reshaping programming and data management in clinical trials. For clinical trial professionals, now is the time to become familiar with these tools, understand their capabilities and limitations, and engage with cross-functional teams to ensure responsible and impactful implementation. Ultimately our goal is to shorten drug development timelines and improve patient outcomes. With AI, we can be part of the solution to provide improved treatments for patients.

 

Interested in learning more?

Join Steven Thacker, Sheree King, Kunal Sanghavi, and Juan Pablo Garcia Martinez for their upcoming webinar, “How AI Enhances Biometrics Services: Streamlining Data Management and Improving Statistical Accuracy in Clinical Trials” on Thursday, August 28 at 10 am ET:

Trustworthy AI in Action: Predicting Stroke Risk Transparently with Claims-Based Machine Learning

In recent years, deep learning and large neural networks have garnered most of the attention in the machine learning (ML) community. Their ability to model complex, high-dimensional data is indeed impressive. But in healthcare — where decisions can have serious consequences and interpretability is paramount — simpler, transparent models like logistic regression still have an important role to play.

Not every problem requires a black box. When it comes to predicting disease risk using structured data, such as insurance claims, traditional models can offer accuracy and insight.

 

Claims databases: An untapped resource for disease risk prediction

Claims databases are an increasingly valuable source of real-world data (RWD). Unlike clinical trial data, which is highly controlled but limited in scale and scope, administrative claims datasets cover millions of lives over multiple years, reflecting real patient behavior and care patterns.

These databases include information on diagnoses, procedures, prescriptions, and demographics — elements that, while lacking granular clinical detail, can still reveal important patterns in disease progression and risk. The scale of these datasets allows for robust statistical modeling, even for rare outcomes.

 

The case for explainable machine learning in claims-based risk prediction

When working with claims data, models like logistic regression, Lasso, or Ridge regression are not just sufficient — they are often ideal. These models:

  • Produce coefficients that quantify the relationship between features and outcomes.
  • Allow for transparent understanding of why a prediction was made.
  • Are easier to validate and communicate to clinicians, payers, and regulators.

In contrast, deep learning models often deliver slightly higher accuracy at the cost of interpretability — a trade-off that may not be acceptable in regulated healthcare environments.

 

A real-world example: Predicting stroke risk with claims data

In a recent study, Cytel used data from over 2.5 million insured individuals to predict the risk of stroke hospitalization. Using only claims-based features such as age, medication use, comorbidities (e.g., diabetes, hypertension), and health service utilization, we compared the performance of several models, including:

  • Logistic Regression
  • Regularized linear models (Lasso and Ridge)
  • XGBoost (a state-of-the-art ML algorithm)

The results? All models achieved similar predictive performance, with area under the ROC curve (AUC) values around 0.81. Logistic regression — simple, explainable, and well-established — performed on par with XGBoost, demonstrating that advanced complexity wasn’t necessary to achieve meaningful predictive power.

 

Transparency enables trust and action

What sets models like logistic regression apart is their explainability. Stakeholders can see precisely how risk factors like atrial fibrillation, hypercholesterolemia, or age contribute to predicted stroke risk. This level of clarity is essential not only for clinicians making decisions, but also for data governance, compliance, and patient communication.

In a time when “black box” AI models are under increasing scrutiny, explainable models offer a pragmatic path forward — especially when paired with large-scale real-world datasets like claims data.

 

Keep it simple, keep it transparent

Healthcare doesn’t just need powerful algorithms — it needs trustworthy ones. As our study shows, standard machine learning models remain highly relevant, especially when applied to well-structured real-world data. Claims databases, in particular, offer a rich foundation for developing these models and making preventive healthcare smarter, earlier, and more accessible.

Blending Power and Flexibility: How AI-Generated R Code is Reshaping Clinical Trial Design

In today’s fast-evolving clinical research landscape, designing robust and efficient trials is more critical than ever. As statistical designs grow in sophistication, biostatisticians are increasingly relying on both commercial platforms and open-source tools to meet unique modeling needs. But this hybrid approach also comes with challenges, particularly for those new to advanced simulation software or lacking programming experience.

At Cytel, we’ve been exploring how artificial intelligence (AI) can help bridge this gap. At the 2025 Joint Statistical Meetings (JSM), we will present on our latest innovation: AI-powered R code generation for clinical trial design, a feature embedded in our East Horizon™ platform. This assistant, called RCACTS (R Coding Assistant for Clinical Trial Simulation), represents a significant step forward in making custom trial design faster, more accessible, and more reliable.

 

Why talk about this now? The open-source imperative

While commercial clinical trial design software offers rapid design development through validated and user-friendly workflows, it doesn’t always address the full complexity of real-world problems. Trial statisticians often face challenges in areas such as oncology, rare diseases, and adaptive designs that require tailored statistical tests, unique outcome generation models, or alternative randomization techniques.

This is where open-source tools like R become invaluable. R allows statisticians to write custom code to simulate complex trial designs, perform Bayesian analyses, or integrate evolving regulatory guidance. Over the years, a vibrant ecosystem of R packages has emerged, offering a high degree of transparency, flexibility, and academic rigor.

Yet this flexibility comes with trade-offs: code development can be time-consuming, error-prone, and requires significant programming expertise. As a result, many biostatisticians find themselves switching between validated commercial workflows and custom R functions, leading to a process that is often fragmented and inefficient.

Recognizing this, Cytel’s East Horizon platform has introduced R integration points, enabling users to inject custom code directly into validated simulation workflows. This integration delivers the best of both worlds: the speed and structure of commercial software with the creativity and control of open-source.

 

Enter AI: Speed, simplicity, and smarter coding

Our next logical question was: can AI make this process even easier?

The answer, increasingly, is yes. With recent advances in generative AI, particularly large language models (LLMs), it’s now possible to assist in the generation of R code for simulation-based design tasks. At Cytel, we’ve harnessed OpenAI’s GPT-4o via API, securely deployed within Microsoft Azure, to create RCACTS, a coding assistant purpose-built for biostatisticians.

Unlike generic AI tools that produce standalone R scripts, RCACTS generates R code specifically tailored for the East Horizon simulation engine. It ensures that the generated functions:

  • Match expected input/output structures,
  • Include pre-defined parameters as shown in our internal statistical package CyneRgy,
  • Are immediately ready for integration and testing within a live trial design workflow.

With RCACTS, users can simply describe what they want in plain English and receive functioning R code that can be integrated into East Horizon.

 

Who benefits? Everyone from newcomers to experts

One of the major advantages of this AI-enhanced workflow is lowering the barrier to entry. For a new user unfamiliar with Cytel’s R integration or syntax requirements, writing compatible code from scratch can be daunting. RCACTS significantly reduces the learning curve by providing validated function templates, sensible defaults, and clear parameterization, all supported by generative AI.

At the same time, experienced statisticians benefit by spending less time on repetitive coding tasks, debugging, or remembering function signatures. This allows them to focus on higher-level design questions, such as: What analysis method is most robust? How sensitive is the design to different outcome distributions? What dropout patterns pose the greatest risk?

Our assistant supports a wide range of trial design elements:

  • Simulating patient responses: Binary, Continuous, Time-to-event, and Repeated-measure endpoints.
  • Analyzing simulated data: Statistical analysis for these endpoints.
  • Randomization: Flexible randomization of patients across treatment groups.
  • Enrollment and dropout modeling: Custom mechanisms for realistic patient enrollment and dropout scenarios.
  • Treatment selection: Supporting multi-arm multi-stage (MAMS) trial designs.

 

Balancing innovation with responsibility

Of course, like any AI solution, there are caveats. AI-generated code must be carefully reviewed for correctness, appropriateness, and regulatory readiness. RCACTS includes a built-in testing functionality to flag structural or syntactic errors, but statistical validation remains the user’s responsibility. Also note that all data interactions adhere to Azure OpenAI’s stringent data protection policies to ensure security and compliance.

There’s also a broader concern: will over-reliance on AI limit the creativity and deep statistical thinking that define our profession? At Cytel, we view AI not as a replacement for expertise, but as a tool to amplify it. Our goal is to give statisticians more time and mental space to explore, iterate, and innovate rather than reduce them to prompt engineers.

 

Looking ahead

The future of clinical trial design lies in intelligent integration: combining the strengths of validated commercial tools, flexible open-source frameworks, and AI-powered coding assistance. With East Horizon and RCACTS, we believe we’re building the blueprint for this future, with a platform that supports both scientific rigor and operational speed.

As the field continues to evolve, biostatisticians will need tools that not only keep up with complexity but also support creativity, collaboration, and efficiency. AI-generated R code, embedded within a powerful simulation engine, is one such tool and is already transforming how we approach design flexibility in clinical trials.

 

Catch us at JSM 2025 to learn more about how AI is transforming the future of clinical trial design within Cytel.

Leveraging Mobile and Wearable Technology for Outcomes Research in Depression

As mobile and wearable technologies become increasingly integrated into daily life, their applications have expanded far beyond convenience and lifestyle. In the field of outcomes research — particularly within mental health — these technologies are opening new frontiers for understanding and monitoring clinical endpoints. A notable case is depression, where continuous digital monitoring can provide rich insights into both the course of illness and treatment impact.

This post draws on our findings from a recent systematic review and poster presentation to examine how mobile and wearable tools are currently deployed in depression monitoring and how this aligns with broader outcomes research goals.

 

Digital monitoring as a tool for mental health outcomes

Over the past five to six years, depression has seen a marked rise across youth and adult populations globally, underscoring the need for scalable and effective monitoring strategies. In parallel, smartphones and wearables have become ubiquitous, capable of capturing passive, longitudinal health data. These digital tools offer unprecedented potential for outcomes research by providing real-time behavioral and physiological markers relevant to depression.

To map the current landscape, we conducted a comprehensive literature review focused on how smartphones and wearables are used to monitor depression in research contexts. This synthesis aimed to highlight prevailing methods, feature usage, and the extent to which demographic variability is accounted for — critical considerations in health outcomes analysis.

 

Key findings from the literature

We reviewed 140 studies and identified 22 that met our inclusion criteria. The following themes emerged:

 

Study characteristics

  • Recency: Most studies were published in 2024, reflecting the field’s rapid acceleration.
  • Geography: The U.S. and Pakistan emerged as leading contributors.
  • Sample Size: Studies included an average of 465 participants, suggesting moderately powered observational designs.

 

Demographic reporting

  • Gender and age: Captured in 20 of the 22 studies.
  • Ethnicity: Reported in just 9 studies.
  • Education and marital status: Only 4 studies reported these variables — yet both are key social determinants of health and influence depression outcomes.

 

Monitoring technologies and features

  • Smartphones were used in 20 of the 22 studies, highlighting their dominance.
  • Key features monitored included:
    • Mood tracking: 20 studies
    • Movement (accelerometer data): 10 studies
    • Heart Rate Variability (HRV): 5 studies
    • Word usage tracking: 4 studies
    • Sleep patterns: 2 studies

 

Clinical assessment tools

Self-reported clinical scales were commonly used as outcome anchors:

  • PHQ-9 (Patient Health Questionnaire-9): 6 studies
  • GAD-7 (Generalized Anxiety Disorder-7): 7 studies

(See our original poster for a visual breakdown of these features and tools.)

 

Implications for outcomes research

From an outcomes research perspective, these technologies offer compelling advantages:

  • Continuous and passive monitoring: Enables longitudinal capture of clinically relevant endpoints like mood, behavior, and sleep — reducing bias from intermittent self-reporting.
  • Scalability and reach: Mobile-based data collection can extend to underserved and geographically dispersed populations, improving study generalizability.
  • Early signal detection: Passive data streams can flag deterioration or improvement earlier than clinical visits alone, offering potential for timely interventions.

However, a consistent limitation observed in the literature is the underreporting of demographic variables — especially education and marital status. This omission constrains subgroup analysis and limits insights into how different populations experience depression and respond to interventions. In outcomes research, such data are essential for contextualizing and stratifying results across socioeconomic or cultural dimensions.

 

The path forward

As wearable and mobile sensors become more refined, their integration into real-world data frameworks will likely become standard practice in outcomes research. But to truly capitalize on this potential, researchers must enhance demographic reporting and examine interactions between digital phenotypes and traditional health indicators across diverse populations.

These tools not only offer more granular tracking of mental health status — they also help researchers and health systems better understand the dynamics of treatment effectiveness, burden of illness, and quality of life over time.

 

Interested in learning more?

This blog summarizes findings from the poster presentation, “Exploring Mobile and Wearable Technology for Early Depression Detection and Monitoring,” presented by Lyuboslav Ivanov and Manuel Cossio at Cytel and Universitat de Barcelona.