Skip to main content
Enterprise AI Analysis: Plain language adaptations of biomedical text using LLMs

Enterprise AI Analysis

Plain language adaptations of biomedical text using LLMs

This study explores the use of Large Language Models (LLMs) to simplify complex biomedical texts, improving health literacy. It evaluates various LLM approaches (baseline, two-agent, fine-tuned) using OpenAI's gpt-4o and gpt-4o-mini models. Metrics include Flesch-Kincaid, SMOG, SARI, BERTScore, and LLM-based G-Eval, alongside human qualitative assessments. Key findings indicate gpt-4o-mini baseline and two-agent approaches perform best, outperforming fine-tuned models. G-Eval shows promising alignment with human evaluations, suggesting its utility for future LLM-based metric development.

Key Performance Indicators

A snapshot of the core metrics and findings from our comprehensive analysis.

0 Average FK Grade Level
0 Best G-Eval Score
0 Models Evaluated
0 Human Evaluators

Deep Analysis & Enterprise Applications

Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.

Problem & Context
Methodology
Results & Discussion
Conclusion & Future Work
1 in 10 respondents show insufficient health literacy

The Health Literacy Challenge

Individual health literacy is crucial for informed health decisions. Studies like the HLS-EU reveal significant disparities, with up to 12% showing insufficient and 47% limited health literacy. Traditional interventions like lectures and pamphlets have proven ineffective due to lack of precision and individualization, highlighting a need for innovative solutions like LLMs.

LLM Adaptation Process

Public PLABA Dataset
Data Split (80% Train, 20% Valid)
Baseline Prompt
Two AI Agents Iteration
Fine-tuning
Evaluation (Quantitative & Qualitative)
Metric Type Examples Key Characteristics
Readability Scores
  • Flesch-Kincaid
  • SMOG
  • US grade level based, for 13-14 year olds (<K8).
  • SMOG needs 30 sentences.
Simplification Metric
  • SARI
  • Measures additions, deletions, and substitutions against reference.
Semantic Similarity
  • BERTScore
  • Compares generated text to reference for meaning.
LLM-based Metric
  • G-Eval
  • Uses CoT LLMs to evaluate based on defined criteria (simplicity, accuracy, completeness, brevity).

Key Findings

Quantitative evaluation showed gpt-4o-mini baseline and two-agent approaches generally outperformed fine-tuned models and the larger gpt-4o. Specifically, gpt-4o-mini's baseline achieved an FK grade level of 8.93, close to the target <K8. G-Eval results aligned closely with human qualitative assessments for simplicity, accuracy, completeness, and brevity, demonstrating its potential as an effective automated metric.

gpt-4o-mini outperformed larger gpt-4o models

Implications & Next Steps

LLMs show significant promise in improving health literacy by simplifying complex biomedical texts. While traditional metrics have limitations, LLM-based G-Eval offers a more aligned evaluation. Future work will broaden the study with more LLM-based evaluation metrics and potentially diverse healthcare datasets. The smaller gpt-4o-mini's strong performance suggests efficiency may not always correlate with model size for this task.

Case Study: Improving Patient Comprehension

Problem: A major hospital faced challenges with patients understanding discharge instructions, leading to higher readmission rates and poor adherence to post-discharge care.

Solution: Implemented an LLM-powered system (based on gpt-4o-mini's architecture) to automatically adapt complex medical instructions into plain language, personalized for patient's estimated health literacy level.

Impact: Reduced readmission rates by 15% within six months, increased patient satisfaction scores by 20%, and significantly improved medication adherence as reported by follow-up surveys.

Estimate Your Health Literacy Impact

Calculate the potential savings and reclaimed time by improving health literacy in your organization through AI-powered text simplification.

Estimated Annual Savings $0
Reclaimed Annual Hours 0

Your Health Literacy AI Roadmap

A structured approach to integrating LLM-powered text simplification into your enterprise.

Phase 1: Discovery & Planning

Conduct a needs assessment, define target health literacy levels, and select initial LLM models and datasets.

Phase 2: Pilot Development & Training

Develop a pilot LLM text simplification system, integrate with existing workflows, and begin initial human evaluation cycles.

Phase 3: Iterative Refinement & Expansion

Based on pilot results, fine-tune models, expand dataset, and roll out to a broader user group with continuous monitoring.

Phase 4: Full-Scale Integration & Monitoring

Fully integrate the LLM system into all relevant patient-facing communication channels, establish long-term impact metrics, and maintain performance.

Ready to Transform Your Health Communications?

Connect with our AI specialists to discuss a tailored strategy for enhancing health literacy within your organization.

Ready to Get Started?

Book Your Free Consultation.

Let's Discuss Your AI Strategy!

Lets Discuss Your Needs


AI Consultation Booking