Injecting Structured Biomedical Knowledge into Language Models: Continual Pretraining vs. GraphRAG
arXiv cs.CL / 4/21/2026
📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper compares two ways to inject structured biomedical knowledge from the UMLS Metathesaurus into language models: continual pretraining (embedding knowledge into model parameters) and GraphRAG (querying a knowledge graph at inference time).
- It builds a large UMLS-derived biomedical knowledge graph (3.4M concepts, 34.2M relations) in Neo4j, and generates an ~100M-token text corpus to continually pretrain BERT-based models (BERTUMLS, BioBERTUMLS).
- Across six BLURB benchmarks, BERTUMLS outperforms the base BERT, especially on knowledge-intensive QA tasks, while benefits for BioBERTUMLS are more mixed due to potential diminishing returns when the base model already has biomedical knowledge.
- On QA evaluations (PubMedQA and BioASQ), GraphRAG applied to LLaMA 3-8B improves accuracy by over 3 points on PubMedQA and 5 points on BioASQ without retraining, providing transparent, multi-hop, and easily updatable knowledge access.
- The authors release the processed UMLS Neo4j graph to enable reproducible research.
Related Articles

Black Hat USA
AI Business

Black Hat Asia
AI Business

A practical guide to getting comfortable with AI coding tools
Dev.to

Every time a new model comes out, the old one is obsolete of course
Reddit r/LocalLLaMA

We built it during the NVIDIA DGX Spark Full-Stack AI Hackathon — and it ended up winning 1st place overall 🏆
Dev.to