Evaluating Post-hoc Explanations of the Transformer-based Genome Language Model DNABERT-2

arXiv cs.LG / 4/24/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper evaluates whether post-hoc explanation techniques that work for CNN-based genome models also transfer to Transformer-based genome language models, using DNABERT-2 as the test case.
  • Researchers adapt AttnLRP—an extension of layer-wise relevance propagation to attention mechanisms—and develop strategies to map explanations at both token and nucleotide levels.
  • The approach is assessed on multiple genomic datasets with several evaluation metrics to judge the reliability of the resulting explanations.
  • Results are compared extensively against a baseline CNN, showing that AttnLRP produces explanations aligned with known biological patterns, supporting biological hypothesis generation.
  • The study advances gLM explainability and specifically targets comparability of relevance attributions across different neural network architectures.

Abstract

Explaining deep neural network predictions on genome sequences enables biological insight and hypothesis generation-often of greater interest than predictive performance alone. While explanations of convolutional neural networks (CNNs) have been shown to capture relevant patterns in genome sequences, it is unclear whether this transfers to more expressive Transformer-based genome language models (gLMs). To answer this question, we adapt AttnLRP, an extension of layer-wise relevance propagation to the attention mechanism, and apply it to the state-of-the-art gLM DNABERT-2. Thereby, we propose strategies to transfer explanations from token and nucleotide level. We evaluate the adaption of AttnLRP on genomic datasets using multiple metrics. Further, we provide an extensive comparison between the explanations of DNABERT-2 and a baseline CNN. Our results demonstrate that AttnLRP yields reliable explanations corresponding to known biological patterns. Hence, like CNNs, gLMs can also help derive biological insights. This work contributes to the explainability of gLMs and addresses the comparability of relevance attributions across different architectures.

Evaluating Post-hoc Explanations of the Transformer-based Genome Language Model DNABERT-2 | AI Navigate