Robust Multi-Source Covid-19 Detection in CT Images

arXiv cs.CV / 4/7/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • 既存のCT画像によるCOVID-19検出モデルは、同一施設内の学習・評価では良好でも、複数センターのCT(装置・撮像プロトコル・患者背景が異なる)では性能が落ちやすいという課題がある。
  • 提案手法はマルチタスク学習により、COVID-19診断に加えて各スキャンの「データ出所センター」も同時に予測させ、共有バックボーン(EfficientNet-B7)で施設横断的な特徴表現を学習させる。
  • 学習データがセンター間で偏っている点に対し、出所センター分類ヘッドにはlogit-adjusted cross-entropyを用いて、少数派センターが無視されるのを抑える。
  • SSFLフレームワークにKDSを組み合わせて各スキャンから代表スライスを8枚選択し、検証データ308 scansでF1=0.9098、AUC-ROC=0.9647を達成した。
  • 公開コードが提示されており(GitHubリンクあり)、再現・発展のための参照実装として利用可能である。

Abstract

Deep learning models for COVID-19 detection from chest CT scans generally perform well when the training and test data originate from the same institution, but they often struggle when scans are drawn from multiple centres with differing scanners, imaging protocols, and patient populations. One key reason is that existing methods treat COVID-19 classification as the sole training objective, without accounting for the data source of each scan. As a result, the learned representations tend to be biased toward centres that contribute more training data. To address this, we propose a multi-task learning approach in which the model is trained to predict both the COVID-19 diagnosis and the originating data centre. The two tasks share an EfficientNet-B7 backbone, which encourages the feature extractor to learn representations that hold across all four participating centres. Since the training data is not evenly distributed across sources, we apply a logit-adjusted cross-entropy loss [1] to the source classification head to prevent underrepresented centres from being overlooked. Our pre-processing follows the SSFL framework with KDS [2], selecting eight representative slices per scan. Our method achieves an F1 score of 0.9098 and an AUC-ROC of 0.9647 on a validation set of 308 scans. The code is publicly available at https://github.com/Purdue-M2/-multisource-covid-ct.