Representation-Level Adversarial Regularization for Clinically Aligned Multitask Thyroid Ultrasound Assessment

arXiv cs.CV / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a clinically aligned multitask model for thyroid ultrasound that jointly predicts nodule segmentation masks and TI-RADS risk categories to reduce inconsistencies caused by reader variability.
  • It introduces a training approach that grounds TI-RADS classification in a compact, TI-RADS-aligned radiomics target while still leveraging deep features for discriminative performance.
  • The authors argue that multitask failures under annotator variability stem from competing gradients in shared representations and present RLAR (Representation-Level Adversarial Regularization) to make this competition explicit.
  • RLAR regularizes latent-space task sensitivity by using each task’s normalized adversarial direction as a geometric probe and penalizing excessive angular alignment between tasks’ adversarial directions.
  • Experiments on a public TI-RADS dataset show improved risk stratification versus single-task training and standard multitask baselines while maintaining segmentation quality, with code and pretrained models planned for release.

Abstract

Thyroid ultrasound is the first-line exam for assessing thyroid nodules and determining whether biopsy is warranted. In routine reporting, radiologists produce two coupled outputs: a nodule contour for measurement and a TI-RADS risk category based on sonographic criteria. Yet both contouring style and risk grading vary across readers, creating inconsistent supervision that can degrade standard learning pipelines. In this paper, we address this workflow with a clinically guided multitask framework that jointly predicts the nodule mask and TI-RADS category within a single model. To ground risk prediction in clinically meaningful evidence, we guide the classification embedding using a compact TI-RADS aligned radiomics target during training, while preserving complementary deep features for discriminative performance. However, under annotator variability, naive multitask optimization often fails not because the tasks are unrelated, but because their gradients compete within the shared representation. To make this competition explicit and controllable, we introduce RLAR, a representation-level adversarial gradient regularizer. Rather than performing parameter-level gradient surgery, RLAR uses each task's normalized adversarial direction in latent space as a geometric probe of task sensitivity and penalizes excessive angular alignment between task-specific adversarial directions. On a public TI-RADS dataset, our clinically guided multitask model with RLAR consistently improves risk stratification while maintaining segmentation quality compared to single-task training and conventional multitask baselines. Code and pretrained models will be released.