Feature Weighting Improves Pool-Based Sequential Active Learning for Regression

arXiv cs.LG / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies pool-based sequential active learning for regression and identifies a gap in prior methods that compute inter-sample distances without accounting for which features matter most.
  • It introduces five feature-weighted active learning variants (three single-task and two multi-task), using ridge regression coefficients from a small labeled set to weight features during distance calculations.
  • Experiments indicate the proposed feature-weighting approach is easy to implement and almost always improves the performance of four existing active learning baselines for both single-task and multi-task regression.
  • The authors suggest the strategy can be extended to stream-based active learning and potentially adapted to classification algorithms as well.
  • Overall, the work offers a practical enhancement to sample-selection quality under limited labeling budgets by making representativeness/diversity computations more feature-aware.

Abstract

Pool-based sequential active learning for regression (ALR) optimally selects a small number of samples sequentially from a large pool of unlabeled samples to label, so that a more accurate regression model can be constructed under a given labeling budget. Representativeness and diversity, which involve computing the distances among different samples, are important considerations in ALR. However, previous ALR approaches do not incorporate the importance of different features in inter-sample distance computation, resulting in sub-optimal sample selection. This paper proposes three feature weighted single-task ALR approaches and two feature weighted multi-task ALR approaches, where the ridge regression coefficients trained from a small amount of previously labeled samples are used to weight the corresponding features in inter-sample distance computation. Experiments showed that this easy-to-implement enhancement almost always improves the performance of four existing ALR approaches, in both single-task and multi-task regression problems. The feature weighting strategy may also be easily extended to stream-based ALR, and classification algorithms.