From Baselines to Preferences: A Comparative Study of LoRA/QLoRA and Preference Optimization for Mental Health Text Classification

arXiv cs.CL / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper provides a systematic comparison of optimization strategies for mental-health text classification, starting from strong vanilla and classical baselines and progressively moving to more specialized methods.
  • It evaluates parameter-efficient supervised fine-tuning using LoRA/QLoRA across multiple objective and optimization settings, while also testing preference optimization methods including DPO, ORPO, and KTO with class-rebalanced training.
  • Results indicate that gains are highly method-dependent: some approaches produce stable, transferable improvements, while others are sensitive to configuration choices and data balance.
  • Preference optimization shows especially large variation across objective formulations, suggesting that simply adding preference-training is not sufficient—method selection matters more than the presence of a preference stage.
  • The authors propose a reproducible “optimization narrative” for mental-health NLP: begin with transparent baselines, apply controlled tuning, and use preference optimization only when it demonstrably improves performance.

Abstract

Mental health text classification has rapidly adopted modern adaptation methods, yet practical guidance on which optimization strategy to use, when, and why remains limited. This paper presents a systematic comparative study of optimization pathways for a joint mental-health classification task, moving from strong vanilla baselines to progressively more specialized techniques. We first establish classical and encoder references, then examine parameter-efficient supervised fine-tuning with LoRA/QLoRA under multiple objective and optimization settings, and finally evaluate preference-based optimization with DPO, ORPO, and KTO, including class-rebalanced training. Rather than emphasizing a single headline score, we focus on methodological insight: how performance changes with objective formulation, adapter choice, optimizer behavior, context windowing, and class-balance intervention. The results show that optimization effects are highly method-dependent: some approaches deliver stable, transferable gains, while others are sensitive to configuration and data balance. Preference optimization, in particular, exhibits large variation across objectives, indicating that method selection is more consequential than simply adding a preference-training stage. The central contribution is a clear optimization narrative for mental health NLP: start from transparent baselines, apply controlled tuning, and use preference optimization selectively where its gains are demonstrable. This provides a reproducible and practically grounded framework for choosing effective training strategies beyond architecture choice alone.