Fuzzy Fingerprinting Encoder Pre-trained Language Models for Emotion Recognition in Conversations: Human Assessment and Validity Study

arXiv cs.CL / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses Emotion Recognition in Conversations (ERC) where standard encoder pre-trained language models can struggle to align with nuanced human judgments and lack interpretability.
  • It highlights a key failure mode on imbalanced datasets: minority emotions are often misclassified as the dominant “neutral” class.
  • The authors propose an interpretable method that combines pre-trained language models with Fuzzy Fingerprints (FFPs), which create emotion-specific prototypes from fuzzy, class-activation patterns in the PLM latent space.
  • During inference, each utterance is converted into a fuzzy fingerprint and compared to emotion prototypes using a fuzzy similarity over intersections of the fuzzy sets.
  • Experiments indicate that adding FFPs reduces overclassification into the neutral class and human evaluation supports the adequacy of the predictions, while offering insight into the classification process.

Abstract

In Emotion Recognition in Conversations (ERC), model decisions should align with nuanced human perception and ideally provide insights on the classification process. Standard encoder pre-trained language models (PLMs) are the state-of-the-art at these tasks but offer little insight into why a certain prediction is made. This is especially problematic in imbalanced datasets, where most utterances are labeled as neutral, making these models frequently misclassify minority emotions as the majority neutral class. To tackle this issue, we introduced a novel, interpretable approach to ERC by combining PLMs with Fuzzy Fingerprints (FFPs). FFP provide class-specific prototypes that reflect the characteristic class activation patterns in the PLM's latent space. They are derived by ranking and fuzzifying the activations of the pooled conversational context-dependent embeddings across training instances for each emotion. At inference time, each input utterance is similarly fuzzy fingerprinted and matched to the emotion prototypes using a fuzzy similarity function based on the aggregation of the intersection of the fuzzy sets that define each FFP. Experimental results show that FFP integration reduces overclassification into the neutral class and human evaluation further supports the adequacy of FFP predictions. Our proposed method thus bridges the gap between deep neural inference and human perception, performing at state-of-the-art level while simultaneously offering valuable insights into the classification procedure.