Job Skill Extraction via LLM-Centric Multi-Module Framework

arXiv cs.CL / 4/24/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces SRICL, an LLM-centric multi-module framework for extracting skills from job ads at span level to support candidate–job matching and labor-market analytics.
  • SRICL combines semantic retrieval from ESCO, in-context learning, and supervised fine-tuning, using format-constrained prompts to stabilize span boundaries and reduce errors.
  • A deterministic verifier is added to enforce structural rules such as correct BIO tagging, non-overlapping spans, and valid span pairing, with only minimal retries.
  • Experiments on six publicly available span-labeled corpora across sectors, languages, and domains show substantial STRICT-F1 gains over GPT-3.5 prompting baselines and a strong reduction in malformed/invalid tags and hallucinated spans.
  • The approach is positioned as enabling more dependable sentence-level deployment, particularly in low-resource, multi-domain environments where long-tail terms and distribution shifts are challenging.

Abstract

Span-level skill extraction from job advertisements underpins candidate-job matching and labor-market analytics, yet generative large language models (LLMs) often yield malformed spans, boundary drift, and hallucinations, especially with long-tail terms and cross-domain shift. We present SRICL, an LLM-centric framework that combines semantic retrieval (SR), in-context learning (ICL), and supervised fine-tuning (SFT) with a deterministic verifier. SR pulls in-domain annotated sentences and definitions from ESCO to form format-constrained prompts that stabilize boundaries and handle coordination. SFT aligns output behavior, while the verifier enforces pairing, non-overlap, and BIO legality with minimal retries. On six public span-labeled corpora of job-ad sentences across sectors and languages, SRICL achieves substantial STRICT-F1 improvements over GPT-3.5 prompting baselines and sharply reduces invalid tags and hallucinated spans, enabling dependable sentence-level deployment in low-resource, multi-domain settings.