Just Pass Twice: Efficient Token Classification with LLMs for Zero-Shot NER
arXiv cs.CL / 4/8/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper explains that causal LLMs struggle with zero-shot token classification for NER when disambiguation needs future context due to left-to-right (causal) attention.
- It proposes “Just Pass Twice (JPT),” which concatenates the input sentence with itself so tokens in the second pass can effectively attend to the full bidirectional context without any model architecture changes.
- JPT combines these bidirectionally informed representations with definition-guided entity embeddings to improve flexible zero-shot generalization across entity types.
- On zero-shot NER benchmarks (CrossNER and MIT), the method reports state-of-the-art performance with an average +7.9 F1 improvement over the previous best approach.
- The approach is also claimed to be over 20x faster than comparable generative decoding-based NER methods, while reducing issues like slow autoregressive generation, hallucinated entities, and formatting errors.
Related Articles
30 Days, $0, Full Autonomy: The Real Report on Running an AI Agent Without a Credit Card
Dev.to
We are building an OS for AI-built software. Here's what that means
Dev.to
Claude Code Forgot My Code. Here's Why.
Dev.to

Whats'App Ai Assistant
Dev.to
I Built a $70K Security Bounty Pipeline with AI — Here's the Exact Workflow
Dev.to