ReHARK: Refined Hybrid Adaptive RBF Kernels for Robust One-Shot Vision-Language Adaptation
arXiv cs.CV / 3/13/2026
📰 NewsModels & Research
Key Points
- The paper addresses the stability-plasticity trade-off in adapting large-scale vision-language models to downstream tasks with extremely limited data, highlighting limitations of prior training-free methods that rely on local estimators.
- ReHARK reinterprets few-shot adaptation as global proximal regularization in a reproducing kernel Hilbert space (RKHS) and introduces a training-free, multistage refinement pipeline to improve robustness.
- The pipeline includes Hybrid Prior Construction (fusing zero-shot textual knowledge from CLIP and GPT-3 with visual class prototypes), Support Set Augmentation (bridging), Adaptive Distribution Rectification, and Multi-Scale RBF Kernels.
- On 11 benchmarks, it achieves an average accuracy of 65.83%, setting a new state-of-the-art for one-shot vision-language adaptation, with code released at GitHub for practical adoption.
Related Articles

I made a 'benchmark' where LLMs write code controlling units in a 1v1 RTS game.
Dev.to

My AI Does Not Have a Clock
Dev.to
How to settle on a coding LLM ? What parameters to watch out for ?
Reddit r/LocalLLaMA

Andrej Karpathy's autonomous AI research agent ran 700 experiments in 2 days and gave a glimpse of where AI is heading
Reddit r/artificial

So cursor admits that Kimi K2.5 is the best open source model
Reddit r/LocalLLaMA