AI Navigate

Tested 14 embedding models on Thai — here's how they rank

Reddit r/LocalLLaMA / 3/16/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Qwen3-Embedding-4B leads the Thai MTEB benchmark with a score of 74.41 across 15 Thai tasks, followed by KaLM-Gemma3-12B (73.92), BOOM_4B_v1 (71.84), and other models.
  • Qwen3-Embedding-0.6B is notably strong for its small size, nearly matching 4B models on Thai tasks, while bge-m3 is solid but not particularly Thai-focused.
  • All benchmarks ran on Thailand's national supercomputer (LANTA) and the results were merged into the official MTEB repository, with an interactive per-task leaderboard available.
  • The findings inform model selection for Thai NLP and are useful for researchers and ML engineers evaluating embedding options.

Ran MTEB benchmarks on 15 Thai tasks using A100 GPUs. Results:

  1. Qwen3-Embedding-4B — 74.41
  2. KaLM-Gemma3-12B — 73.92
  3. BOOM_4B_v1 — 71.84
  4. jina-v5-text-small — 71.69
  5. Qwen3-Embedding-0.6B — 69.08
  6. multilingual-e5-large — 67.22
  7. jina-v5-text-nano — 66.85
  8. bge-m3 — 64.77
  9. jina-v3 — 57.81

Qwen3-0.6B is impressive for its size — nearly matches 4B models on Thai. bge-m3 is solid but nothing special for Thai specifically.

Interactive leaderboard with per-task breakdown: https://anusoft.github.io/thai-mteb-leaderboard/

All benchmarks ran on Thailand's national supercomputer (LANTA). Results merged into the official MTEB repo.

submitted by /u/anusoft
[link] [comments]