Reproduction of TurboQuant

Reddit r/LocalLLaMA / 4/16/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The post argues that recent TurboQuant implementations across multiple inference frameworks may be noisy or AI-generated, and asks which paper claims have been independently validated.
  • The author focuses on reproducing specific claims—especially “lossless compression”—and comparing real-world performance against other low-bit quantization approaches.
  • After spending a full day reproducing the TurboQuant+QJL setup, the author reports that results were worse in their tests and questions whether QJL delivers practical benefits.
  • The discussion implicitly calls for more rigorous third-party benchmarks and clearer evidence to separate verified gains from overstated or unvalidated claims.
  • The content is framed as reproduction/replicability research rather than an announcement of new TurboQuant releases or tools.

There have been many TurboQuant implementations recently in llama.cpp, mlx, vllm, and sglang, but a lot of the discussion and code around them feels pretty noisy and looks to be AI-generated.

I’m trying to understand which claims from the paper have actually been validated by independent third parties. For example, has the lossless compression claim been reproduced, and how does TurboQuant perform in practice compared with other low-bit quantization methods?

I spent an entire day reproducing the TurboQuant+QJL setup, and it only made performance worse in my tests. I was wondering whether QJL is providing a meaningful practical benefit here.

submitted by /u/ExpensivePilot1431
[link] [comments]