Is Min P sampling really the preferred modern alternative to Top K/Top P?

Reddit r/LocalLLaMA / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The post questions whether Min P sampling is truly the preferred modern alternative to Top K/Top P, based on community consensus and anecdotal guidance.
  • It observes that recently published LLMs on Hugging Face and elsewhere still recommend using Top K and/or Top P sampling parameters.
  • The author asks whether this continued reliance on Top K/Top P is merely due to legacy inertia or if there are other technical reasons.
  • Overall, the discussion frames a comparison between sampling strategies and their practical adoption in current model setups.

According to what I've been reading (and also according to all models I've asked about this), the consensus seems to be that Min P is the better/more modern approach to sampling and that it should be preferred over Top P/Top K, which should be used only if Min P isn't available or for legacy reasons...

Yet, looking and recently published LLM on huggingface and elsewhere, the recommended parameters for sampling are still largely Top K and/or Top P. Is this only for legacy reasons? Or some other reason?

submitted by /u/bgravato
[link] [comments]