When fractional quasi p-norms concentrate

arXiv stat.ML / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies whether distances concentrate in high-dimensional spaces when distances are measured using fractional quasi p-norms with p in (0,1), a question tied to long-running theoretical and empirical debates.
  • It provides conditions under which fractional quasi p-norms do concentrate and conditions under which they do not, showing that for broad distribution classes they can satisfy strong concentration bounds that hold uniformly in p.
  • The results contradict earlier claims that one could generally mitigate concentration issues by “optimally” choosing p within (0,1), effectively ruling out such approaches for the identified distribution families.
  • It also characterizes distribution classes where concentration control is still possible through appropriate p selection, while highlighting that near regimes with concentration there exist uncountably many distributions exhibiting anti-concentration.
  • The authors connect these findings to practical implications for designing data encoding or representation schemes that can either encourage or discourage distance concentration.

Abstract

Concentration of distances in high dimension is an important factor for the development and design of stable and reliable data analysis algorithms. In this paper, we address the fundamental long-standing question about the concentration of distances in high dimension for fractional quasi p-norms, p\in(0,1). The topic has been at the centre of various theoretical and empirical controversies. Here we, for the first time, identify conditions when fractional quasi p-norms concentrate and when they don't. We show that contrary to some earlier suggestions, for broad classes of distributions, fractional quasi p-norms admit exponential and uniform in p concentration bounds. For these distributions, the results effectively rule out previously proposed approaches to alleviate concentration by "optimal" setting the values of p in (0,1). At the same time, we specify conditions and the corresponding families of distributions for which one can still control concentration rates by appropriate choices of p. We also show that in an arbitrarily small vicinity of a distribution from a large class of distributions for which uniform concentration occurs, there are uncountably many other distributions featuring anti-concentration properties. Importantly, this behavior enables devising relevant data encoding or representation schemes favouring or discouraging distance concentration. The results shed new light on this long-standing problem and resolve the tension around the topic in both theory and empirical evidence reported in the literature.