A Scalable Nystrom-Based Kernel Two-Sample Test with Permutations
arXiv stat.ML / 4/21/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses two-sample hypothesis testing and proposes a scalable approach for determining whether two datasets come from the same distribution.
- It improves the practicality of maximum mean discrepancy (MMD)-based testing for large-scale settings by using a Nyström approximation of the MMD.
- The authors provide finite-sample theoretical guarantees, including a bound on the test’s power when the two distributions are sufficiently separated in MMD.
- They show that the derived separation rate achieves the known minimax-optimal rate for this problem setting.
- Numerical experiments demonstrate the method’s applicability to realistic scientific data and highlight computational efficiency.
Related Articles

Every time a new model comes out, the old one is obsolete of course
Reddit r/LocalLLaMA

We built it during the NVIDIA DGX Spark Full-Stack AI Hackathon — and it ended up winning 1st place overall 🏆
Dev.to

Stop Losing Progress: Setting Up a Pro Jupyter Workflow in VS Code (No More Colab Timeouts!)
Dev.to

Building AgentOS: Why I’m Building the AWS Lambda for Insurance Claims
Dev.to

Where we are. In a year, everything has changed. Kimi - Minimax - Qwen - Gemma - GLM
Reddit r/LocalLLaMA