Neighbor Embedding for High-Dimensional Sparse Poisson Data
arXiv stat.ML / 4/21/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses dimensionality reduction for high-dimensional, sparse count data that are well-modeled by Poisson distributions, which often violate the assumptions of standard methods like PCA and t-SNE.
- It introduces p-SNE (Poisson Stochastic Neighbor Embedding), a nonlinear neighbor-embedding approach tailored to Poisson count structure.
- p-SNE defines pairwise dissimilarity using KL divergence between Poisson distributions and optimizes the embedding using Hellinger distance.
- Experiments on synthetic and real datasets show p-SNE can recover meaningful structure, including communication weekday patterns, topic clusters in OpenReview papers, and neural spike temporal drift/stimulus gradients.
- The results suggest that incorporating the underlying probabilistic model of sparse counts can improve embedding quality versus geometry-assuming techniques designed for continuous data.
Related Articles

Rethinking Coding Education for the AI Era
Dev.to

We Shipped an MVP With Vibe-Coding. Here's What Nobody Tells You About the Aftermath
Dev.to

Agent Package Manager (APM): A DevOps Guide to Reproducible AI Agents
Dev.to

3 Things I Learned Benchmarking Claude, GPT-4o, and Gemini on Real Dev Work
Dev.to

Open Source Contributors Needed for Skillware & Rooms (AI/ML/Python)
Dev.to