A Comparative Study of Demonstration Selection for Practical Large Language Models-based Next POI Prediction

arXiv cs.CL / 4/9/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper compares multiple demonstration selection strategies for next point-of-interest (POI) prediction using large language models (LLMs) and in-context learning (ICL) over historical check-in data.
  • It shows that the choice of demonstrations strongly affects ICL effectiveness, motivating a systematic comparison of selection methods rather than relying on arbitrary or single-purpose approaches.
  • Across three real-world datasets, simple heuristics like geographical proximity, temporal ordering, and sequential patterns outperform more complex and costly embedding-based demonstration selection methods on both accuracy and computational cost.
  • In some cases, LLMs prompted with demonstrations chosen via these heuristics outperform existing fine-tuned models without additional training, suggesting practical deployment advantages.
  • The authors release the associated codebase, enabling replication and further experimentation for real-world POI/trajectory prediction systems.

Abstract

This paper investigates demonstration selection strategies for predicting a user's next point-of-interest (POI) using large language models (LLMs), aiming to accurately forecast a user's subsequent location based on historical check-in data. While in-context learning (ICL) with LLMs has recently gained attention as a promising alternative to traditional supervised approaches, the effectiveness of ICL significantly depends on the selected demonstration. Although previous studies have examined methods such as random selection, embedding-based selection, and task-specific selection, there remains a lack of comprehensive comparative analysis among these strategies. To bridge this gap and clarify the best practices for real-world applications, we comprehensively evaluate existing demonstration selection methods alongside simpler heuristic approaches such as geographical proximity, temporal ordering, and sequential patterns. Extensive experiments conducted on three real-world datasets indicate that these heuristic methods consistently outperform more complex and computationally demanding embedding-based methods, both in terms of computational cost and prediction accuracy. Notably, in certain scenarios, LLMs using demonstrations selected by these simpler heuristic methods even outperform existing fine-tuned models, without requiring further training. Our source code is available at: https://github.com/ryonsd/DS-LLM4POI.