RHYTHM: Reasoning with Hierarchical Temporal Tokenization for Human Mobility
Abstract
RHYTHM uses hierarchical temporal tokenization and large language models to predict human mobility, capturing long-range dependencies and multi-scale periodic behaviors efficiently.
Predicting human mobility is inherently challenging due to complex long-range dependencies and multi-scale periodic behaviors. To address this, we introduce RHYTHM (Reasoning with Hierarchical Temporal Tokenization for Human Mobility), a unified framework that leverages large language models (LLMs) as general-purpose spatio-temporal predictors and trajectory reasoners. Methodologically, RHYTHM employs temporal tokenization to partition each trajectory into daily segments and encode them as discrete tokens with hierarchical attention that captures both daily and weekly dependencies, thereby significantly reducing the sequence length while preserving cyclical information. Additionally, we enrich token representations by adding pre-computed prompt embeddings for trajectory segments and prediction targets via a frozen LLM, and feeding these combined embeddings back into the LLM backbone to capture complex interdependencies. Computationally, RHYTHM freezes the pretrained LLM's backbone to reduce attention complexity and memory cost. We evaluate our model against state-of-the-art methods using three real-world datasets. Notably, RHYTHM achieves a 2.4% improvement in overall accuracy, a 5.0% increase on weekends, and a 24.6% reduction in training time. Code is publicly available at https://github.com/he-h/rhythm.
Community
RHYTHM (Reasoning with Hierarchical Temporal Tokenization for Human Mobility) is a novel framework that recasts mobility modeling as a temporal-token reasoning task. By segmenting trajectories into daily tokens (and pooling them hierarchically), RHYTHM dramatically shortens sequence length while retaining cyclical structure. It then enriches those tokens with semantic embeddings generated by a frozen LLM, and feeds them into the same model as a reasoning backbone. This combination of compressed temporal abstraction + LLM-based reasoning delivers stronger predictive accuracy with far lower computational cost than prior methods.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Zero-Shot Human Mobility Forecasting via Large Language Model with Hierarchical Reasoning (2025)
- Adapting LLMs to Time Series Forecasting via Temporal Heterogeneity Modeling and Semantic Alignment (2025)
- From Values to Tokens: An LLM-Driven Framework for Context-aware Time Series Forecasting via Symbolic Discretization (2025)
- ST-LINK: Spatially-Aware Large Language Models for Spatio-Temporal Forecasting (2025)
- GSTM-HMU: Generative Spatio-Temporal Modeling for Human Mobility Understanding (2025)
- TransLLM: A Unified Multi-Task Foundation Framework for Urban Transportation via Learnable Prompting (2025)
- T3Time: Tri-Modal Time Series Forecasting via Adaptive Multi-Head Alignment and Residual Fusion (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper