Papers
arxiv:2509.23115

RHYTHM: Reasoning with Hierarchical Temporal Tokenization for Human Mobility

Published on Sep 27
· Submitted by Haozheng Luo on Sep 30
Authors:
,
,

Abstract

RHYTHM uses hierarchical temporal tokenization and large language models to predict human mobility, capturing long-range dependencies and multi-scale periodic behaviors efficiently.

AI-generated summary

Predicting human mobility is inherently challenging due to complex long-range dependencies and multi-scale periodic behaviors. To address this, we introduce RHYTHM (Reasoning with Hierarchical Temporal Tokenization for Human Mobility), a unified framework that leverages large language models (LLMs) as general-purpose spatio-temporal predictors and trajectory reasoners. Methodologically, RHYTHM employs temporal tokenization to partition each trajectory into daily segments and encode them as discrete tokens with hierarchical attention that captures both daily and weekly dependencies, thereby significantly reducing the sequence length while preserving cyclical information. Additionally, we enrich token representations by adding pre-computed prompt embeddings for trajectory segments and prediction targets via a frozen LLM, and feeding these combined embeddings back into the LLM backbone to capture complex interdependencies. Computationally, RHYTHM freezes the pretrained LLM's backbone to reduce attention complexity and memory cost. We evaluate our model against state-of-the-art methods using three real-world datasets. Notably, RHYTHM achieves a 2.4% improvement in overall accuracy, a 5.0% increase on weekends, and a 24.6% reduction in training time. Code is publicly available at https://github.com/he-h/rhythm.

Community

Paper author Paper submitter

RHYTHM (Reasoning with Hierarchical Temporal Tokenization for Human Mobility) is a novel framework that recasts mobility modeling as a temporal-token reasoning task. By segmenting trajectories into daily tokens (and pooling them hierarchically), RHYTHM dramatically shortens sequence length while retaining cyclical structure. It then enriches those tokens with semantic embeddings generated by a frozen LLM, and feeds them into the same model as a reasoning backbone. This combination of compressed temporal abstraction + LLM-based reasoning delivers stronger predictive accuracy with far lower computational cost than prior methods.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2509.23115 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2509.23115 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2509.23115 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.