ActionPiece: Contextually Tokenizing Action Sequences for Generative Recommendation
Abstract
Generative recommendation (GR) is an emerging paradigm where user actions are tokenized into discrete token patterns and autoregressively generated as predictions. However, existing GR models tokenize each action independently, assigning the same fixed tokens to identical actions across all sequences without considering contextual relationships. This lack of context-awareness can lead to suboptimal performance, as the same action may hold different meanings depending on its surrounding context. To address this issue, we propose ActionPiece to explicitly incorporate context when tokenizing action sequences. In ActionPiece, each action is represented as a set of item features, which serve as the initial tokens. Given the action sequence corpora, we construct the vocabulary by merging feature patterns as new tokens, based on their co-occurrence frequency both within individual sets and across adjacent sets. Considering the unordered nature of feature sets, we further introduce set permutation regularization, which produces multiple segmentations of action sequences with the same semantics. Experiments on public datasets demonstrate that ActionPiece consistently outperforms existing action tokenization methods, improving NDCG@10 by 6.00% to 12.82%.
Community
We propose ActionPiece, the first context-aware action sequence tokenizer for generative recommendation, which can tokenize the same action into different tokens based on the surrounding context in different sequences.
thanks. Looks very interesting
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- PatchRec: Multi-Grained Patching for Efficient LLM-based Sequential Recommendation (2025)
- Multimodal Large Models Are Effective Action Anticipators (2025)
- TALKPLAY: Multimodal Music Recommendation with Large Language Models (2025)
- Order-agnostic Identifier for Large Language Model-based Generative Recommendation (2025)
- Token Assorted: Mixing Latent and Text Tokens for Improved Language Model Reasoning (2025)
- Stop Looking for Important Tokens in Multimodal Language Models: Duplication Matters More (2025)
- Full-Stack Optimized Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper