Paper status: completed

OpenOneRec Technical Report

Published:12/31/2025
Original LinkPDF
Price: 0.100000
Price: 0.100000
Price: 0.100000
2 readers
This analysis is AI-generated and may not be fully accurate. Please refer to the original paper.

TL;DR Summary

The paper introduces RecIF-Bench, a holistic benchmark for generative recommendation and a large dataset of 96 million interactions. It open-sources a comprehensive training pipeline, demonstrating scalable capabilities while reducing catastrophic forgetting, with the OneRec foun

Abstract

While the OneRec series has successfully unified the fragmented recommendation pipeline into an end-to-end generative framework, a significant gap remains between recommendation systems and general intelligence. Constrained by isolated data, they operate as domain specialists-proficient in pattern matching but lacking world knowledge, reasoning capabilities, and instruction following. This limitation is further compounded by the lack of a holistic benchmark to evaluate such integrated capabilities. To address this, our contributions are: 1) RecIF Bench & Open Data: We propose RecIF-Bench, a holistic benchmark covering 8 diverse tasks that thoroughly evaluate capabilities from fundamental prediction to complex reasoning. Concurrently, we release a massive training dataset comprising 96 million interactions from 160,000 users to facilitate reproducible research. 2) Framework & Scaling: To ensure full reproducibility, we open-source our comprehensive training pipeline, encompassing data processing, co-pretraining, and post-training. Leveraging this framework, we demonstrate that recommendation capabilities can scale predictably while mitigating catastrophic forgetting of general knowledge. 3) OneRec-Foundation: We release OneRec Foundation (1.7B and 8B), a family of models establishing new state-of-the-art (SOTA) results across all tasks in RecIF-Bench. Furthermore, when transferred to the Amazon benchmark, our models surpass the strongest baselines with an average 26.8% improvement in Recall@10 across 10 diverse datasets (Figure 1). This work marks a step towards building truly intelligent recommender systems. Nonetheless, realizing this vision presents significant technical and theoretical challenges, highlighting the need for broader research engagement in this promising direction.

Mind Map

In-depth Reading

English Analysis

1. Bibliographic Information

1.1. Title

The central topic of the paper is "OpenOneRec Technical Report: An Open Foundation Model and Benchmark to Accelerate Generative Recommendation."

1.2. Authors

The paper is authored by the OneRec Team, with a detailed list of contributors provided in Appendix A. The Core Contributors include Guorui Zhou, Honghui Bao, Jiaming Huang, Jiaxin Deng, Jinghao Zhang, Junda She, Kuo Cai, Lejian Ren, Lu Ren, Qiang Luo, Qianqian Wang, Qigen Hu, Rongzhou Zhang, Ruiming Tang, Shiyao Wang, Wuchao Li, Xiangyu Wu, Xinchen Luo, Xingmei Wang, Yifei Hu, Yunfan Wu, Zhanyu Liu, Zhiyang Zhang, and Zixing Zhang. Additional Contributors are Bo Chen, Bin Wen, Chaoyi Ma, Chengru Song, Chenglong Chu, Defu Lian, Fan Yang, Feng Jiang, Hongtao Cheng, Huanjie Wang, Kun Gai, Pengfei Zheng, Qiang Wang, Rui Huang, Siyang Mao, Tingting Gao, Wei Yuan, Yan Wang, Yang Zhou, Yi Su, Zexuan Cheng, Zhixin Ling, and Ziming Li. Their affiliations are not explicitly stated in the provided text, but the GitHub and Hugging Face links suggest they are associated with Kuaishou.

1.3. Journal/Conference

The paper is published as a preprint on arXiv (https://arxiv.org/abs/2512.24762). arXiv is a widely recognized open-access repository for preprints of scientific papers in fields like mathematics, physics, computer science, and quantitative biology. While it is not a peer-reviewed journal or conference in itself, it serves as a crucial platform for rapid dissemination of research findings and allows for community feedback before formal publication. Its influence is significant in facilitating timely academic discourse.

1.4. Publication Year

The paper was published at (UTC): 2025-12-31T10:15:53.000Z.

1.5. Abstract

The abstract states that the OneRec series has successfully integrated the fragmented recommendation pipeline into an end-to-end generative framework. However, a significant gap persists between recommendation systems and general intelligence. Current recommendation systems act as domain specialists, adept at pattern matching but lacking world knowledge, reasoning capabilities, and instruction following due to isolated data. This limitation is exacerbated by the absence of a holistic benchmark. To address this, the authors make three contributions:

  1. RecIF-Bench & Open Data: They propose RecIF-Bench, a holistic benchmark with 8 diverse tasks, evaluating capabilities from fundamental prediction to complex reasoning. They also release a training dataset of 96 million interactions from 160,000 users.

  2. Framework & Scaling: They open-source a comprehensive training pipeline (data processing, co-pretraining, post-training) to ensure reproducibility. They demonstrate predictable scaling of recommendation capabilities while mitigating catastrophic forgetting of general knowledge.

  3. OneRec-Foundation: They release OneRec-Foundation models (1.7B and 8B parameters) that achieve new state-of-the-art (SOTA) results across all RecIF-Bench tasks. When transferred to the Amazon benchmark, these models surpass baselines with an average 26.8% improvement in Recall@10 across 10 datasets.

    The work is presented as a step towards intelligent recommender systems, acknowledging significant technical and theoretical challenges and calling for broader research engagement.

The official source link is https://arxiv.org/abs/2512.24762 and the PDF link is https://arxiv.org/pdf/2512.24762v1.pdf. The paper is currently a preprint on arXiv.

2. Executive Summary

2.1. Background & Motivation

The paper addresses a critical challenge in the field of recommender systems: the significant gap between current generative recommender systems and general intelligence. While recent advancements, particularly the OneRec series, have successfully unified the traditional multi-stage recommendation pipeline into an end-to-end generative framework, these systems remain constrained by several limitations:

  • Isolated Data Silos: Existing models are typically trained on isolated, domain-specific data, preventing them from leveraging the massive data scaling that drives the emergent capabilities of Large Language Models (LLMs).

  • Lack of General Intelligence: Consequently, these models operate as domain specialists, excelling at collaborative pattern matching but lacking crucial world knowledge, reasoning capabilities, and instruction following abilities that are hallmarks of general intelligence. This limits their adaptability and broader utility.

  • Catastrophic Forgetting: Attempts to integrate LLM capabilities, such as aligning discrete recommendation identifiers with LLM linguistic space, often lead to catastrophic forgetting of the LLM backbone's inherent generalization capabilities, especially when fine-tuned on limited, task-homogeneous data.

  • Inadequate Benchmarking: A major impediment is the absence of a holistic benchmark capable of evaluating the integrated capabilities (prediction, reasoning, instruction-following) essential for these next-generation recommendation foundation models. Traditional benchmarks are confined to narrow, specialized tasks, usually focusing on closed-set ranking accuracy within single domains.

    The core problem is to bridge this semantic and functional gap to build truly intelligent recommender systems that can understand complex user intent, reason about recommendations, and follow diverse instructions, much like LLMs do in general domains.

2.2. Main Contributions / Findings

To address the identified challenges, the paper presents several key contributions:

  • RecIF-Bench: A Holistic Recommendation Instruction-Following Benchmark & Open Data:
    • Contribution: Introduction of RecIF-Bench, a multi-dimensional benchmark featuring 8 diverse tasks spanning 4 capability layers (from semantic alignment to complex reasoning). It covers short-video, e-commerce, and online advertising domains, designed to rigorously assess the multifaceted capabilities of recommender foundation models via instruction following. It is the first benchmark to cover long-sequence interactions, multi-task, multi-modal, cross-domain, multi-behavioral interactions, interleaved data (text and itemic tokens), and recommendation explanation.
    • Finding: It provides a robust testbed for quantifying the synergy between instruction following and recommendation. Additionally, the OneRec-Foundation models are evaluated on 7 widely-recognized general benchmarks to verify retention of broad reasoning and coding skills. A comprehensive training dataset (96 million interactions from 160,000 users) is released to facilitate reproducible research.
  • Open-Source Framework & Validated Scaling Laws:
    • Contribution: Open-sourcing of a complete training pipeline (built on PyTorch and VeRL) encompassing data processing, co-pretraining, and post-training protocols. A novel two-stage alignment strategy is introduced, combining on-policy distillation and recommendation-oriented Reinforcement Learning (Rec-RL).
    • Finding: Empirical validation of scaling laws in the recommendation domain. The study demonstrates predictable capability scaling, with optimal compute allocation requiring data volume to scale more aggressively than model parameters (a data-intensive regime where NoptC0.44N _ { \mathrm { o p t } } \propto C ^ { 0 . 4 4 } and DoptC0.56D _ { \mathrm { o p t } } \propto C ^ { 0 . 5 6 }). This framework effectively mitigates catastrophic forgetting of general knowledge.
  • OneRec-Foundation Model Family:
    • Contribution: Release of the OneRec-Foundation series, including 1.7B and 8B parameter models, built upon Qwen. This series moves beyond simple fine-tuning, endowing the LLM backbone with intrinsic recommendation capabilities. Standard versions are trained on the open-source dataset, and Pro versions are enhanced with a hundred-billion-token industrial corpus.
    • Finding: The models achieve new state-of-the-art (SOTA) performance across all RecIF-Bench tasks. Furthermore, they demonstrate exceptional cross-domain transferability, surpassing strong baselines on 10 Amazon datasets with an average 26.8% improvement in Recall@10, underscoring their robustness as foundation models.

3. Prerequisite Knowledge & Related Work

3.1. Foundational Concepts

To understand the OpenOneRec paper, a beginner should be familiar with several fundamental concepts in machine learning, particularly in the domains of natural language processing and recommender systems.

  • Large Language Models (LLMs): These are advanced artificial intelligence models trained on vast amounts of text data, enabling them to understand, generate, and process human language. They are typically based on the Transformer architecture and are known for their ability to learn complex patterns and perform a wide range of tasks, from translation to creative writing, through next-token prediction. Their "emergent capabilities" refer to advanced behaviors (like complex reasoning) that appear only when scaled to very large sizes and data volumes.
  • Generative Models: In machine learning, a generative model learns the distribution of data points in a dataset and can then generate new data samples that are similar to the training data. For example, a generative LLM can generate novel sentences or paragraphs. In recommendation, generative recommenders predict the next item by generating its identifier or description, rather than just ranking existing items.
  • Recommender Systems: These systems aim to predict users' preferences for items (e.g., movies, products, videos) and suggest items they are likely to enjoy.
    • Collaborative Filtering: A common technique in recommender systems that predicts user preferences based on the preferences of similar users or the characteristics of similar items. It essentially finds patterns in user-item interaction data.
    • Sequential Recommendation: A sub-field of recommendation focusing on modeling the temporal dynamics of user behavior. It predicts the next item a user will interact with based on their sequence of past interactions.
  • Transformer Architecture: The dominant neural network architecture for LLMs. It relies heavily on self-attention mechanisms to weigh the importance of different parts of the input sequence when processing each element. This allows it to capture long-range dependencies in data, unlike recurrent neural networks.
  • Autoregressive Modeling: A type of statistical model where the output at a given time step is dependent on previous outputs. In LLMs, this means generating text one token at a time, with each new token being predicted based on all the preceding tokens in the sequence. The paper formulates recommendation as a Next-Token Prediction problem within this paradigm.
  • Instruction Following: The ability of an LLM to understand and execute commands or requests given in natural language. Instruction tuning is a training method where models are fine-tuned on a diverse set of tasks presented as natural language instructions, enabling them to generalize to new, unseen instructions.
  • Catastrophic Forgetting: A phenomenon in neural networks where training a model on a new task causes it to forget previously learned information or skills from older tasks. This is a significant challenge when adapting general LLMs to specific domains like recommendation, as fine-tuning might erase their world knowledge.
  • Reinforcement Learning (RL): A paradigm of machine learning where an "agent" learns to make decisions by performing actions in an environment to maximize a cumulative reward signal. It's often used to fine-tune LLMs for better alignment with human preferences or specific task objectives.
  • Itemic Tokens: A novel representation for discrete items (like videos or products) that bridges the gap between item IDs and natural language. Instead of arbitrary IDs, items are represented by sequences of discrete tokens (codes) derived from their semantic embeddings, allowing LLMs to process them as if they were words.

3.2. Previous Works

The paper builds upon and differentiates itself from several lines of prior research:

  • OneRec Series (Deng et al., 2025; Zhou et al., 2025a,b): The OpenOneRec paper is presented as part of the OneRec series, which is noted for successfully unifying the multi-stage recommendation pipeline into an end-to-end generative framework. This means moving from separate stages like retrieval, ranking, and explanation to a single model that can generate recommendations directly.
  • Generative Recommenders (Geng et al., 2022; Wu et al., 2023): This trend in recommendation treats the task as a language processing problem. The paper highlights this shift, noting that LLMs offer a new paradigm.
  • LLM-Recommendation Alignment (Zheng et al., 2024; Liu et al., 2025):
    • LC-Rec (Zheng et al., 2024): A study that aligns discrete recommendation identifiers with the linguistic space of LLMs. The paper cites LC-Rec as a baseline and notes its limitation to a limited set of downstream tasks, potentially causing catastrophic forgetting.
    • OneRec-Think (Liu et al., 2025): Another related work aiming to bridge the semantic gap.
  • Traditional Discriminative Recommender Models: The paper compares OpenOneRec against several established sequential recommendation models:
    • BERT4Rec (Sun et al., 2019): Uses a Transformer-based architecture with bidirectional encoder representations to predict masked items in a user's sequence, capturing rich contextual information.
    • GRU4Rec (Hidasi et al., 2016): Employs Gated Recurrent Units (GRUs) to model sequential user behavior, particularly effective for session-based recommendations.
    • SASRec (Kang and McAuley, 2018): Self-Attentive Sequential Recommendation model that uses the self-attention mechanism to capture long-range dependencies and item transitions in user behavior sequences.
    • HSTU (Zhai et al., 2024): Trillion-parameter sequential transducers for generative recommendations, indicating very large-scale models.
    • ReaRec (Tang et al., 2025): Focuses on unleashing latent reasoning power for sequential recommendation, suggesting a move towards more intelligent recommendation.
  • Generative Recommender Baselines:
    • TIGER (Rajput et al., 2023): A generative retrieval model for recommender systems, serving as a strong baseline for generative recommendation.
  • Item Tokenization Techniques (Luo et al., 2025; Zhou et al., 2025a): The paper adopts Itemic Tokens and mentions RQ-Kmeans (Luo et al., 2025) for discretizing item semantic embeddings into hierarchical discrete codes. This is a crucial step for integrating items into an LLM's token space. Finite Scalar Quantization (FSQ) (Mentzer et al., 2023) is also used for extending token depth in transfer learning.
  • Base LLM Backbone:
    • Qwen (Yang et al., 2025): The OneRec-Foundation models are built on Qwen, a family of open-source LLMs. This provides the foundational linguistic and reasoning capabilities.
  • LLM Scaling Laws (Kaplan et al., 2020; Hoffmann et al., 2022): The paper references these works for its analysis of how model performance scales with model parameters (NN) and training tokens (DD). Chinchilla scaling laws (Hoffmann et al., 2022) are specifically mentioned for comparison.
  • Reinforcement Learning Frameworks:
    • Group Relative Policy Optimization (GRPO) (Shao et al., 2024): Used in the Rec-RL stage for fine-tuning. It's noted for computing advantage relative to a group of sampled trajectories, reducing computational overhead compared to Actor-Critic algorithms like PPO.
  • Data Deduplication:
    • MinHash Algorithm (Broder, 1997): Used for efficient fuzzy deduplication to ensure evaluation benchmarks are not leaked into the training data.

3.3. Technological Evolution

The field of recommender systems has seen a significant evolution:

  1. Early Systems (e.g., Matrix Factorization): Focused on discovering latent factors from user-item interaction matrices. These were often discriminative models, predicting ratings or clicks for existing items.

  2. Deep Learning for Sequential Recommendation: With the rise of deep learning, models like GRU4Rec and BERT4Rec started to capture temporal dependencies in user behavior, treating interactions as sequences. This improved the ability to predict the next item.

  3. Multi-modal and Cross-domain Approaches: Researchers began incorporating rich metadata (text, images) and leveraging data from multiple domains to enrich user and item representations.

  4. Generative Recommender Systems: The recent paradigm shift, inspired by LLMs, views recommendation as a generative task. Instead of just predicting scores or ranking, these systems generate item IDs, descriptions, or explanations. The OneRec series represents a significant step in this direction, unifying previously fragmented pipelines.

  5. LLM-based Foundation Models for Recommendation: This paper marks the latest stage, aiming to integrate the general intelligence, reasoning, and instruction-following capabilities of LLMs into recommendation. The goal is to move beyond domain-specific expertise to truly intelligent agents that can engage in complex dialogues about recommendations.

    OpenOneRec fits into this timeline at the cutting edge, striving to establish foundation models for recommendation that can leverage LLM scalability and generalizability, while also providing a holistic benchmark to evaluate these advanced capabilities.

3.4. Differentiation Analysis

Compared to existing methods, OpenOneRec introduces several core innovations and differentiations:

  • Holistic Benchmarking with RecIF-Bench:
    • Differentiation: Unlike existing benchmarks (e.g., PixelRec, KuaiSAR, NineRec, Yelp, Amazon) that partially address multi-modal, multi-task, or cross-domain aspects, RecIF-Bench is the first to simultaneously cover all seven critical dimensions: long-sequence interactions, multi-task, multi-modal, cross-domain, multi-behavioral interactions, interleaved data (text and itemic tokens), and recommendation explanation. This provides a truly comprehensive testbed for foundation models.
    • Innovation: The inclusion of Interleaved Data and Recommendation Explanation is particularly novel, directly assessing the instruction-following and reasoning capabilities central to LLM-based recommenders.
  • Unified Generative Framework with Scalable Pre-training:
    • Differentiation: While OneRec series and TIGER also adopt generative frameworks, OpenOneRec specifically integrates collaborative signals with general semantics through Itemic-Text Alignment and mixed-domain Co-Pretraining. This approach is designed to overcome the data silos and catastrophic forgetting issues prevalent in prior attempts to adapt LLMs for recommendation.
    • Innovation: The two-stage pre-training (Itemic-Text Alignment then Full-Parameter Co-Pretraining) systematically bridges the modality gap and injects recommendation knowledge while preserving general world knowledge from the Qwen3 backbone.
  • Hybrid Post-Training for Capability Balance:
    • Differentiation: Many LLM-based recommenders struggle to balance domain-specific precision with general reasoning abilities. OpenOneRec introduces a novel hybrid post-training pipeline: Multi-task Supervised Fine-tuning (SFT) is followed by alternating On-policy Distillation for General Capability and Reinforcement Learning for Recommendation (Rec-RL).
    • Innovation: On-policy distillation actively recovers general reasoning abilities by using the original LLM as a teacher, while Rec-RL (specifically Group Relative Policy Optimization with a rule-based Hit reward) directly optimizes for discrete ranking metrics in recommendation, effectively balancing both objectives.
  • Validated Scaling Laws in Recommendation:
    • Differentiation: The paper empirically validates scaling laws specifically in the recommendation domain. This extends the understanding from general LLMs to a specialized application.
    • Innovation: The finding of a data-intensive scaling regime (DoptC0.56D_{\mathrm{opt}} \propto C^{0.56} vs. NoptC0.44N_{\mathrm{opt}} \propto C^{0.44}) suggests that, unlike Chinchilla's near-equiproportional split for general text, optimal compute allocation in recommendation requires scaling data volume more aggressively than model parameters.
  • Exceptional Transferability and Few-Shot Learning Performance:
    • Differentiation: OpenOneRec-Foundation models demonstrate superior cross-domain transferability compared to baselines like TIGER, especially in few-shot learning scenarios.
    • Innovation: The Text-Augmented Itemic Tokens strategy, which concatenates original itemic tokens with keyword representations, effectively leverages the foundation model's diverse capabilities (collaborative filtering, knowledge, semantic understanding) while maintaining pre-trained structural integrity. This adaptability is crucial for robust performance in new domains with limited data.

4. Methodology

4.1. Principles

The core idea behind the OpenOneRec framework is to bridge the semantic and functional gap between traditional recommendation systems and general intelligence by treating recommendation as an autoregressive generation problem within a Large Language Model (LLM) framework. The theoretical basis is that LLMs, with their vast world knowledge, reasoning capabilities, and instruction-following prowess, can transcend the limitations of domain-specific recommenders that are constrained by isolated data and lack broader intelligence.

The intuition is that if LLMs can process and generate natural language, they should also be able to process and generate item identifiers (or representations) if these items are properly aligned with the LLM's token space. This involves:

  1. Bridging the Modality Gap: Representing discrete items as itemic tokens that the LLM can understand and generate, akin to natural language tokens.
  2. Unified Generative Formulation: Framing all recommendation tasks, from basic prediction to complex reasoning, as a next-token prediction problem, allowing a single LLM architecture to handle diverse scenarios.
  3. Scalable Pre-training: Integrating collaborative signals and general semantics during pre-training to endow the LLM backbone with intrinsic recommendation capabilities while preserving its general world knowledge.
  4. Hybrid Post-training: Employing a multi-stage post-training process that includes supervised fine-tuning, on-policy distillation, and reinforcement learning to simultaneously enhance instruction-following for recommendation tasks and restore general intelligence capabilities, effectively balancing domain-specific precision with broader reasoning.
  5. Holistic Evaluation: Developing a comprehensive benchmark (RecIF-Bench) to thoroughly assess these integrated capabilities, pushing beyond traditional ranking metrics to include instruction following and explanation generation.

4.2. Core Methodology In-depth (Layer by Layer)

The OpenOneRec framework is structured into three main phases: Pre-Training, Post-Training, and Evaluation, as depicted in Figure 2.

4.2.1. Items as Tokens: Bridging the Modality Gap

A foundational challenge in applying LLMs to recommendation is the inherent mismatch between the continuous semantic space of language and the discrete identifier space of items. Representing items with extensive textual descriptions is inefficient due to long context lengths for user histories and doesn't guarantee the generation of in-corpus items.

To address this, the paper adopts Itemic Tokens (Luo et al., 2025; Zhou et al., 2025a), treating items as a distinct modality. This involves:

  • Hierarchical Quantization: The RQ-Kmeans (Luo et al., 2025) algorithm is used to discretize pre-trained semantic embeddings of item metadata (e.g., text, visual features) into hierarchical discrete codes. This process compresses item semantics into short, fixed-length sequences of tokens.
  • Vocabulary Extension: These itemic tokens are appended to the LLM's original vocabulary, creating a unified vocabulary V=VtextVitem\mathcal { V } = \mathcal { V } _ { t e x t } \cup \mathcal { V } _ { i t e m }.
  • Benefits:
    • Efficient Long-Context Modeling: Short itemic token sequences allow for processing longer user histories.
    • Preservation of Collaborative Structure: The hierarchical nature ensures items with similar semantics share common prefixes, enabling the model to transfer knowledge based on token proximity, analogous to natural language.
    • In-corpus Generation: Facilitates the generation of actual, known items.

4.2.2. Recommendation as Autoregressive Modeling

With items represented as tokens, the recommendation task is unified into a standard autoregressive generation problem. A user's interaction history is treated as a long context sequence comprising item tokens, optionally interleaved with text.

Tasks, from prediction (e.g., retrieval) to reasoning (e.g., explanation), are formulated as a Next-Token Prediction problem. Given an instruction I\boldsymbol { \mathcal { I } } and a user context CC (which includes interaction history and optional queries), the model maximizes the likelihood of the target response YY. The objective function is: $ \mathcal { L } ( \theta ) = - \sum _ { t = 1 } ^ { | Y | } \log P _ { \theta } ( y _ { t } | \mathcal { I } , C , y _ { < t } ) $ Where:

  • L(θ)\mathcal{L}(\theta): The negative log-likelihood loss, which the model aims to minimize. θ\theta represents the parameters of the model.

  • Y|Y|: The length of the target response sequence.

  • Pθ(ytI,C,y<t)P_{\theta}(y_t | \mathcal{I}, C, y_{<t}): The probability of predicting the tt-th token yty_t in the target sequence, conditioned on the task instruction I\mathcal{I}, the user context CC, and all previously generated tokens y<ty_{<t} (i.e., tokens y1,,yt1y_1, \ldots, y_{t-1}).

  • yty_t: The tt-th token in the target response sequence.

  • I\mathcal{I}: The task-specific instruction provided to the model.

  • CC: The personalized user context, which can be the user's interaction history (Hu\mathcal{H}_u) or their User Portrait (Pu\mathcal{P}_u).

  • y<ty_{<t}: The sequence of tokens generated before the current token yty_t.

    This formulation allows leveraging standard Transformer architectures and LLM training infrastructure without task-specific architectural changes.

4.2.3. RecIF-Bench: Task Taxonomy

RecIF-Bench organizes 8 distinct tasks into a four-layer capability hierarchy to assess different levels of intelligence in recommendation foundation models. The tasks are formalized as sequence-to-sequence problems Y=F(X)Y = { \mathcal { F } } ( X ), where X=[Iˉ,C]X = [ \bar { \mathcal { I } } , C ] is the input (instruction and context) and YY is the target (item ID, sequence of IDs, or natural language explanation).

The four layers are:

  • Layer 0: Semantic Alignment (Item Understanding)
    • Input (X): Item ii.
    • Target (Y): Item Description (textual metadata).
    • Evaluation Focus: Verifies if the model can map itemic tokens to natural language, bridging the modality gap.
  • Layer 1: Fundamental Recommendation
    • Short Video Recommendation
      • Input (X): History Hvideo\mathcal{H}_{video} (user's video viewing history).
      • Target (Y): Next item ivideoi_{video}.
      • Evaluation Focus: Canonical next-item prediction within a single domain.
    • Ad Recommendation
      • Input (X): History Hvideo+Had\mathcal{H}_{video} + \mathcal{H}_{ad} (video viewing history plus ad click history).
      • Target (Y): Next item iadi_{ad}.
      • Evaluation Focus: Cross-domain interest transfer and holistic user modeling.
    • Product Recommendation
      • Input (X): History Hvideo+Hproduct\mathcal{H}_{video} + \mathcal{H}_{product} (video viewing history plus product click history).
      • Target (Y): Next item iproducti_{product}.
      • Evaluation Focus: Cross-domain interest transfer and holistic user modeling.
    • Label Prediction
      • Input (X): History Hvideo+Item ivideo\mathcal{H}_{video} + \text{Item } i_{video}.
      • Target (Y): Binary (Yes/No) engagement.
      • Evaluation Focus: Pointwise estimation complementing generative recommendation.
  • Layer 2: Instruction Following
    • Interactive Recommendation
      • Input (X): Portrait P+Query q\mathcal{P} + \text{Query } q (user profile plus natural language intent).
      • Target (Y): Item ii user engages with.
      • Evaluation Focus: Model's ability to adapt predictions based on explicit natural language instructions.
    • Label-Conditional Recommendation
      • Input (X): History Hvideo+Action a\mathcal{H}_{video} + \text{Action } a (video history plus target behavior label, e.g., "Like").
      • Target (Y): Item ii with action aa.
      • Evaluation Focus: Fine-grained behavior modeling based on specified actions.
  • Layer 3: Reasoning
    • Recommendation Explanation
      • Input (X): Portrait P+Hvideo+Item i\mathcal{P} + \mathcal{H}_{video} + \text{Item } i.

      • Target (Y): Explanation (natural language justification).

      • Evaluation Focus: Model's ability to synthesize information and articulate user-item compatibility reasoning.

        Figure 7 from the original paper (Figure 4 in original) provides a visual summary of the Task Taxonomy of RecIF-Bench.

        Figur 4 | Task Taxonomy of RecF-Bench. Weorganize 8 tasks across 4 capability layers, speing the instruction, context, and target. 该图像是示意图,展示了 RecIF-Bench 中 8 个任务的分类架构,涵盖了 4 个能力层级,并展示了任务之间的关系与结构。

Figure 7 | Task Taxonomy of RecIF-Bench. Weorganize 8 tasks across 4 capability layers, speing the instruction, context, and target.

4.2.4. Pre-Training

The pre-training phase is designed to inject recommendation knowledge and align itemic tokens while preserving general world knowledge. It uses a Qwen3 architecture as its backbone and is structured into two stages.

4.2.4.1. Pre-training Data

The pre-training corpus consists of Recommendation Corpora and General-Domain Corpora.

  • Recommendation Corpora: Derived from anonymized user logs from Kuaishou, covering user-side, item-side, and interaction-side metadata.
    • Itemic Tokenization: Item multimodal embeddings are quantized into 3-layer hierarchical itemic tokens using RQ-Kmeans, with a codebook size of 8192 per layer. Each item ii is mapped to a tuple of codes Si=(c1,c2,c3)S _ { i } = \left( c _ { 1 } , c _ { 2 } , c _ { 3 } \right), flattened into a token sequence wrapped by special tokens (e.g., <|item_begin|><item_a_C1><item_b_C2><item_c_C3><|item_end|>).
    • Data Types:
      • Itemic Dense Caption Data: Trains the model to generate natural-language captions given itemic tokens, establishing a semantic bridge.
      • Sequential User Behavior Data: Captures chronological user-item interactions (views, likes, shares) for next-item prediction and learning collaborative filtering signals.
      • Interleaved User Persona Grounding Data: Constructs narrative-style User Portraits by interleaving discrete item representations with heterogeneous user metadata (demographics, search history, interaction sequences, summarized interests). This fosters deeper semantic grounding between user characteristics and behaviors.
    • Scale: The primary training corpus for the standard OneRec variant involves ~160k users, 13 million item captions, and corresponding interactions. For the OneRec-Pro variant, this scales to ~20 million users and 98 million item captions, leveraging an in-house industrial corpus.
  • General-Domain Corpora: High-quality general-domain text corpora are mixed in during co-pretraining to mitigate catastrophic forgetting and maintain general intelligence.
    • Diversity: Includes multiple languages (Chinese, English) and specialized domains (Mathematic, Medical).
    • Reasoning-Intensive Data: Prioritizes data like mathematical derivations, logical puzzles, and code-centric corpora to enhance reasoning capabilities.
    • Deduplication: The MinHash algorithm is used to filter out general-domain samples similar to evaluation benchmarks, ensuring reliable generalization.

4.2.4.2. Training Recipe

The pre-training uses the Qwen3 architecture. Two model variants are developed: OneRec (standard, 33B tokens, 41.3 million samples from public datasets) and OneRec-Pro (enhanced, 130B tokens, 179.1 million samples from in-house corpus).

The pre-training methodology involves two distinct stages:

  • Stage 1: Itemic-Text Alignment:
    • Objective: Establish a preliminary alignment between itemic tokens and text tokens.
    • Process: The vocabulary is expanded by appending itemic special tokens to the Qwen3 tokenizer. The embedding parameters for these itemic tokens are initialized from a multivariate normal distribution (mean and covariance of existing embeddings). Only these itemic token embedding parameters are trainable, while all other model parameters are frozen. For larger models (8B+), output projection parameters corresponding to itemic tokens are also trainable.
  • Stage 2: Full-Parameter Co-Pretraining:
    • Objective: Inject recommendation knowledge while preserving general world knowledge.
    • Process: All model parameters are unfrozen. Full-parameter pre-training is conducted on a mixed dataset of recommendation-domain samples and a considerable proportion of general-domain knowledge data to prevent catastrophic forgetting.
  • Training Recipe Details:
    • Optimizer: AdamW with β1=0.9\beta _ { 1 } = 0 . 9, β2=0.95\beta _ { 2 } = 0 . 9 5, and weight decay of 0.1.
    • Learning Rate Schedule: Cosine decay with a linear warmup phase.
      • Peak LR: 1×1031 \times 1 0 ^ { - 3 } for Stage 1, 1×1041 \times 1 0 ^ { - 4 } for Stage 2.
      • Minimum LR: 1×1041 \times 1 0 ^ { - 4 } for Stage 1, 2×1052 \times 1 0 ^ { - 5 } for Stage 2.
      • Warmup Duration: First 10% of training steps.
    • Maximum Context Length: 32K tokens to accommodate long sequential user behavior data.

4.2.4.3. Scaling Laws in Recommendation

To optimize the allocation of compute budget CC between model parameters NN and training tokens DD, the paper follows Hoffmann et al. (2022). The Qwen3 architecture is evaluated across N{0.6,1.7,4,8,14}×109N \in \{ 0 . 6 , 1 . 7 , 4 , 8 , 1 4 \} \times 1 0 ^ { 9 } parameters. The compute budget is approximated as C6NDC \approx 6 N D.

The compute-optimal frontier is derived from the convex hull of the final training loss. Power-law scaling relations are fitted: $ N _ { \mathrm { o p t } } \propto C ^ { a } , \quad D _ { \mathrm { o p t } } \propto C ^ { b } $

  • Scaling Laws on Recommendation Data: Empirical fit yields exponents:

    • NoptC0.44N _ { \mathrm { o p t } } \propto C ^ { 0 . 4 4 }
    • DoptC0.56D _ { \mathrm { o p t } } \propto C ^ { 0 . 5 6 } This indicates a data-intensive scaling regime (b>ab > a), suggesting that optimal compute in recommendation requires scaling data volume more aggressively than model parameters, a deviation from Chinchilla scaling laws (which imply a0.5,b0.5a \approx 0.5, b \approx 0.5).
  • Parametric Fit and Interpretation: The final loss L(N, D) is modeled using a parametric function: $ L ( N , D ) = E + \frac { A } { N ^ { \alpha } } + \frac { B } { D ^ { \beta } } $ Where:

    • EE: Irreducible loss floor (minimum possible loss).

    • ANα\frac{A}{N^\alpha}: Component of loss due to finite model capacity, where AA is a coefficient and α\alpha is the model capacity exponent.

    • BDβ\frac{B}{D^\beta}: Component of loss due to finite data size, where BB is a coefficient and β\beta is the data exponent.

      Fitting this to experimental data yields: $ L ( N , D ) = 0 . 4 2 3 2 + \frac { 5 0 2 . 3 2 } { N ^ { 0 . 3 3 2 5 } } + \frac { 7 . 0 2 } { D ^ { 0 . 1 8 6 5 } } $ From these coefficients, three insights are derived:

    • Data-Hungry Scaling (α>β\alpha > \beta): The model capacity exponent (α0.33\alpha \approx 0.33) is consistent with LLM literature, but the data exponent (β0.19\beta \approx 0.19) is lower than typical text-domain values (βtext0.28\beta_{text} \approx 0.28). Since α>β\alpha > \beta, this necessitates b>0.5b > 0.5, confirming the need for more aggressive data scaling.

    • Impact of Warm-Starting (High A, Low B): The imbalance between AA (502.32) and BB (7.02) is attributed to transfer learning from the Qwen3 backbone, which lowers initial data entropy (low BB). The inflated AA captures performance gains from larger models being trained with more data.

    • Low Entropy of Recommendation Tasks (Low E): The estimated irreducible loss floor E=0.42E = 0.42 is substantially lower than for natural text (1.69\approx 1.69). This suggests recommendation tasks with structured features (like Itemic Dense Captions) have lower inherent entropy, allowing the model to approach a more deterministic state, emphasizing the need for diverse and high-quality recommendation corpora.

4.2.5. Post-Training

After pre-training, the model aligns itemic tokens and encodes collaborative filtering signals but may still lack advanced instruction-following and reasoning capabilities. The post-training phase (Figure 6) aims to enhance recommendation capabilities and restore general task performance using identical training data and strategies for both OneRec and OneRec-Pro.

Figure 6 | Post-training pipeline of the OneRec series models. 该图像是OneRec系列模型的后训练流程示意图,展示了模型从预训练到多任务微调、一般策略蒸馏,再到推荐任务的转变过程。每个阶段的处理流程清晰可见,强调了推荐系统的构建步骤。

Figure 6 | Post-training pipeline of the OneRec series models.

4.2.5.1. Multi-task Supervised Fine-tuning (SFT)

  • Objective: Restore and enhance foundational instruction-following and reasoning capabilities across both general and recommendation domains, establishing a robust base for subsequent stages.
  • Data: A specialized SFT corpus is curated by blending complex instruction-response pairs from RecIF-Bench-specific cleaned metadata (from 160K users) with high-quality open-source general-domain datasets focusing on instruction-following and complex reasoning. This corpus ensures no leakage from evaluation benchmarks.
  • Format: All instances are organized into a conversational format and serialized using the Qwen3 chat template.
  • Training: Fine-tuning on this unified dataset uses a recipe consistent with pre-training but with a reduced learning rate (from 2×1052 \times 1 0 ^ { - 5 } to 5×1065 \times 1 0 ^ { - 6 }).
  • Outcome: Successfully resuscitates instruction-following. Reasoning ability from general-domain data cross-fertilizes with recommendation tasks, leading to coherent reasoning trajectories for complex recommendation queries even without explicit supervision in recommendation samples.

4.2.5.2. On-policy Distillation for General Capability

  • Objective: Address a persistent capability gap in general-domain reasoning observed after SFT, likely due to distributional shift and RL-initialized backbones.

  • Method: On-policy distillation via Policy Gradient. Unlike off-policy distillation (learning from static pre-generated data), on-policy distillation involves the student model generating its own trajectories, which are then evaluated and supervised by a teacher.

  • Objective Function: Per-token reverse KL divergence between the student's distribution (πθ\pi_{\theta}) and the teacher's (πteacher\pi_{\mathrm{teacher}}). $ \mathbb { D } _ { K L } \left( \pi _ { \theta } \parallel \pi _ { \mathsf { t e a c h e r } } \right) = \mathbb { E } _ { x \sim \pi _ { \theta } } \left[ \log \pi _ { \theta } ( x _ { t + 1 } | x _ { 1 . . t } ) - \log \pi _ { \mathsf { t e a c h e r } } ( x _ { t + 1 } | x _ { 1 . . t } ) \right] $ Where:

    • DKL\mathbb{D}_{KL}: The Kullback-Leibler (KL) divergence, a measure of how one probability distribution (πθ\pi_{\theta}) is different from a second, reference probability distribution (πteacher\pi_{\mathrm{teacher}}). Minimizing this divergence makes the student policy's outputs similar to the teacher's.
    • πθ\pi_{\theta}: The student policy (the model being trained), parameterized by θ\theta.
    • πteacher\pi_{\mathrm{teacher}}: The teacher policy (a pre-trained, typically stronger model).
    • Exπθ[]\mathbb{E}_{x \sim \pi_{\theta}}[\cdot]: Expected value over trajectories xx sampled from the student policy πθ\pi_{\theta}.
    • logPθ(xt+1x1..t)\log P_{\theta}(x_{t+1}|x_{1..t}): Log probability of the next token xt+1x_{t+1} given the current trajectory prefix x1..tx_{1..t} under the student policy.
    • logPteacher(xt+1x1..t)\log P_{\mathrm{teacher}}(x_{t+1}|x_{1..t}): Log probability of the next token xt+1x_{t+1} given the current trajectory prefix x1..tx_{1..t} under the teacher policy.
  • Policy Gradient Optimization: The policy πθ\pi_{\theta} is optimized using policy gradient methods. For each input prompt qq, a trajectory oo is sampled, and the reverse KL divergence is used as a reward signal. The objective is to maximize the expected reward via gradient ascent: $ \nabla _ { \theta } J ( \theta ) = \mathbb { E } _ { o \sim \mathcal { D } , x \sim \pi _ { \theta } } \left[ \sum _ { t = 1 } ^ { T } \nabla _ { \theta } \log \pi _ { \theta } ( x _ { t } | o , x _ { < t } ) \cdot R _ { K L } ( o , x ) \right] $ Where:

    • θJ(θ)\nabla_{\theta} J(\theta): The gradient of the objective function J(θ)J(\theta) with respect to model parameters θ\theta. Maximizing this objective means updating θ\theta to increase the likelihood of actions that lead to higher rewards.
    • EoD,xπθ[]\mathbb{E}_{o \sim \mathcal{D}, x \sim \pi_{\theta}}[\cdot]: Expected value over prompts oo sampled from dataset D\mathcal{D} and trajectories xx sampled from the student policy πθ\pi_{\theta}.
    • TT: The length of the sampled trajectory xx.
    • θlogπθ(xto,x<t)\nabla_{\theta} \log \pi_{\theta}(x_t | o, x_{<t}): The gradient of the log-probability of taking action xtx_t at time tt, given the prompt oo and previous actions x<tx_{<t}, under the student policy. This is the "score function" or "log-likelihood gradient."
    • RKL(o,x)R_{KL}(o, x): The per-token reward derived from the teacher's distribution for the trajectory xx generated from prompt oo.
  • Reward Clipping: To mitigate numerical instability from extreme log-probability ratios, a clipping mechanism is applied to the reverse KL divergence: $ R _ { K L } ( o , x ) = \mathrm { c l i p } \left( - \mathbb { D } _ { K L } ( \pi _ { \theta } | | \pi _ { \mathrm { t e a c h e r } } ) , \alpha , \beta \right) $ Where:

    • clip(,α,β)\mathrm{clip}(\cdot, \alpha, \beta): A function that limits the value of its input. If the input is less than α\alpha, it becomes α\alpha; if it's greater than β\beta, it becomes β\beta; otherwise, it remains unchanged.

    • α,β\alpha, \beta: Lower and upper clipping thresholds. This prevents outlier reward signals from destabilizing training.

      The comprehensive pipeline is illustrated in Figure 10 from the original paper (Figure 7 in original).

      Figure 7 | The pipeline of On-policy Distillation via Policy Gradient. The student model (Policy Model) samples a trajectory o for the given prompt \(q\) , while the Teacher Model provides feedback through Reverse KL divergence as reward \(r\) The Policy Model is iteratively optimized using policy gradient methods based on reward. 该图像是示意图,展示了通过策略梯度进行在线政策蒸馏的流程。图中,学生模型(Policy Model)针对给定的提示 qq 采样出轨迹 oo,同时教师模型通过反向KL散度提供反馈作为奖励 rr。学生模型基于奖励 rr 通过策略梯度方法进行迭代优化。

    Figure 7 | The pipeline of On-policy Distillation via Policy Gradient. The student model (Policy Model) samples a trajectory o for the given prompt qq , while the Teacher Model provides feedback through Reverse KL divergence as reward rr The Policy Model is iteratively optimized using policy gradient methods based on reward.

  • On-Policy Distillation on General-Domain:

    • Teacher Model: The original Qwen3 model (of the same parameter scale) serves as the teacher (πteacher\pi_{\mathrm{teacher}}).
    • Vocabulary Discrepancy Strategy: The Qwen3 teacher cannot recognize itemic tokens. To handle this:
      • Prompt Selection: Queries qq are sampled exclusively from general-domain datasets, expecting pure text generation.
      • Itemic Token Penalty & Truncation: If an itemic token appears in a sampled trajectory oo at step tt, logπteacher(xtx<t)\log \pi _ { \mathrm { t e a c h e r } } ( x _ { t } | \boldsymbol x _ { < t } ) is set to a minimal value (e.g., -1e9) to simulate zero probability, and the trajectory is truncated. This provides a strong negative signal.
      • Enhanced Exploration: A high temperature coefficient during sampling encourages exploration, allowing the distillation process to identify and correct itemic token activations in general-domain tasks.
    • Thinking Paradigms: To recover instruction-following on thinking, a suffix (/think, /no_think, or empty) is randomly appended to user prompts to align with forced-thinking, non-thinking, and auto-thinking paradigms, as described in the Qwen3 technical report.

4.2.5.3. Reinforcement Learning for Recommendation (Rec-RL)

  • Objective: Directly optimize discrete ranking metrics (e.g., Recall, NDCG) that SFT doesn't directly address, which often suffers from exposure bias and struggles to distinguish "near-misses."

  • Method: Group Relative Policy Optimization (GRPO) (Shao et al., 2024). This framework avoids a separate critic model by computing the advantage of a response relative to a group of sampled trajectories for the same prompt, reducing computational overhead while maintaining stability.

  • Objective Function: Maximize: $ \mathcal { L } _ { G R P O } ( \theta ) = \frac { 1 } { G } \sum _ { i = 1 } ^ { G } \left( \mathsf { A d v } _ { i } \cdot \log \pi _ { \theta } ( R _ { i } | q ) \right) - \beta \cdot K L ( \pi _ { \theta } | | \pi _ { r e f } ) $ Where:

    • LGRPO(θ)\mathcal{L}_{GRPO}(\theta): The GRPO objective function to be maximized.
    • GG: The number of candidate responses sampled for a given prompt qq.
    • RiR_i: The ii-th sampled candidate response.
    • Advi\mathsf{Adv}_i: The relative advantage of response RiR_i, calculated by normalizing rewards within the group. This term dictates how much the probability of RiR_i should be increased or decreased.
    • logπθ(Riq)\log \pi_{\theta}(R_i | q): The log-probability of generating response RiR_i given prompt qq under the current policy πθ\pi_{\theta}.
    • β\beta: A hyperparameter controlling the strength of the KL penalty.
    • KL(πθπref)KL(\pi_{\theta} || \pi_{ref}): The KL divergence between the current policy πθ\pi_{\theta} and a reference policy πref\pi_{ref}. This penalty encourages the new policy not to deviate too far from the previous, stable policy (after on-policy distillation), preventing instability and catastrophic forgetting of general intelligence.
  • Rule-based Recommendation Reward: A sparse, rule-based reward function r(Ri)r(R_i) is designed for the 5 core recommendation tasks (Short Video Rec, Ad Rec, Product Rec, Interactive Rec, Label-Conditional Rec). The reward is focused on "Hit" events: $ r ( R _ { i } ) = \left{ \begin{array} { l l } { + 1 . 0 } & { { \mathrm { i f ~ t h e ~ t a r g e t ~ i t e m i c ~ t o k e n ~ } } s \in R _ { i } } \ { 0 . 0 } & { { \mathrm { o t h e r w i s e } } } \end{array} \right. $ This encourages the model to assign higher probability mass to itemic tokens that lead to successful hits, effectively performing "Soft Ranking" within the generative space.

  • Implementation Details:

    • Initialization: The RL trainer is initialized with the model after on-policy distillation.
    • KL Penalty: A strict KL penalty (β\beta) against πref\pi_{ref} is maintained to prevent sacrificing general intelligence for domain-specific precision.
    • Dataset: Uses the same dataset as the SFT stage.
  • Outcome: Significant boost in recommendation metrics, aligning generative behavior with recommender system goals.

5. Experimental Setup

5.1. Datasets

The experiments in OpenOneRec utilize several datasets to evaluate both recommendation-specific and general intelligence capabilities.

5.1.1. RecIF-Bench

RecIF-Bench is a newly proposed, comprehensive benchmark designed to rigorously evaluate recommendation foundation models.

  • Source & Scale: Aggregates approximately 120 million (120M) interactions from 200,000 (200K) distinct users.

  • Domain Coverage: Spans three heterogeneous industrial domains, each capturing different user behavior patterns:

    • Short Video (Content Domain): Short-form videos from Kuaishou, including viewing behaviors across various app tabs. Provides impression sequences with corresponding interaction types.
    • Ad (Commercial Domain): Promotional short videos sponsored by advertisers on the Kuaishou platform, typically with clickable redirects. Provides click sequences of user ad click behaviors.
    • Product (E-commerce Domain): Products listed in the Kuaishou Mall. Provides click sequences of user product click behaviors.
  • Rich Metadata: Beyond interaction logs, RecIF-Bench provides comprehensive metadata:

    • User-side: Each user has a User Portrait—a unified narrative interleaving natural language descriptions with itemic tokens. This portrait includes demographics (gender, age), content creation history, recent searches, followed creator types, viewing preferences, comments, livestream views, purchase records, shopping cart items, local service coupons, ad exposures, and commercial intent signals.
    • Item-side: Each item is associated with multimodal embeddings (4096-dim text embedding and 5-frame visual embeddings with 1152-dim per frame). Dense captions are also provided for approximately 13 million videos.
    • Interaction-side: For each user-video pair in the exposure sequence, multi-label behavioral signals are recorded, including like, follow, comment, effective view, and dislike.
  • Itemic Tokenization: All items are pre-tokenized into tuples of discrete tokens s = ( c _ { 1 } , c _ { 2 } , . . . , c _ { k } ) using a hierarchical quantization strategy, enabling direct consumption by LLM-based recommenders. The paper notes flexibility for researchers to train custom itemic tokenizers or use traditional item IDs.

  • Data Splitting Strategy: A strict user-based splitting strategy is used. 20% of users are randomly selected as the held-out test set, ensuring zero leakage. For each user, interactions are partitioned temporally: interactions before a timestamp form the history H\mathcal{H}, and those after serve as the target YY.

    The following are the statistics of RecIF-Bench from Table 1 of the original paper:

    Domain# Users# Items# InteractionsAvg. Hist. ItemAvg. Tgt. Item
    Short Video195,02613,107,67594,443,611458.18.6
    Ad151,259177,5485,341,91129.95.5
    Product144,3072,055,24020,087,210132.56.7
    Total202,35915,340,463119,872,732574.917.5

The following are the data distribution analysis of RecIF-Bench from Figure 3 of the original paper:

Figure 3 | Data distribution analysis of RecIF-Bench. (a) Item popularity distribution (log-log scale) across domains. (b-d) Distribution of user history lengths for Short Video, Ad, and Product domains, respectively. 该图像是图表,展示了 RecIF-Bench 数据分布分析。其中 (a) 显示各域项目的热门程度分布(对数-对数坐标),(b-d) 展示短视频、广告和产品域的用户历史长度分布。各子图反映了不同类型项目的互动计数和历史长度特点。

Figure 3 | Data distribution analysis of RecIF-Bench. (a) Item popularity distribution (log-log scale) across domains. (b-d) Distribution of user history lengths for Short Video, Ad, and Product domains, respectively.

5.1.2. Amazon Benchmark

To evaluate cross-domain transferability and generalization capabilities, the models are tested on 10 real-world datasets from the popular Amazon review benchmark (McAuley et al., 2015).

  • Domains: Baby, Beauty, Cell Phones and Accessories, Grocery and Gourmet Food, Health and Personal Care, Home and Kitchen, Pet Supplies, Sports and Outdoors, Tools and Home Improvement, and Toys and Games.
  • Purpose: These diverse domains rigorously validate whether the foundation model, pre-trained on open-domain data, provides a fundamental transfer advantage for specific downstream recommendation distributions.
  • Data Pre-processing: Sparse users and items with fewer than 5 interactions are discarded.
  • Splitting Strategy: Leave-one-out strategy (Rajput et al., 2023; Wang et al., 2024) for training and evaluation in the sequential recommendation setting.

5.1.3. General Intelligence Sanity Check

To ensure that recommendation-focused models retain general intelligence, a sanity check suite is included, covering four categories:

  • Math & Text Reasoning: MATH-500, GSM8K, and AIME'24.
  • General Tasks: MMLU-Pro and GPQA-Diamond.
  • Alignment: IFEVAL (strict prompt).
  • Coding: LiveCodeBench v5.

5.1.4. Pre-training Data

The pre-training data for OneRec (standard variant) consists of 33B tokens across 41.3 million samples, derived from a mixture of general-domain and recommendation-domain corpora. OneRec-Pro uses 130B tokens and 179.1 million samples.

The following are the detailed data composition and token budgets for pre-training from Table 13 of the original paper:

DatasetWeight (%)CategorySubtotal (%)Token Budget
Nemotron_CC_Math_v137.41%Math62.34%29B
Nemotron_Pretraining_Code_v112.91%Code
Nemotron_CC_v25.59%Math, General, Code4B
reasoning_v1_20m4.04%General
OpenMathReasoning1.16%Math
NuminaMath-QwQ-CoT-5M0.79%Math
KodCode_V1_SFT_R10.26%Code
Chinese-Reasoning-Distil-Data0.09%General
medical-o1-reasoning-SFT0.09%Medical
Itemic Dense Caption Data12.01%Reco37.66%
Interleaved User Persona Grounding Data10.03%Reco
Sequential User Behavior Data15.62%Reco
Total100%100%33B

The following are the data composition and token budgets for pre-training stages from Table 14 of the original paper:

ModelStageTrainingGeneral-DomainReco-DomainToken Budget
OneRecstage1Itemic-related Parameters--16B
stage2Full-Parameter62.34%37.66%33B
OneRec-Prostage1Itemic-related Parameters--30B
stage2Full-Parameter53%47%130B

5.1.5. Multi-task SFT Data

The Multi-task Supervised Fine-tuning (SFT) stage uses a blended corpus of general-domain reasoning samples and recommendation-specific tasks.

The following are the data mixture for Multi-task SFT from Table 15 of the original paper:

DatasetWeight (%)CategorySubtotal (%)
OpenMathReasoning12.971%Math64.978%
R1-Distill-SFT12.784%General Instruction
Infinity_Instruct11.359%Instruction
OpenCoderReasoning11.130%Code
Chinese-Reasoning-Distil-Data4.552%General
Reasoning_Multi_subject_RLVR4.376%Multi-subject
Reasoning_KodCode_V1_SFT_R14.167%Code
DeepMath103K2.362%Math
medical-o1-reasoning-SFT1.277%Medical
Label prediction7.800%Reco35.022%
SID to Caption generation7.493%Reco
Interactive recommendation6.392%Reco
Video recommendation3.971%Reco
Label conditional recommendation3.575%Reco
Total100%100%

5.2. Evaluation Metrics

The paper employs a dual-metric evaluation system to cover both recommendation accuracy and generation quality.

5.2.1. Recommendation Metrics (for Layer 1 & 2 tasks)

  • Pass@K:
    • Conceptual Definition: Pass@K measures whether the ground truth item is present among the top KK items generated or recommended by the model. It is a binary hit/miss metric. If any of the KK predicted items matches a ground truth relevant item, it's considered a "pass."
    • Mathematical Formula: $ \mathrm{Pass@K} = \frac{1}{|U|} \sum_{u \in U} \mathbb{I}(\text{relevant_item}{u} \in \text{TopK_predictions}{u}) $
    • Symbol Explanation:
      • U|U|: Total number of users in the evaluation set.
      • I()\mathbb{I}(\cdot): An indicator function that returns 1 if the condition inside is true, and 0 otherwise.
      • relevant_itemu\text{relevant\_item}_{u}: The actual relevant item(s) for user uu in the test set.
      • TopK_predictionsu\text{TopK\_predictions}_{u}: The set of top KK items recommended by the model for user uu.
  • Recall@K:
    • Conceptual Definition: Recall@K measures the proportion of relevant items that are successfully retrieved by the recommendation system within its top KK recommendations. It focuses on how many of the truly relevant items the model was able to "find."
    • Mathematical Formula: $ \mathrm{Recall@K} = \frac{1}{|U|} \sum_{u \in U} \frac{|\text{relevant_items}{u} \cap \text{TopK_predictions}{u}|}{|\text{relevant_items}_{u}|} $
    • Symbol Explanation:
      • U|U|: Total number of users in the evaluation set.
      • relevant_itemsu\text{relevant\_items}_{u}: The set of all relevant items for user uu in the test set.
      • TopK_predictionsu\text{TopK\_predictions}_{u}: The set of top KK items recommended by the model for user uu.
      • |\cdot|: Denotes the cardinality (number of elements) of a set.
      • \cap: Set intersection operator.
  • AUC (Area Under the Receiver Operating Characteristic Curve):
    • Conceptual Definition: AUC is a commonly used metric for binary classification problems (like Label Prediction). It measures the ability of a model to distinguish between positive and negative classes across all possible classification thresholds. A higher AUC indicates better discrimination power. An AUC of 0.5 means the model performs no better than random guessing, while an AUC of 1.0 indicates perfect classification.
    • Mathematical Formula: $ \mathrm{AUC} = \frac{\sum_{i \in \text{positive}} \sum_{j \in \text{negative}} \mathbb{I}(P(i) > P(j)) + 0.5 \cdot \mathbb{I}(P(i) = P(j))}{|\text{positive}| \cdot |\text{negative}|} $
    • Symbol Explanation:
      • positive: Set of positive samples (e.g., actual engagements).
      • negative: Set of negative samples (e.g., non-engagements).
      • P(i): The predicted probability (or score) that sample ii belongs to the positive class.
      • I()\mathbb{I}(\cdot): An indicator function.
      • positive|\text{positive}|: Number of positive samples.
      • negative|\text{negative}|: Number of negative samples.
  • NDCG@K (Normalized Discounted Cumulative Gain at K):
    • Conceptual Definition: NDCG@K is a measure of ranking quality, especially useful when relevance judgments are graded (e.g., highly relevant, somewhat relevant, not relevant). It takes into account both the relevance of recommended items and their position in the ranked list. Highly relevant items ranked higher contribute more to the score. It normalizes the DCG by the ideal DCG to get a score between 0 and 1.
    • Mathematical Formula: $ \mathrm{DCG@K} = \sum_{i=1}^{K} \frac{2^{\mathrm{rel}_i} - 1}{\log_2(i+1)} $ $ \mathrm{NDCG@K} = \frac{\mathrm{DCG@K}}{\mathrm{IDCG@K}} $
    • Symbol Explanation:
      • reli\mathrm{rel}_i: The relevance score of the item at position ii in the ranked list.
      • log2(i+1)\log_2(i+1): The discount factor for items at lower positions.
      • DCG@K\mathrm{DCG@K}: Discounted Cumulative Gain at rank KK.
      • IDCG@K\mathrm{IDCG@K}: Ideal Discounted Cumulative Gain at rank KK, which is the maximum possible DCG if all relevant items were perfectly ranked.

5.2.2. Text Generation Metrics (for Layer 0 & 3 tasks)

  • LLM-as-Judge Score:
    • Conceptual Definition: For Item Understanding and Recommendation Explanation tasks, an independent LLM (Gemini-2.5-Flash-Lite) is used as a judge to rate the quality of generated text on dimensions like accuracy and coherence. This method overcomes the limitations of traditional N-gram based metrics (like BLEU or ROUGE) which may not capture semantic quality for open-ended generation.
    • Evaluation Process (Appendix B.1):
      1. Information Point Extraction: The LLM judge decomposes both the ground truth captions (generated by Gemini-2.5-Pro) and model-generated captions into atomic Weighted Information Points (WIPs). Each WIP contains a fact statement and an importance score (1-5).
      2. Semantic Matching: The LLM judge aligns model-generated WIPs with ground truth WIPs, identifying valid matches, hallucinations (unmatched model WIPs), and omissions (unmatched GT WIPs).
      3. Weighted Scoring: A Double-Weighted F1 Score is computed. For each matched pair (wgt,wmodel)(w_{gt}, w_{model}), a match quality score q[0,1]q \in [0, 1] is calculated using BERTScore (with bert-base-chinese).
    • Mathematical Formula for F1 Score: $ TP _ { i } = \sum _ { ( w _ { g t } , w _ { m o d e l } ) \in \mathrm { M a t c h e s } } \mathrm { s c o r e } ( w _ { g t } ) \times q $ $ F N _ { i } = \sum _ { ( w _ { g t } , w _ { m o d e l } ) \in \mathrm { M a t c h e s } } \mathrm { s c o r e } ( w _ { g t } ) \times ( 1 - q ) + \sum _ { w _ { g t } \in \mathrm { U n m a t c h e d ~ G T } } \mathrm { s c o r e } ( w _ { g t } ) $ $ F P _ { i } = \sum _ { ( w _ { g t } , w _ { m o d e l } ) \in \mathrm { M a t c h e s } } \mathrm { s c o r e } ( w _ { m o d e l } ) \times ( 1 - q ) + \sum _ { w _ { m o d e l } \in \mathrm { U n m a t c h e d ~ M o d e l } } \mathrm { s c o r e } ( w _ { m o d e l } ) $ $ F 1 _ { i } = \frac { 2 \cdot T P _ { i } } { 2 \cdot T P _ { i } + F P _ { i } + F N _ { i } } $ $ \mathrm { L L M - J u d g e ~ S c o r e } = \frac { 1 } { N } \sum _ { i = 1 } ^ { N } F 1 _ { i } $
    • Symbol Explanation:
      • TPiTP_i: Weighted True Positives for sample ii. Sum of scores of matched ground truth WIPs, weighted by their BERTScore match quality qq.
      • FNiFN_i: Weighted False Negatives for sample ii. Sum of scores of matched GT WIPs weighted by (1-q) (missed aspects of matched WIPs) plus scores of completely unmatched GT WIPs (omissions).
      • FPiFP_i: Weighted False Positives for sample ii. Sum of scores of model WIPs that matched GT WIPs weighted by (1-q) (incorrect aspects of matched WIPs) plus scores of completely unmatched model WIPs (hallucinations).
      • score(w)\mathrm{score}(w): Importance score of a WIP.
      • qq: BERTScore (F1) match quality for a pair of WIPs.
      • Matches: Set of matched WIP pairs.
      • Unmatched GT: Set of ground truth WIPs that were not matched.
      • Unmatched Model: Set of model-generated WIPs that were not matched.
      • F1_i: The F1 score for sample ii.
      • LLMJudge Score\mathrm{LLM-Judge~Score}: The final average F1 score across all NN samples.

5.3. Baselines

The OneRec-Foundation models are compared against two groups of competitive baselines:

5.3.1. Discriminative Recommender Models

These are traditional models that typically focus on predicting scores or ranking items.

  • BERT4Rec (Sun et al., 2019)
  • GRU4Rec (Hidasi et al., 2016)
  • SASRec (Kang and McAuley, 2018)
  • HSTU (Zhai et al., 2024)
  • ReaRec (Tang et al., 2025)
  • Adaptation: These methods are inherently task-specific, requiring separate training for each RecIF-Bench task. They were adapted with task-specific modifications to support the diverse RecIF tasks.

5.3.2. Generative Recommender Models

These models also aim to generate recommendations, but OpenOneRec distinguishes itself by its holistic approach and foundation model capabilities.

  • TIGER (Rajput et al., 2023)
  • LC-Rec (Zheng et al., 2024): This model was specifically implemented as "LC-Rec-8B" using a comparable Qwen3-8B backbone to ensure a fair comparison at the foundation model scale.

6. Results & Analysis

6.1. Core Results Analysis

6.1.1. Main Results on RecIF-Bench

The evaluation on RecIF-Bench demonstrates OpenOneRec-Foundation's superior performance across a wide range of recommendation tasks.

The following are the unified performance comparison across all tasks from Table 4 of the original paper:

OneRec-1.7BOneRec-1.7B-ProOneRec-8BOneRec-8B-Pro
TaskMetricBest BaselineLC-Rec-8BSASRecBERT4RecGRU4RecHSTUReaRecTIGER
Short Video RecPass@10.00450.00400.00510.00430.00520.01680.03410.0496
Pass@320.09510.09930.10030.10100.10020.10610.13060.1710
Recall@320.01130.01190.01190.01200.01170.01320.01800.0272
Ad RecPass@10.00440.00610.00590.00350.00760.01250.01970.0169
Pass@320.09800.12250.11020.10540.12660.17690.20960.2037
Recall@320.02930.03810.03360.03270.04090.05810.07230.0707
Product RecPass@10.00520.00540.00470.00300.00550.01200.01780.0144
Pass@320.09140.09360.08210.09070.09140.12760.18090.1571
Recall@320.01750.01930.01610.01890.01780.02830.04160.0360
Label-Cond. RecPass@10.00260.00260.00320.00260.00270.00440.00790.0064
Pass@320.03800.03720.03930.03810.03830.03370.04200.0431
Recall@320.01400.01350.01430.01370.01390.01230.01700.0184
Label Pred.AUC0.62440.65980.66400.62040.65810.66750.61390.6184
Interactive RecPass@1------0.08900.0660
Pass@32------0.37300.3170
Recall@32------0.23940.1941
Item Understand.LLM-Judge Score------0.25170.3175
Rec. ExplanationLLM-Judge Score------3.93503.3540
  • State-of-the-Art Recommendation Performance: OneRec-Foundation consistently outperforms all baselines (both discriminative and generative) across the vast majority of tasks in RecIF-Bench. This is evident from the higher Pass@K, Recall@K, and AUC scores for OneRec-Foundation variants compared to LC-Rec-8B and other baselines.
  • Scaling Effects:
    • Data Scaling: OneRec-Pro models (trained on a larger industrial corpus) consistently surpass OneRec models of the same parameter size (e.g., OneRec-8B-Pro generally outperforms OneRec-8B). This validates the importance of data scale in enhancing recommendation capabilities.
    • Model Scaling: Larger models (8B parameters) generally outperform smaller models (1.7B parameters) across all variants (e.g., OneRec-8B > OneRec-1.7B). This confirms predictable capability scaling with increasing model size, analogous to findings in general LLMs.

6.1.2. Trade-off on General Capabilities

While excelling in recommendation, the models exhibit a trade-off in general capabilities.

The following are the performance comparison on general capability (Thinking) from Table 5 of the original paper:

CategoryTaskQwen3-1.7BOneRec-1.7BOneRec-1.7B-ProQwen3-8BOneRec-8BOneRec-8B-Pro
Math & Text ReasoningMATH-5000.87800.88400.88400.95200.94600.9380
GSM8K0.91210.89840.89990.95680.95750.9575
AIME'240.49380.41040.41460.79170.72500.7188
General TasksMMLU-Pro0.54220.35480.39320.72350.53420.5204
GPQA-Diamond0.37880.32320.33330.56060.50000.5051
Alignment TasksIFEVALstrict prompt0.69690.54710.54160.85770.78930.7634
CodingLiveCodeBench v50.39070.28320.28320.54840.49100.4667

The following are the performance comparison on general capability (Non-Thinking) from Table 6 of the original paper:

CategoryTaskQwen3-1.7BOneRec-1.7BOneRec-1.7B-ProQwen3-8BOneRec-8BOneRec-8B-Pro
Math & Text ReasoningMATH-5000.69800.70600.69400.83800.82400.7980
GSM8K0.82180.80360.81580.93030.93100.9196
AIME'240.13130.12710.12500.27290.24170.2271
General TasksMMLU-Pro0.43840.30720.28040.66320.57950.4521
GPQA-Diamond0.30300.31310.27780.39900.40400.3939
Alignment TasksIFEVALstrict prompt0.67470.47690.52500.83920.73570.7098
CodingLiveCodeBench v50.12190.12190.11470.27600.24010.2401
  • Retention of General Capabilities: The models successfully retain most of the general capabilities of the Qwen3 backbone. Performance on mathematical benchmarks (MATH-500, GSM8K) shows minimal degradation, and in some cases (e.g., GSM8K for OneRec-8B), even slight improvements over the base Qwen3-8B in Thinking mode.
  • Performance Trade-off: A noticeable performance trade-off exists, particularly in broader general knowledge (MMLU-Pro, GPQA-Diamond) and coding (LiveCodeBench v5) tasks. OneRec-Foundation variants generally score lower than their base Qwen3 counterparts in these categories. This suggests that while the distillation process effectively preserves reasoning proficiency, the limited diversity or quality of the general data used during post-training might constrain the model's broader general capabilities, indicating a need for more refined data strategies to achieve a better balance.

6.1.3. Transfer Learning on Amazon Benchmark

OneRec-Foundation demonstrates exceptional transferability and cross-domain generalization on the Amazon benchmark.

The following are the Cross-Domain generalization performance on Amazon domains from Table 7 of the original paper:

ModelMetricBabyBeautyCellGroceryHealthHomePetSportsToolsToys
SASRecR@50.02320.03930.04820.04800.02950.01330.03770.02400.02690.0420
R@100.03810.06390.07820.07890.05060.02120.06070.03890.04370.0658
N@50.01370.02090.02810.02620.01730.00700.02220.01300.01490.0217
N@100.01850.02890.03780.03610.02420.00980.02960.01780.02030.0294
BERT4RecR@50.01170.02190.03250.03070.02040.00630.02180.01510.01450.0200
R@100.02280.04190.05690.05340.03530.01130.04120.02610.02640.0362
N@50.00650.01200.01900.01740.01170.00380.01230.00830.00830.0102
N@100.01010.01850.02680.02470.01650.00540.01860.01190.01210.0154
GRU4RecR@50.02020.03220.04300.03620.02560.00900.02640.01740.01760.0266
R@100.03460.05390.06760.05910.04230.01560.04490.02780.03050.0453
N@50.01240.02010.02750.02300.01640.00580.01630.01100.01160.0171
N@100.01700.02710.03550.03030.02170.00790.02220.01440.01580.0231
HSTUR@50.02260.04560.04750.04580.03300.01340.03620.02270.02310.0489
R@100.03500.06430.07250.07120.04850.01970.05210.03470.03370.0649
N@50.01560.03080.03140.02970.02150.00920.02390.01510.01590.0339
N@100.01960.03680.03950.03780.02650.01120.02900.01900.01930.0391
ReaRecR@50.01970.04880.04440.04540.03260.01500.02990.02310.02190.0517
R@100.03200.07020.07110.07300.04810.02100.04860.03480.03100.0706
N@50.01230.03410.02690.02890.02130.01010.01890.01520.01430.0369
N@100.01630.04090.03550.03780.02630.01210.02490.01890.01730.0430
TIGERR@50.01910.04130.05400.04470.03280.01420.03430.02160.02280.0367
R@100.03180.06280.07860.06910.05340.02160.05420.03310.03440.0527
N@50.01250.02770.03500.02950.02220.00940.02320.01450.01480.0255
N@100.01620.03460.04290.03730.02890.01180.02950.01820.01840.0307
LC-RecR@50.02320.04950.05850.05010.04120.01990.03880.02690.02880.0350
R@100.03440.07640.08830.07900.06160.02930.06120.04180.04380.0549
N@50.01510.03380.03920.03280.02720.01380.02470.01770.01870.0221
N@100.01870.04240.04880.04210.03380.01680.03200.02250.02350.0285
OursR@50.03520.06460.07170.06880.05340.02790.05630.03650.04120.0693
R@100.05130.09240.10360.10290.07680.03900.08340.05470.05930.0953
N@50.02380.04560.04900.04600.03760.02020.03890.02520.02950.0496
N@100.02890.05450.05930.05700.04520.02370.04760.03100.03540.0579
Improve (%) R@1034.6↑20.9↑17.3↑30.3↑24.7↑33.1↑36.3↑30.9↑35.4↑35.0↑
  • SOTA Results Across Domains: OneRec-Foundation (Ours) achieves new state-of-the-art results across all 10 Amazon datasets.
  • Significant Improvement: The model secures an average improvement of 26.8% in Recall@10 over the second-best baseline on each domain. This empirically confirms that large-scale generative pre-training endows the model with robust transfer capabilities far exceeding traditional collaborative filtering approaches.

6.1.3.1. Adaptive Strategies for Pre-trained Model Utilization

To address the distributional shift of item identifiers in transfer learning (where the pre-trained tokenizer might not granularly distinguish items in specific target domains), three strategies were explored.

The following are the Performance Comparison of Adaptive Strategies for Pre-trained Model Utilization from Table 8 of the original paper:

StrategyMetricBabyBeautyCellGroceryHealthHomePetSportsToolsToys
Extended Residual QuantizationR@50.02880.05340.05740.05620.04790.02270.05180.03150.03500.0511
R@100.04070.07990.08300.08610.06730.03130.07580.04470.04950.0701
N@50.02010.03640.03890.03830.03330.01620.03560.02150.02430.0360
N@100.02390.04490.04710.04800.03960.01900.04330.02580.02890.0421
Text-Only AdaptationR@50.03170.06300.06880.06870.05290.02850.05480.03680.04140.0668
R@100.04480.08830.09850.10480.07520.03980.08500.05480.06150.0931
N@50.02270.04450.04730.04600.03680.01990.03820.02560.02880.0483
N@100.02690.05260.05690.05760.04400.02350.04780.03140.03540.0568
Text-Augmented Itemic TokensR@50.03520.06460.07170.06880.05340.02850.05630.03680.04140.0693
R@100.05130.09240.10360.10290.07680.03980.08340.05470.05930.0953
N@50.02380.04560.04900.04600.03760.02020.03890.02560.02950.0496
N@100.02890.05450.05930.05760.04520.02370.04780.03140.03540.0579
  • Strategy 1: Extended Residual Quantization: Extends the hierarchical depth of itemic tokens using Finite Scalar Quantization (FSQ). Reduces collision rate to 3.05%. Achieves a 10.0% improvement in average R@10 over LC-Rec. However, the non-pre-trained fourth layer disrupts original hierarchical semantics.
  • Strategy 2: Text-Only Adaptation: Bypasses itemic tokens entirely, representing items via 5 keywords from metadata. Collision rate is 4.27%. Achieves an 18.8% improvement in average R@10 over Extended Residual Quantization, as the model's linguistic core remains intact and natural language representations are more expressive in narrow domains. This sacrifices collaborative filtering signals.
  • Strategy 3: Text-Augmented Itemic Tokens: Concatenates the original three-layer pre-trained itemic tokens with keyword representations ([itemictokens]+[keywords][itemic_tokens]^+[keywords]). Preserves original itemic tokens structure, maintaining hierarchical semantics. Keywords provide semantic disambiguation (collision rate 0.47%) and enable full utilization of linguistic capabilities. This strategy achieves state-of-the-art performance across nearly all datasets, validating that effective transfer learning requires maximizing utilization of the foundation model's diverse capabilities while preserving pre-trained structural integrity.

6.1.3.2. Domain-Specific Training vs. Multi-Domain Joint Training

The paper investigates the impact of training strategies on transfer performance.

The following are the Impact of Training Strategies (Domain-Specific vs Multi-Domain Joint) and Few-Shot Learning on Transfer Performance from Figure 11 of the original paper:

Figure 8 | Impact of Training Strategies (Domain-Specific vs Multi-Domain Joint) and Few-Shot Learning on Transfer Performance. We compare OneRec-Foundation (Ours) against TIGER across four Amazon domains under three settings: (1) Few-shot learning with \(1 0 \\%\) training data, (2) Full-data training with domain-specific strategy, and (3) Full-data training with joint multi-domain strategy. The green dashed line represents the performance gain (Recall `@ 1 0` difference) of Ours over TIGER. 该图像是图表,展示了不同训练策略(少量训练、领域特定及多领域联合)对转移性能的影响。通过比较 OneRec-Foundation 和 TIGER 在四个亚马逊领域的 Recall@10,显示了我们方法在不同设置下的表现。绿虚线代表我们与 TIGER 的性能差异(Recall @ 10 之差)。

Figure 8 | Impact of Training Strategies (Domain-Specific vs Multi-Domain Joint) and Few-Shot Learning on Transfer Performance. We compare OneRec-Foundation (Ours) against TIGER across four Amazon domains under three settings: (1) Few-shot learning with 101 0 \\% training data, (2) Full-data training with domain-specific strategy, and (3) Full-data training with joint multi-domain strategy. The green dashed line represents the performance gain (Recall @ 1 0 difference) of Ours over TIGER.

  • Divergent Performance: TIGER (a traditional generative recommender) shows a consistent performance decline (average 10.6% drop in Recall@10) under joint training across multiple domains. In contrast, OneRec-Foundation demonstrates an average 2.3% improvement.
  • Foundation Model Advantage: This divergence highlights that OneRec-Foundation, with its pre-trained recommendation knowledge and semantic understanding, extracts generalizable patterns rather than memorizing domain-specific statistics. Multi-domain joint training further enriches the model by exposing it to diverse interaction patterns, enabling effective cross-domain knowledge transfer. The massive parameter capacity allows encoding domain-specific nuances while maintaining shared high-level patterns.

6.1.3.3. Few-Shot Learning: Amplified Transfer Advantage

  • Significant Gap Widening: The transfer learning advantage of foundation models is amplified under data scarcity. While OneRec-Foundation surpasses TIGER by an average of 77.7% in Recall@10 with full training data, this gap dramatically widens to 219.7% in the 10% few-shot regime.
  • Resilience under Data Constraints: OneRec-Foundation preserves 45.2% of its full-data performance with only 10% of the data, whereas TIGER retains only 23.0%. This striking resilience validates that large-scale pre-training confers robust, transferable representations that enable effective domain adaptation even under severe data constraints.

6.2. Ablation Study

6.2.1. Ablation Study on Pre-training Strategies

An ablation study quantifies the contribution of the itemic-text alignment (Stage 1) in the pre-training pipeline. The full model is compared against a variant w/o Align that bypasses this initial phase. Both variants undergo Multi-task SFT before evaluation.

The following are the Ablation study on pre-training strategies from Table 9 of the original paper:

TaskMetric0.6B1.7B8B
Oursw/o AlignOursw/o AlignOursw/o Align
Short Video RecPass@320.14010.13970.16360.16050.20340.1933
Recall@320.02100.02100.02540.02510.03340.0310
Ad RecPass@320.17400.16800.19610.19220.23500.2401
Recall@320.05860.05690.06730.06690.08210.0841
Product RecPass@320.11390.10640.15120.13950.18930.1911
Recall@320.02570.02430.03430.03120.04470.0442
Label-Cond. RecPass@320.03500.03430.04260.04010.05370.0537
Recall@320.01460.01450.01810.01710.02270.0230
Interactive RecPass@320.24600.23600.31100.30500.46500.4490
Recall@320.14020.13570.19080.17700.30390.2910
Label Pred.AUC0.64880.58070.63920.57960.68790.6285
Item UnderstandingLLM-Judge Score0.31740.31120.31700.31810.32250.3103
Rec. ExplanationLLM-Judge Score2.99602.86353.09223.31603.94203.9329
  • Semantic Bridge: Stage 1 (itemic-text alignment) acts as a fundamental semantic bridge for cold-started itemic token embeddings. Aligning these initialized parameters with the pre-trained latent space before full-parameter fine-tuning establishes a robust semantic foundation.
  • Model Size Dependence: The marginal gains from Stage 1 scale inversely with model size, meaning it's particularly necessary for smaller models (0.6B, 1.7B). Larger backbones possess greater inherent generalization capabilities, reducing the relative impact of this initial alignment stage.
  • Domain-Specific Precision: This stage remains essential for domain-specific precision, especially in Label Prediction (significant AUC drop without alignment) and Interactive Recommendation. This underscores that explicit alignment is a prerequisite for optimizing recommendation performance across all model scales.

6.2.2. Evolution of Model Capabilities Across Post-training Stages

The performance of the model is analyzed after each key post-training phase (Multi-task SFT, On-policy Distillation, Reinforcement Learning).

The following are the Performance comparison on general capability across post-training stages (Thinking) from Table 10 of the original paper:

math_500gsm8kAIME'24mmlu_progpqa_diamondIFEVALLiveCodeBench
Qwen3-8B (Base)0.9520.95680.79170.72350.56060.85770.5484
Stage 1: Multi-task SFT0.9360.90830.51040.53070.49490.61740.4516
Stage 2: On-Policy Distillation0.9480.95380.71250.54540.50.76530.4659
Stage 3: Reinforcement Learning0.9380.95750.71880.52040.50510.76340.4667

The following are the Performance comparison on general capability across post-training stages (Non-Thinking) from Table 11 of the original paper:

math_500gsm8kAIME'24mmlu_progpqa_diamondIFEVALLiveCodeBench
Qwen3-8B (Base)0.8380.93030.27290.66320.3990.83920.276
Stage 1: Multi-task SFT0.8760.9060.06880.49090.33840.56380.1756
Stage 2: On-Policy Distillation0.8480.92340.25210.5830.40910.76890.2545
Stage 3: Reinforcement Learning0.7980.91960.22710.45210.39390.70980.2401

The following are the Recommendation benchmark performance across post-training stages from Table 12 of the original paper:

ModelVideo RecAd RecProduct RecLabel Cond.InteractiveLabel Pred.Item UnderstandingReco Reason
Stage 1: Multi-task SFT0.03240.09250.05320.02290.34610.69790.32743.8795
Stage 2: On-Policy Distillation0.03040.05960.03300.02000.24190.69440.33193.9479
Stage 3: Reinforcement Learning0.03700.09670.05360.02360.34580.69080.32094.0381
  • Impact of On-policy Distillation on General Capabilities:

    • Restoration of General Capabilities: Comparing Stage 1 (Multi-task SFT) with Stage 2 (On-policy Distillation) (Tables 10 and 11) shows that on-policy distillation significantly restores general capabilities, realigning the model with the Qwen3 baseline on most general benchmarks, particularly in Math & Text Reasoning and Alignment Tasks. For example, AIME'24 (Thinking) jumps from 0.5104 (SFT) to 0.7125 (Distillation) close to Qwen3-8B's 0.7917.
    • Persistent Gap: Despite restoration, a performance gap relative to the original Qwen3 base model persists across several metrics (MMLU-Pro, GPQA-Diamond, LiveCodeBench). This is attributed to current data composition and quality during the distillation phase, suggesting potential for further improvement with refined data strategies.
    • Mitigation of Instruction Drift: Multi-task SFT occasionally led to instruction drift where the model disregarded /no_think tags and generated Chain-of-Thought (CoT) reasoning in "Non-Thinking" mode, leading to inflated scores. On-policy Distillation effectively mitigates this, restoring the model's ability to faithfully switch between distinct reasoning modes.
  • Advancements through RL for Recommendation:

    • Targeted Improvements: Stage 3 (Reinforcement Learning) demonstrates targeted improvements on core recommendation tasks (Table 12). The RL-trained model achieves consistent gains on Video Rec, Ad Rec, Product Rec, Label Cond., and Interactive Rec compared to Stage 2.
    • Optimization for Ranking Accuracy: These improvements stem from the rule-based Hit reward in Rec-RL which directly optimizes for ranking accuracy, encouraging the model to assign higher probability mass to target itemic tokens.
    • Transfer to Explanation Generation: Interestingly, the Reco Reason task (Layer 3) also benefits from RL training, showing an increase in LLM-Judge Score (from 3.9479 to 4.0381). This suggests that the refined "recommendation intuition" acquired through RL transfers to explanation generation, producing more coherent and relevant reasoning.

6.3. Data Presentation (Figure 1 from original paper)

The following are the holistic performance overview of OpenOneRec-Foundation from Figure 1 of the original paper:

Figure 1 | Holistic Performance Overview. Left: Evaluation on RecIF-Bench and general LLM bencharks. ue ive penedat sk hiey general knowledge. "Best Baseline" denotes the highest performance achieved by existing methods for each specific task. Right: Amazon Benchmark results. Our model demonstrates exceptional cross-domain transferability, consistently surpassing leading baselines across 10 diverse datasets. 该图像是图表,展示了在 RecIF-Bench 的多项任务中的能力评估(左侧)及在亚马逊基准上的召回率结果(右侧)。左侧为不同任务的表现,红色和黄色线条分别代表 OneRec-Pro-8B 和最佳基线,右侧柱状图比较了各模型在不同类别中的 Recall@10 结果。

Figure 1 | Holistic Performance Overview. Left: Evaluation on RecIF-Bench and general LLM bencharks. ue ive penedat sk hiey general knowledge. "Best Baseline" denotes the highest performance achieved by existing methods for each specific task. Right: Amazon Benchmark results. Our model demonstrates exceptional cross-domain transferability, consistently surpassing leading baselines across 10 diverse datasets.

The left chart of Figure 1 visually confirms the SOTA performance of OneRec-Pro-8B across RecIF-Bench tasks, often significantly outperforming the "Best Baseline." The general LLM benchmarks section on the left also illustrates the trade-off, where OneRec-Pro-8B retains strong performance in areas like MATH-500 but shows some decrease compared to the base Qwen3-8B in broader General Tasks. The right chart (Amazon Benchmark) clearly shows OneRec-Pro-8B consistently outperforming "Best Baseline" in Recall@10 across all 10 Amazon datasets, visually reinforcing its cross-domain transferability.

7. Conclusion & Reflections

7.1. Conclusion Summary

In this work, the authors successfully presented OpenOneRec, a comprehensive framework aimed at bridging the critical gap between traditional recommendation systems and Large Language Models (LLMs). Their contributions are multi-faceted:

  1. RecIF-Bench: They introduced RecIF-Bench, the first holistic benchmark specifically designed to evaluate recommendation instruction-following capabilities. This benchmark covers 8 diverse tasks, ranging from fundamental prediction to complex reasoning, and incorporates crucial dimensions like multi-modal, cross-domain, multi-behavioral interactions, interleaved data, and recommendation explanation.

  2. Open-Sourced Framework & Scaling Laws: To ensure reproducibility and foster scalable research, they open-sourced a full-stack training pipeline, encompassing data processing, co-pretraining strategies (including Itemic-Text Alignment), and hybrid post-training protocols (Multi-task SFT, On-policy Distillation, and Rec-RL). Through empirical analysis, they validated scaling laws in the recommendation domain, demonstrating a data-intensive scaling regime where increasing training data is more critical than merely scaling model parameters.

  3. OneRec-Foundation Models: Leveraging their framework, they released the OneRec-Foundation model family (1.7B and 8B parameters). Extensive experiments demonstrated that these models achieve new state-of-the-art (SOTA) performance across all RecIF-Bench tasks. Furthermore, they exhibited exceptional cross-domain transferability to the Amazon benchmark, surpassing strong baselines with an average 26.8% improvement in Recall@10.

    This work signifies a crucial step towards building truly intelligent recommender systems that can understand, reason, and adapt to diverse user instructions, much like general LLMs.

7.2. Limitations & Future Work

The authors acknowledge several limitations and outline important future research directions:

  • Tokenizer Transferability: While OpenOneRec enhances downstream performance, the magnitude of gains is currently limited by tokenizer transferability. The pre-trained tokenizer, optimized on broad open-domain data, may not optimally distinguish items in a specific vertical, leading to collision rates.
    • Future Direction: Maximizing the reuse of foundation model priors while ensuring high-quality item indexing (code quality) for downstream tasks is a promising avenue. This could involve more adaptive or domain-specific itemic tokenization strategies during transfer.
  • Balancing General Intelligence and Domain Precision: Maintaining the model's general intelligence and reasoning capabilities necessitates mixing vast amounts of general-domain text during training. This process currently presents a trade-off.
    • Future Direction: Investigating optimal data mixing ratios and improving data utilization efficiency are urgent challenges to better balance domain-specific precision with general capabilities, possibly through more advanced curriculum learning or data weighting techniques.
  • Chain-of-Thought Reasoning Consistency: Chain-of-Thought (CoT) reasoning currently yields improvements only in limited settings.
    • Future Direction: A more rigorous exploration of test-time scaling strategies is needed to unlock consistent reasoning gains across diverse recommendation scenarios. This could involve exploring novel prompting techniques, self-consistency methods, or integrating external knowledge graphs more effectively.

7.3. Personal Insights & Critique

The OpenOneRec paper makes a highly significant contribution by systematically addressing the integration of LLMs into recommendation systems. The authors' commitment to open-sourcing their benchmark and training pipeline is commendable and truly accelerates research in this emerging field.

Personal Insights:

  • Paradigm Shift Validation: This paper strongly validates the paradigm shift towards generative recommendation and the potential of foundation models to bring general intelligence to this domain. The improvements on RecIF-Bench and the Amazon benchmark are substantial, particularly the few-shot learning results, which highlight the power of transferable representations under data scarcity. This suggests a future where recommender systems are not just prediction engines but intelligent conversational agents.
  • Bridging Modalities: The itemic tokenization strategy is a crucial innovation. Effectively bridging the discrete item space with the LLM's linguistic token space is fundamental to this integration. The detailed exploration of adaptive strategies for tokenizer transferability (e.g., Text-Augmented Itemic Tokens) is particularly insightful, providing concrete guidance for future work.
  • Balanced Training Regimen: The meticulously designed hybrid post-training pipeline (SFT, On-policy Distillation, Rec-RL) is a key strength. It demonstrates a sophisticated understanding of how to mitigate catastrophic forgetting and simultaneously optimize for diverse objectives (general intelligence vs. domain-specific metrics). This multi-stage approach is likely to become a standard for adapting LLMs to specialized domains.
  • Scaling Laws in a New Domain: The empirical validation of scaling laws specifically for recommendation data is a valuable theoretical contribution. The finding of a data-intensive scaling regime (where data scaling is more impactful than parameter scaling) provides critical guidance for resource allocation in developing future generative recommenders. This challenges the direct applicability of general LLM scaling laws and underscores the unique characteristics of recommendation data.

Critique & Areas for Improvement:

  • General Intelligence Trade-off: While the paper successfully mitigates catastrophic forgetting, the persistent performance gap on some general intelligence benchmarks (e.g., MMLU-Pro, LiveCodeBench) compared to the base Qwen3 suggests that the current data mixing and distillation strategies could be further refined. Future work could explore more dynamic or adaptive data sampling techniques during pre-training and post-training to better balance the two types of knowledge.
  • Complexity of LLM-as-Judge: While LLM-as-Judge is a good approach for text generation, its multi-step process (WIP extraction, semantic matching, weighted F1) is complex and potentially sensitive to the choice of the judge LLM itself. More transparency on its robustness and potential biases would be beneficial.
  • Computational Cost: The scale of training (hundred-billion-token corpus, 8B parameter models, multi-stage training) is substantial. While scaling laws are explored, practical deployment considerations (inference cost, latency) for such large generative recommenders in real-world industrial settings remain a significant challenge. The paper could delve more into the efficiency of these models at inference time.
  • Interpretability of Itemic Tokens: While itemic tokens are explained as maintaining hierarchical semantics, their direct interpretability for humans or for debugging purposes might be limited compared to natural language descriptions. Exploring ways to make itemic tokens more human-understandable could enhance trust and utility.

Transferability to Other Domains: The methodology presented in OpenOneRec has high transferability to other specialized domains where LLMs are being adapted. The core principles—bridging modality gaps (e.g., sensor data, medical images, financial time series to LLM-understandable tokens), scalable co-pretraining with general-domain knowledge, and hybrid post-training with on-policy distillation for general skills and RL for domain-specific metrics—could be applied to:

  • Medical AI: Integrating LLMs with structured medical data (e.g., patient records, lab results, imaging reports) for diagnosis, treatment planning, or drug discovery.

  • Financial Services: Combining LLMs with market data, financial reports, and expert knowledge for quantitative trading, risk management, or personalized financial advice.

  • Robotics/Control: Using LLMs for high-level planning and instruction in robotic systems, where sensor data and control signals need to be aligned with natural language commands.

    By open-sourcing their framework, OpenOneRec provides a solid foundation, inviting the broader research community to tackle these remaining challenges and further accelerate the evolution towards truly intelligent AI systems across various applications.

Similar papers

Recommended via semantic vector search.

No similar papers found yet.