Shortcut Learning in Generalist Robot Policies: The Role of Dataset Diversity and Fragmentation
TL;DR Summary
This paper attributes generalist robot policy's poor generalization to shortcut learning, caused by limited intra-sub-dataset diversity and inter-sub-dataset fragmentation. Via theoretical and empirical analysis, it shows data augmentation effectively mitigates these dataset issu
Abstract
Generalist robot policies trained on large-scale datasets such as Open X-Embodiment (OXE) demonstrate strong performance across a wide range of tasks. However, they often struggle to generalize beyond the distribution of their training data. In this paper, we investigate the underlying cause of this limited generalization capability. We identify shortcut learning -- the reliance on task-irrelevant features -- as a key impediment to generalization. Through comprehensive theoretical and empirical analysis, we uncover two primary contributors to shortcut learning: (1) limited diversity within individual sub-datasets, and (2) significant distributional disparities across sub-datasets, leading to dataset fragmentation. These issues arise from the inherent structure of large-scale datasets like OXE, which are typically composed of multiple sub-datasets collected independently across varied environments and embodiments. Our findings provide critical insights into dataset collection strategies that can reduce shortcut learning and enhance the generalization ability of generalist robot policies. Moreover, in scenarios where acquiring new large-scale data is impractical, we demonstrate that carefully selected robotic data augmentation strategies can effectively reduce shortcut learning in existing offline datasets, thereby improving generalization capabilities of generalist robot policies, e.g., , in both simulation and real-world environments. More information at https://lucky-light-sun.github.io/proj/shortcut-learning-in-grps/.
Mind Map
In-depth Reading
English Analysis
1. Bibliographic Information
- Title: Shortcut Learning in Generalist Robot Policies: The Role of Dataset Diversity and Fragmentation
- Authors: Youguang Xing, Xu Luo, Junlin Xie, Lianli Gao, Hengtao Shen, Jingkuan Song.
- Affiliations: The authors are affiliated with the University of Electronic Science and Technology of China (UESTC) and Tongji University.
- Journal/Conference: The paper is available on arXiv, which is a preprint server. This means it has not yet undergone formal peer review for a specific conference or journal at the time of this version's publication.
- Publication Year: The preprint was submitted in 2025 (as per the paper's bibliography, though the arXiv ID suggests a different year, this is likely a placeholder). The specific version analyzed is
2508.06426v1. - Abstract: The paper investigates why generalist robot policies, despite being trained on large-scale datasets like Open X-Embodiment (OXE), exhibit poor generalization. The authors identify shortcut learning—relying on task-irrelevant features—as the primary cause. They argue that this issue stems from two structural properties of large datasets: (1) limited diversity within individual sub-datasets and (2) significant distributional gaps between these sub-datasets, which they term dataset fragmentation. Through theoretical and empirical analysis, the paper provides guidelines for future data collection to mitigate these problems. For existing datasets, it demonstrates that robotic data augmentation techniques can effectively reduce shortcut learning and improve the generalization of policies like π₀ in both simulation and the real world.
- Original Source Link:
- Official Page: https://lucky-light-sun.github.io/proj/shortcut-learning-in-grps/
- arXiv Link: http://arxiv.org/abs/2508.06426
- PDF Link: http://arxiv.org/pdf/2508.06426v1
- Publication Status: Preprint.
2. Executive Summary
-
Background & Motivation (Why):
- Core Problem: The field of robotics is attempting to replicate the success of large-scale models in vision and language by training "generalist robot policies" on massive datasets. However, unlike their counterparts in other domains, these policies show surprisingly poor generalization to new tasks and environments, even when trained on datasets containing over a million episodes.
- Importance & Gap: This limitation suggests that simply scaling up data is not a sufficient solution. The paper addresses a critical gap: understanding why these large-scale robot policies fail to generalize. Prior work has noted the generalization failure, but the underlying structural causes within the datasets themselves have not been deeply analyzed.
- Innovation: The paper introduces a fresh perspective by diagnosing the problem as shortcut learning, where policies learn spurious correlations (e.g., associating a specific background with a specific task) instead of the true causal relationships. It provides a formal framework to show that this behavior is a direct result of two intertwined dataset flaws: low diversity within sub-datasets and high fragmentation between them.
-
Main Contributions / Findings (What):
- Problem Diagnosis: The paper is the first to systematically identify and analyze shortcut learning as a key impediment to generalization in generalist robot policies trained on large, composite datasets like OXE.
- Theoretical Framework: It develops a mathematical framework using mutual information to formally prove that low intra-dataset diversity and high inter-dataset fragmentation (disparity) lead to spurious correlations, which are the root cause of shortcut learning.
- Empirical Validation: It provides comprehensive empirical evidence through controlled experiments in simulation (LIBERO, SIMPLER) and the real world. These experiments show that manipulating dataset diversity and disparity directly impacts a policy's reliance on shortcuts and its out-of-distribution (OOD) performance.
- Actionable Guidelines: Based on its findings, the paper proposes three concrete guidelines for future robot data collection to build more robust datasets.
- Practical Mitigation Strategy: For situations where new data collection is infeasible, it demonstrates that targeted data augmentation (viewpoint and object augmentation) can effectively "bridge the gaps" in existing offline datasets, reducing shortcut learning and improving policy generalization.
3. Prerequisite Knowledge & Related Work
-
Foundational Concepts:
- Generalist Robot Policy: A single machine learning model (often a neural network) trained to perform a wide variety of robotics tasks across different environments and embodiments (robot types). These policies typically map sensory inputs (like images and language instructions) directly to robot actions. A common architecture is the Vision-Language-Action (VLA) model.
- Open X-Embodiment (OXE): A massive, open-source dataset for robot learning, created by combining dozens of smaller, independently collected "sub-datasets." Its scale is intended to enable the training of generalist policies.
- Shortcut Learning: A phenomenon where a machine learning model learns to solve a task by relying on simple, unintended, or spurious correlations in the training data rather than the intended, causal features. For example, a model trained to classify cows might learn to associate them with grassy fields, and fail to recognize a cow on a beach. In robotics, this could mean associating a specific camera angle with a specific action, while ignoring the language command.
- Dataset Diversity: The degree of variation within a dataset or sub-dataset. In this context, it refers to the variety of both task-relevant factors (e.g., different objects, tasks, instructions) and task-irrelevant factors (e.g., backgrounds, viewpoints, lighting).
- Dataset Fragmentation: A structural property of a dataset composed of multiple sub-datasets where there are large distributional gaps and minimal overlap between them. This makes the sub-datasets look like isolated islands of data rather than a single, cohesive whole.
-
Previous Works & Technological Evolution:
- The paper situates itself within the trend of scaling laws, where performance in fields like NLP and computer vision has been shown to improve predictably with increases in model size and data volume. The robotics community has tried to follow this path by creating large datasets (
OXE,BridgeData V2,DROID) and training high-capacity models (RT-1,Octo, π₀). - However, recent work [18] has shown that this scaling approach has not yielded the same generalization benefits in robotics. Policies trained on OXE still struggle with minor changes in viewpoint, language, or object poses.
- The concept of shortcut learning is well-established in computer vision [57, 67] and NLP [71, 73], but this paper is the first to apply it as a primary lens to diagnose the failures of large-scale robot policies.
- While other works have focused on data quality [88] or optimizing data mixtures [89], they have typically analyzed properties within individual trajectories or at the dataset-composition level.
- The paper situates itself within the trend of scaling laws, where performance in fields like NLP and computer vision has been shown to improve predictably with increases in model size and data volume. The robotics community has tried to follow this path by creating large datasets (
-
Differentiation:
- This paper's key innovation is its focus on the structural properties of the dataset as a whole—specifically the interplay between low intra-dataset diversity and high inter-dataset fragmentation—as the fundamental cause of shortcut learning.
- Unlike work on data curation that scores individual trajectories [86], this paper uses mutual information as a diagnostic tool at the dataset-structure level to explain why spurious correlations emerge in the first place.
- It provides not only a diagnosis but also a clear, theoretically-grounded set of solutions: principles for better data collection and practical data augmentation techniques.
4. Methodology (Core Technology & Implementation)
The paper's methodology is divided into three parts: analyzing existing datasets, building a theoretical model to explain the observations, and validating the theory through experiments.
4.1 Analysis of Dataset Diversity and Fragmentation
To quantify the structural properties of datasets like OXE, the authors propose two metrics based on feature similarity.
-
Feature Extraction:
- Visual Features: They use a concatenation of features from pretrained
DINOv2andSigLIPmodels, which capture complementary visual information. They analyze only the first frame of each episode, assuming subsequent frames are similar. - Language Features: They use
CLIPto extract features from language instructions.
- Visual Features: They use a concatenation of features from pretrained
-
Diversity Metric (
S_diversity):- This metric quantifies the diversity within a single sub-dataset . It is adapted from the uniformity metric in contrastive learning.
- Formula:
- Symbol Explanation:
u, v: Feature vectors of two data points sampled from the sub-dataset .- : The squared Euclidean distance between the two feature vectors.
- : A temperature parameter. A larger makes the metric more sensitive to small distances, effectively focusing on very similar pairs.
- Intuition: The metric is the inverse of the average similarity between all pairs of points in the sub-dataset. If points are spread out uniformly (high diversity), their average similarity is low, and thus is high.
-
Disparity Metric (
S_disparity):- This metric quantifies the dissimilarity or fragmentation between multiple sub-datasets.
- Formula:
- Symbol Explanation:
- : The number of sub-datasets.
- : Feature vectors sampled from two different sub-datasets, and .
- Intuition: This metric is inversely proportional to the average similarity between points from different sub-datasets. If sub-datasets are well-separated (high fragmentation), the similarity between them is low, and thus is high.
4.2 Theoretical Framework for Shortcut Learning
The authors formalize shortcut learning to show how dataset structure causes it.
-
Principles:
- An observation is composed of task-relevant factors (which causally affect the correct action ) and task-irrelevant factors (e.g., background, viewpoint).
- Ideally, and should be independent in the training data ().
- Shortcut learning occurs when and are correlated in the training data. This creates a spurious correlation between the irrelevant factor and the action , which the model exploits.
- The paper models a large dataset like OXE as a mixture of sub-datasets .
-
Key Propositions (for sub-datasets): The degree of shortcut learning is quantified using normalized mutual information , which measures the statistical dependency between relevant () and irrelevant () factors.
-
Proposition 3.1 (Disjoint Supports): When the factors in the two sub-datasets are completely separate (e.g., Viewpoint Set A is only in Sub-dataset 1, Viewpoint Set B is only in Sub-dataset 2), the mutual information is:
- Symbol Explanation:
- is the total diversity (measured by Shannon entropy ) within the sub-datasets.
- Implication: Spurious correlation () is inversely proportional to diversity. Lower diversity leads to stronger shortcuts. This is the "limited diversity" problem.
- Symbol Explanation:
-
Proposition 3.2 (Overlapping Supports): When the sub-datasets have some overlap in their factors (e.g., both datasets contain some of the same viewpoints), the mutual information is bounded:
- Symbol Explanation:
- measures the degree of overlap between the factor distributions of the sub-datasets. A higher value means more overlap.
- Implication: As the overlap () increases, the upper bound on the spurious correlation decreases. This means interleaving or "bridging" sub-datasets weakens shortcuts. This addresses the "fragmentation" problem.
- Symbol Explanation:
-
5. Experimental Setup
-
Datasets:
- Open X-Embodiment (OXE): Used for the initial analysis of diversity and fragmentation. Specifically, the
OXE Magic Soup++subset of 27 high-quality sub-datasets was analyzed. - LIBERO: A simulation benchmark used for controlled experiments. The
LIBERO-Spatialtask suite was used, where the task-relevant factor was the object's initial position (and corresponding instruction), and the task-irrelevant factor was the camera viewpoint. - SIMPLER: A realistic simulation environment used to evaluate policies fine-tuned on offline data, mimicking the Bridge and RT-1 datasets from OXE.
- Real-World: A physical setup with an AgileX PIPER robotic arm and two cameras to validate findings in a real environment.
- Open X-Embodiment (OXE): Used for the initial analysis of diversity and fragmentation. Specifically, the
-
Evaluation Metrics:
- OOD Success Rate:
- Conceptual Definition: Measures the policy's ability to generalize by calculating the percentage of successfully completed tasks in out-of-distribution (OOD) settings. An OOD setting is one where the spurious correlations learned during training are broken (e.g., a viewpoint from sub-dataset A is paired with a task from sub-dataset B).
- Mathematical Formula:
- Symbol Explanation: A trial is a single attempt by the policy to complete a task. Success is determined by task-specific criteria (e.g., object placed correctly).
- Degree of Shortcut Learning:
- Conceptual Definition: A human-assessed score that quantifies how strongly a policy relies on a shortcut cue instead of the correct instruction. This provides a more nuanced measure than binary success/failure, capturing cases where the robot tries to perform the wrong task. Lower is better.
- Scoring Rubric (from Appendix G):
- 1.0 (Clear Shortcut): The policy ignores the instruction and performs the action associated with the spurious cue (e.g., viewpoint).
- 0.5 (Ambiguous Shortcut): The behavior is unclear, hesitant, or a mix between the correct and shortcut action.
- 0.0 (No Shortcut): The policy correctly attempts to follow the instruction, regardless of final success.
- The final score is the average over all evaluation trials.
- OOD Success Rate:
-
Baselines (Models Tested):
- Diffusion Policy: A vision-only imitation learning model that uses a diffusion process to generate actions. Used to test if shortcuts exist even without language.
- MiniVLA: A smaller-scale Vision-Language-Action (VLA) model, trained from scratch.
- π₀ (Pi-Zero): A powerful, large-scale VLA model pretrained on extensive robot datasets. It was fine-tuned with LoRA to test if even strong pretrained models are susceptible to shortcut learning on new, biased data.
6. Results & Analysis
6.1 Core Results: Dataset Analysis
- Low Diversity and High Fragmentation in OXE:
-
Figure 2 (Image 8): The analysis shows that both the visual and textual diversity of individual OXE sub-datasets are orders of magnitude lower than standard vision and multimodal datasets (e.g., ImageNet, LAION).
该图像为示意图,展示了两个子数据集在不同视角和指令下的机器人执行场景对比。上方为原始数据集样本,左侧视角A对应指令C,右侧视角B对应指令D;下方为新增样本,分别在相同视角下加入新指令“将纸放入篮子”。画面用绿色圈标注了关注的对象位置,展示了数据集多样性和指令变化对机器人任务表现的影响。 -
Figure 3 (Image 9): A t-SNE visualization of visual features reveals that vision/multimodal datasets are intertwined, while OXE sub-datasets form distinct, separate clusters. This is a clear visual depiction of dataset fragmentation.
该图像为插图,展示了机器人操作场景中的数据增强效果。左侧为原始图像,右侧为经过数据增强后的图像,显示了同一场景中机器人夹取物体的位置和视角略有变化,用以提升训练数据的多样性。 -
Figure 4 (Image 10): The quantitative analysis confirms the visual intuition. The disparity metric and the ratio are significantly higher for OXE, especially at higher temperatures (which focus on closer data points), indicating that sub-datasets are like isolated "points" with large gaps between them.
该图像为示意图,展示了两个子数据集(使用不同摄像头和物品组合)中,原始场景与增强场景的对比。子数据集1使用左摄像头和香蕉、桌布等物品,子数据集2使用右摄像头和西瓜等物品。图中显示通过数据增强,使两个子数据集的视觉元素更加多样化,有助于减少数据集碎片化和提升机器人策略的泛化能力。 -
Figure 15 (Image 6): The paper shows that task diversity (number of unique language instructions) is also extremely low in most sub-datasets, with many containing fewer than 10 unique tasks. This lack of diversity in task-relevant features further contributes to the problem.
该图像为多子图的图表,展示了不同模型(DP、MiniVLA、π₀等)在视角多样性、视角差异性以及位置(语言)多样性和差异性条件下,捷径学习程度与离群成功率之间的关系。图中上排显示捷径学习程度随视角半径、视角中心距离和物体位置数量变化的趋势,下排对应离群成功率的变化,反映数据多样性和碎片化对模型泛化能力的影响。
-
6.2 Core Results: Controlled Experiments on LIBERO
-
Impact of Diversity and Disparity: The experiments systematically validate the theoretical propositions.
- Figure 6: The plots show a clear trend across all three models (
Diffusion Policy,MiniVLA, π₀).-
Increasing Diversity: As viewpoint diversity (an irrelevant factor) or object position diversity (a relevant factor) within sub-datasets increases, the degree of shortcut learning drops, and the OOD success rate rises.
-
Decreasing Disparity: As the distance (disparity) between the viewpoint distributions of the two sub-datasets decreases, shortcut learning is reduced, and OOD performance improves.
-
Key Insight: These results directly support Propositions 3.1 and 3.2, showing that both low diversity and high fragmentation cause shortcut learning and harm generalization.
该图像为多子图的图表,展示了不同模型(DP、MiniVLA、π₀等)在视角多样性、视角差异性以及位置(语言)多样性和差异性条件下,捷径学习程度与离群成功率之间的关系。图中上排显示捷径学习程度随视角半径、视角中心距离和物体位置数量变化的趋势,下排对应离群成功率的变化,反映数据多样性和碎片化对模型泛化能力的影响。
-
- Figure 6: The plots show a clear trend across all three models (
-
Diversity Does Not Always Help:
-
Figure 7: This experiment serves as a crucial control. When diversity is increased in a way that creates a new spurious correlation (by assigning a unique viewpoint to each task), it actually worsens fragmentation. This leads to a complete drop in OOD performance, demonstrating that diversity must be introduced carefully to maintain factor independence.
该图像为示意图和柱状图组合。左侧示意图展示通过增加训练数据的视角数量(从2个视角到10个视角)覆盖更多任务的过程;右侧两柱状图显示在MiniVLA环境中,使用10个视角训练相比2个视角训练,捷径学习程度降低(第一张柱状图中蓝色柱明显高于棕色柱),而OOD(分布外)成功率提升(第二张柱状图中棕色柱更高且带误差线)。整体说明通过增加视角多样性,有助于减轻捷径学习,提高泛化能力。
-
6.3 Core Results: Real-World and Augmentation Experiments
-
Validating the Theory in the Real World:
-
Figure 1 (Image 1): The initial real-world experiment shows π₀ fine-tuned on two confounded sub-datasets (Viewpoint A with Instruction C, Viewpoint B with Instruction D) fails the OOD test (Viewpoint A with Instruction D), confirming its susceptibility to shortcut learning.
该图像为实验示意图,展示了基于OXE训练的模型在SIMPLER OOD测试中的表现及微调过程。左侧通过三组“放勺子到毛巾”任务的起始和结束环境图,指出模型表现出“拿起可乐罐”的捷径行为(shortcut behavior)。右侧展示了通过两个子数据集不同视角和指令微调策略π₀,最后在OOD测试中绿色圆圈处机器人成功完成指令D任务,表明微调后能减弱捷径行为。 -
Table 1 (Manual Transcription): This table shows the effect of mitigation strategies. The baseline π₀ model has a high shortcut degree (0.6) and low OOD success (0.2).
-
Adding a third "bridge" object (seen from both viewpoints) increases diversity and reduces disparity, completely eliminating shortcut behavior (degree 0) and dramatically improving OOD success to 75%.
-
Applying viewpoint augmentation also significantly reduces the shortcut degree (0.15) and improves success (0.55).
Model Shortcut degree ↓ OOD success rate ↑ π₀ baseline 0.6 0.2 + third object 0 0.75 + viewpoint aug 0.15 0.55
-
-
-
Mitigation via Object Augmentation:
-
Table 2 (Manual Transcription): This experiment tests object augmentation in both SIMPLER and the real world on a π₀ model. The non-augmented model exhibits extreme shortcut behavior (degree 1.0 in SIMPLER, 0.8 in real-world). In contrast, the model fine-tuned on augmented data, where objects are swapped between scenes, shows a substantially lower degree of shortcut learning (0.68 and 0.25, respectively), indicating it is better at following language instructions regardless of the visual context.
Model Shortcut degree (SIMPLER) Shortcut degree (Real-world) π₀ 1.0 0.8 π₀ + aug 0.68 0.25
-
7. Conclusion & Reflections
-
Conclusion Summary: The paper provides a compelling argument, backed by theory and extensive experiments, that the limited generalization of current generalist robot policies is not just a matter of needing more data, but a fundamental problem with the structure of that data. The core issue is shortcut learning, which is induced by low diversity within sub-datasets and high fragmentation between them. The authors show that simply adding more similarly structured data may even be harmful. The path forward requires a more disciplined approach to data collection that focuses on controlled diversification of relevant factors while ensuring overlap and consistency across sub-datasets. For existing datasets, targeted data augmentation offers a practical way to repair these structural flaws.
-
Limitations & Future Work: The authors acknowledge several limitations:
- Identifying Specific Shortcuts: The analysis confirms the existence of shortcuts but doesn't provide methods to automatically identify the specific spurious correlations (e.g., background color, object texture) that models are exploiting in massive datasets like OXE.
- Measuring Task-Relevant Diversity: The quantitative analysis focused on visual and textual features. Measuring the diversity of task-relevant factors like grasp affordances or object positions remains a major challenge.
- Scalability of Augmentation: While effective in controlled settings, scaling the proposed data augmentation techniques to the entirety of OXE would be computationally expensive and may introduce visual artifacts.
- Real-World Complexity: The real-world experiments, while valuable, were limited in scale and complexity compared to the variety of scenarios in large datasets.
- Model-Centric Solutions: The paper focuses on data-centric solutions. It does not explore how model architectures or training objectives could be designed to be inherently more robust to shortcut learning.
-
Personal Insights & Critique:
- This paper is a significant contribution to the robot learning community. It provides a clear and intuitive framework for understanding a critical bottleneck that has puzzled many researchers: why scaling isn't working as expected.
- The connection drawn between dataset structure and shortcut learning is powerful and well-supported. The theoretical propositions provide a solid foundation for the empirical findings.
- The distinction between "good" diversity (that maintains factor independence) and "bad" diversity (that creates new correlations) is a crucial insight for anyone involved in data collection.
- The practical demonstration that data augmentation can "repair" fragmented datasets is highly valuable, as it provides an immediate, actionable strategy for researchers working with existing offline data.
- A potential avenue for future work could be to develop a "dataset health check" tool based on the
S_diversityandS_disparitymetrics, allowing researchers to automatically flag sub-datasets or factor combinations that are likely to induce shortcut learning before starting a lengthy training run. - The paper's conclusion—that we must strategically control some factors while diversifying others—is a mature and pragmatic take on the challenges of embodied intelligence, moving the field beyond a naive "more is better" approach to data.
Similar papers
Recommended via semantic vector search.