LIBERO-Para: A Diagnostic Benchmark and Metrics for Paraphrase Robustness in VLA Models

Chanyoung Kim1*, Minwoo Kim1*, Minseok Kang1, Hyunwoo Kim2, Dahuin Jung2†
1Soongsil University, 2Chung-Ang University
* Equal contribution    † Corresponding author
Paraphrase robustness gap in VLA models

VLA models overfit to seen instruction phrasings during fine-tuning and fail to generalize to paraphrased variants at deployment.

Abstract

Vision–Language–Action (VLA) models achieve strong performance in robotic manipulation by leveraging pre-trained vision–language backbones. However, in downstream robotic settings, they are typically fine-tuned with limited data, leading to overfitting to specific instruction formulations and leaving robustness to paraphrased instructions underexplored. To study this gap, we introduce LIBERO-Para, a controlled benchmark that independently varies action expressions and object references for fine-grained analysis of linguistic generalization. Across seven VLA configurations (0.6B–7.5B), we observe consistent performance degradation of 22–52 pp under paraphrasing. This degradation is primarily driven by object-level lexical variation: even simple synonym substitutions cause large drops, indicating reliance on surface-level matching rather than semantic grounding. Moreover, 80–96% of failures arise from planning-level trajectory divergence rather than execution errors, showing that paraphrasing disrupts task identification. Binary success rate treats all paraphrases equally, obscuring whether models perform consistently across difficulty levels or rely on easier cases. To address this, we propose PRIDE, a metric that quantifies paraphrase difficulty using semantic and syntactic factors.

LIBERO-Para Overview

Compared to LIBERO, LIBERO-Para evaluates paraphrase robustness under data-scarce fine-tuning via a controlled two-axis design (action vs. object), enabling interpretable analysis. The benchmark includes 4,000+ paraphrased instructions across 10 evaluation scenarios.

LIBERO-Para overview

Overview of LIBERO-Para. Paraphrases are decomposed along action and object axes, enabling fine-grained analysis of which linguistic variations most severely impact VLA performance.

Paraphrase Taxonomy

Robotic manipulation instructions are structured around what to act on and how to act. LIBERO-Para decomposes paraphrases along these two axes: object-referring expressions (e.g., synonym substitution, addition) and action-referring expressions (lexical, structural, and pragmatic variations). Composing both yields compositional paraphrases, totaling 43 fine-grained variation types.

Paraphrase taxonomy examples

Examples of axis-specific paraphrases. Object variations modify target references; action variations cover lexical, structural, and pragmatic realizations.

PRIDE Metric

Binary success rate treats all paraphrases equally, obscuring whether models succeed consistently or rely on easier cases. PRIDE (Paraphrase Robustness Index in Robotic Instructional DEviation) addresses this by computing a Paraphrase Distance (PD) from keyword similarity (SK) and structural similarity (ST), then weighting success by difficulty. Unlike plain SR, PRIDE gives more credit for succeeding on harder, more deviated paraphrases.

PRIDE metric: SK and ST computation

SK (top) measures keyword-level semantic similarity between task-critical content words. ST (bottom) uses dependency-tree edit distance to capture structural variation.

Experiment

Success Rate: LIBERO-Goal vs. LIBERO-Para

Method LIBERO-Goal SR LIBERO-Para SR Drop
Xiaomi-Robotics-098.876.0-22.8
π0.597.671.4-26.2
OpenVLA-OFTmixed96.163.7-32.4
OpenVLA-OFTgoal97.964.7-33.2
X-VLA97.862.1-35.7
π0.5 (expert-only)78.639.1-39.5
VLA-Adapter98.246.3-51.9

PRIDE Reveals Hidden Severity

Method SR PRIDE Overestimation (%)
VLA-Adapter46.336.122.0
π0.5 (expert-only)39.132.018.2
X-VLA62.152.715.1
OpenVLA-OFTmixed63.756.311.6
OpenVLA-OFTgoal64.758.89.1
Xiaomi-Robotics-076.069.28.9
π0.571.465.48.4

Overestimation = (SR − PRIDE) / SR. Higher values indicate success concentrated on easier paraphrases.

Model-average SR heatmap

Model-average success rate per Object × Action cell. Object-paraphrased rows drop sharply, reaching 30.4% at SP-habitual × Hint.

Finding 1: Paraphrase Fragility Persists

Across seven configurations spanning four architecture families, all models show substantial SR drops under paraphrasing (22.8–51.9 pp). The 7.5B OpenVLA-OFT shows PRIDE scores comparable to the 0.9B X-VLA. Expanding task-level data diversity by 4× (OFTmixed vs. OFTgoal) yields similar drops (32.4 vs. 33.2 pp), suggesting that increasing task diversity does not improve robustness to linguistic variation. Freezing the VLM and fine-tuning only the Action Expert also does not help. Paraphrase fragility cannot be explained by architecture, data scope, or fine-tuning strategy alone.

Finding 2: Object Grounding Is the Primary Bottleneck

When the object is paraphrased—even through common synonyms such as replacing stove with range—performance drops by 19.8–51.0 pp across models. This gap appears consistently across architectures, suggesting that current VLAs rely on surface-level keyword matching rather than semantic understanding. The object space is lexically open-ended, concentrating combinatorial complexity on object references.

Object-preserved vs object-paraphrased SR

SR comparison: object-preserved (None, Addition) vs. object-paraphrased (SP-contextual, SP-habitual). Δ annotated per pair.

Finding 3: Failures Are Planning-Level, Not Execution-Level

We classify failures based on trajectory similarity to successful executions. Far-GT (planning-level): trajectories diverge from the ground-truth early, indicating the model misidentified the task. Near-GT (execution-level): trajectories track the GT but fail due to minor control errors. Across models, 79.5–95.5% of failures are Far-GT, showing that paraphrasing disrupts task identification rather than motor control.

Near-GT vs Far-GT trajectory visualization

Green: success; black: mean GT; orange: Near-GT failure; red: Far-GT failure (diverges early).

Model SR Near-GT Far-GT Far-GT (%)
OpenVLA-OFTgoal64.71.633.795.5
Xiaomi-Robotics-076.01.822.292.5
VLA-Adapter46.34.249.592.2
π0.571.42.426.291.6
OpenVLA-OFTmixed63.73.333.090.9
X-VLA62.15.232.786.3
π0.5 (expert-only)39.112.548.479.5

Implications for Future VLA Development

Our findings collectively point to a fundamental limitation: current VLA models rely on surface-level keyword matching rather than genuine semantic understanding of instructions. This suggests several directions for advancing the next generation of VLA systems:

  • From keyword matching to semantic object grounding. Object-level variation is the dominant failure source (Finding 2), yet current training data typically refers to each object by a single canonical name. Future work should explore richer object representations—grounding objects through visual attributes, spatial relations, or functional descriptions—rather than relying on fixed lexical labels.
  • Instruction-to-task mapping over motor control refinement. 80–96% of failures are planning-level, not execution-level (Finding 3). This means models generate entirely wrong trajectories, not slightly imprecise ones. Improving VLA robustness should prioritize how instructions are mapped to task plans, rather than refining low-level action precision.
  • Linguistic diversity in training. Scaling model size or expanding task-level data diversity does not resolve paraphrase fragility (Finding 1). This suggests that exposure to linguistically diverse instructions during training—including synonyms, structural rephrasing, and indirect expressions—may be essential for robust deployment.
  • Difficulty-aware evaluation. Standard binary success rate masks hidden severity by treating easy and hard paraphrases equally. Metrics like PRIDE that account for paraphrase difficulty provide a more faithful assessment of real-world robustness, and should be adopted alongside SR in future VLA benchmarks.

Ultimately, for VLA models to operate reliably in real-world settings where users express the same intent in diverse ways, the field must move beyond pattern memorization toward robust language understanding grounded in semantic and visual context.

BibTeX

@misc{kim2026liberoparadiagnosticbenchmarkmetrics,
      title={LIBERO-Para: A Diagnostic Benchmark and Metrics for Paraphrase Robustness in VLA Models},
      author={Chanyoung Kim and Minwoo Kim and Minseok Kang and Hyunwoo Kim and Dahuin Jung},
      year={2026},
      eprint={2603.28301},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2603.28301},
}