Papers
Topics
Authors
Recent
Search
2000 character limit reached

LIBERO-Para: A Diagnostic Benchmark and Metrics for Paraphrase Robustness in VLA Models

Published 30 Mar 2026 in cs.LG | (2603.28301v1)

Abstract: Vision-Language-Action (VLA) models achieve strong performance in robotic manipulation by leveraging pre-trained vision-language backbones. However, in downstream robotic settings, they are typically fine-tuned with limited data, leading to overfitting to specific instruction formulations and leaving robustness to paraphrased instructions underexplored. To study this gap, we introduce LIBERO-Para, a controlled benchmark that independently varies action expressions and object references for fine-grained analysis of linguistic generalization. Across seven VLA configurations (0.6B-7.5B), we observe consistent performance degradation of 22-52 pp under paraphrasing. This degradation is primarily driven by object-level lexical variation: even simple synonym substitutions cause large drops, indicating reliance on surface-level matching rather than semantic grounding. Moreover, 80-96% of failures arise from planning-level trajectory divergence rather than execution errors, showing that paraphrasing disrupts task identification. Binary success rate treats all paraphrases equally, obscuring whether models perform consistently across difficulty levels or rely on easier cases. To address this, we propose PRIDE, a metric that quantifies paraphrase difficulty using semantic and syntactic factors. Our benchmark and corresponding code are available at: https://github.com/cau-hai-lab/LIBERO-Para

Summary

  • The paper introduces LIBERO-Para, a controlled benchmark that diagnoses VLA modelsโ€™ vulnerability to paraphrased instructions using 43 distinct paraphrase types.
  • It proposes PRIDE, a novel metric quantifying semantic and syntactic divergence, revealing uniform performance drops of 22โ€“52 percentage points under linguistic variations.
  • The findings highlight that object lexical substitutions critically impair VLA models, underscoring the need for enhanced semantic grounding in instruction-to-task mapping.

LIBERO-Para: A Controlled Benchmark for Diagnosing Paraphrase Robustness in Vision-Language-Action Models

Motivation and Problem Formulation

Vision-Language-Action (VLA) models leverage foundation VLMs and language encoders for robotic manipulation, attaining high task success rates via multimodal instruction following. However, in downstream deployment, data-scarce fine-tuning on limited demonstrations often induces overfitting to specific linguistic formulations, giving rise to a critical vulnerability: failure to generalize to unseen paraphrased instructions in the operational environment. The LIBERO benchmark evaluates VLA models under a fixed instruction set, focusing on visual generalization but neglecting robustness to linguistic variation and paraphrasing. Existing benchmarks (CALVIN, LADEV, LIBERO-PRO/Plus/X, LangGap) insufficiently isolate meaning-preserving linguistic perturbations, confound task semantics, or lack formal difficulty quantification. Figure 1

Figure 1: VLA models' susceptibility to paraphrased instructions after data-scarce fine-tuning, emphasizing the paraphrase robustness gap.

LIBERO-Para: Benchmark Design and Paraphrase Taxonomy

LIBERO-Para proposes a controlled two-axis variation schema (action ร— object) where action expressions and object references, the core semantic components in robotic instructions, are systematically paraphrased. The benchmark isolates meaning-preserving linguistic changes, grounded in the Extended Paraphrase Typology and Directive taxonomies, resulting in 43 distinct paraphrase types. The dataset comprises 4,092 paraphrased instructions generated via LLMs and verified for semantic fidelity. Figure 2

Figure 2: LIBERO-Para's two-axis schema enables interpretable, compositional analysis of paraphrase robustness under limited fine-tuning.

Axis-specific examples clarify the scope of both object and action paraphrasing: object axis covers same-polarity substitutions and additions; action axis includes lexical, structural, and pragmatic reformulations, graded by directness and linguistic complexity. Figure 3

Figure 3: Representative examples of paraphrasing across object and action axes, reflecting lexical, syntactic, and pragmatic variation.

PRIDE: Task-Grounded Paraphrase Distance and Robustness Metric

LIBERO-Para introduces PRIDEโ€”a metric quantifying paraphrase difficulty and robustnessโ€”composed of two complementary components:

  1. Keyword Similarity (SKS_K): Evaluates semantic preservation of task-critical content words (action/object) between original and paraphrased instructions via SBERT embeddings and maximal cosine matching.
  2. Structural Similarity (STS_T): Measures syntactic divergence using normalized dependency-tree edit distance on POS and dependency labels, decoupling surface word changes from structural rewritings.

The overall Paraphrase Distance (PD) is a weighted combination of SKS_K and STS_T, calibrated by parameter ฮฑ\alpha. PRIDE score further incorporates task success, providing a granular robustness measure as models face linguistic deviation. Figure 4

Figure 4: Visualization of SKS_K and STS_T calculations, respectively capturing semantic overlap and syntactic reorganization between instructions.

Strong Empirical Findings

Evaluation across seven VLA models (0.6Bโ€“7.5B parameters, four architectural paradigms) reveals pervasive paraphrase fragility:

  • Uniform Performance Degradation: Success rate drops consistently by 22โ€“52 pp under LIBERO-Para, regardless of architecture, scale, or fine-tuning schema.
  • Object Paraphrasing is the Primary Bottleneck: Lexical substitutions of target object references cause the largest declines, with surface-level matching outweighing semantic grounding.
  • Hidden Non-uniformity Unveiled by PRIDE: Conventional success rate falsely attributes robustness to models that only succeed on โ€œeasyโ€ paraphrases; PRIDE reveals systematic failures on difficult variants, overestimating robustness by up to 22%.
  • Planning-level Failures Dominate: 80โ€“96% of errors result from trajectory divergence (task misidentification) rather than low-level execution, highlighting impaired instruction-to-task mapping. Figure 5

    Figure 5: PRIDE score grid reflects increased paraphrase difficulty in more indirect action types and object substitutions, identifying challenging regions.

    Figure 6

    Figure 6: Model-average success rates sharply decrease under object paraphrasing, especially combined with indirect action types.

Detailed Analysis

Cross-model comparison confirms that paraphrase fragility cannot be attributed to architectural choices, training data scope, or VLM adaptation strategy. Increasing task diversity or freezing the VLM backbone does not mitigate the linguistic robustness gap.

Object grounding emerges as the fundamental challenge: synonym or contextually appropriate substitutions for objects (e.g., โ€œstoveโ€ โ†’\rightarrow โ€œrangeโ€) consistently incur pronounced failures. This is further exacerbated by canonical instruction datasets, which offer only one lexical reference per object during training.

Action indirectness and pragmatic paraphrases cause stepwise declines in performance but are less severe than object lexical variation due to limited ambiguity in motor primitives. However, questions or hints that require inference (indirect speech acts) can be substantially disruptive.

Failure analysis using Dynamic Time Warping (DTW) on end-effector trajectories validates that the vast majority of errors are โ€œFar-GTโ€โ€”the model plans toward incorrect targets under paraphrased instructions. Only a minimal proportion (<5% for most models) are โ€œNear-GT,โ€ indicating correct task planning but physical execution errors.

Implications and Future Directions

LIBERO-Para and PRIDE diagnose a critical and quantifiable limitation in current VLA systems: shallow linguistic grounding precludes robust task identification for paraphrased instructions, severely impacting operational reliability. The findings indicate that enhancing paraphrase robustness requires:

  • Prioritizing instruction-to-task semantic mapping rather than low-level motor control;
  • Explicitly modeling object reference diversity and training on lexical alternatives;
  • Developing architectures and training procedures that emphasize semantic object grounding and compositional generalization;
  • Integrating PRIDE or equivalent metrics in evaluation to accurately reflect robustness across linguistic difficulty gradients.

The benchmark facilitates future research into LLM-based paraphrase augmentation, adaptive object grounding, and multimodal instruction mappings that can bridge the robustness gap. Additionally, validating these failures in real-world settings and extending linguistic variation to compound paraphrases remains an open avenue.

Conclusion

LIBERO-Para establishes a rigorous diagnostic framework for evaluating VLA modelsโ€™ robustness to paraphrased instructions. The work identifies object-level lexical variation as the dominant source of failures, quantifies performance degradation and its hidden non-uniformity, and demonstrates that most errors stem from planning rather than execution. PRIDE offers a task-grounded, interpretable metric, revealing that current VLA models lack generalization to unseen linguistic formulations and motivating advancements in semantic grounding and instruction-language diversity (2603.28301).

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 23 likes about this paper.