VLS: Steering Pretrained Robot Policies via Vision-Language Models
Abstract: Why do pretrained diffusion or flow-matching policies fail when the same task is performed near an obstacle, on a shifted support surface, or amid mild clutter? Such failures rarely reflect missing motor skills; instead, they expose a limitation of imitation learning under train-test shifts, where action generation is tightly coupled to training-specific spatial configurations and task specifications. Retraining or fine-tuning to address these failures is costly and conceptually misaligned, as the required behaviors already exist but cannot be selectively adapted at test time. We propose Vision-Language Steering (VLS), a training-free framework for inference-time adaptation of frozen generative robot policies. VLS treats adaptation as an inference-time control problem, steering the sampling process of a pretrained diffusion or flow-matching policy in response to out-of-distribution observation-language inputs without modifying policy parameters. By leveraging vision-LLMs to synthesize trajectory-differentiable reward functions, VLS guides denoising toward action trajectories that satisfy test-time spatial and task requirements. Across simulation and real-world evaluations, VLS consistently outperforms prior steering methods, achieving a 31% improvement on CALVIN and a 13% gain on LIBERO-PRO. Real-world deployment on a Franka robot further demonstrates robust inference-time adaptation under test-time spatial and semantic shifts. Project page: https://vision-language-steering.github.io/webpage/
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
What this paper is about
Robots often learn by watching examples, like “pick up the cup and place it on the table.” They do great when the scene looks exactly like their training videos. But small changes—like the cup is near the edge, there’s a new object nearby, or the instruction is slightly different—can make them fail. This paper introduces a way to help a robot adapt on the fly, without retraining it, by using a vision-LLM (a model that understands pictures and words) to “steer” the robot’s existing skills. The method is called Vision-Language Steering (VLS).
The main goals, in simple terms
- Make a trained robot policy (its “playbook” for actions) work in new, slightly different situations it wasn’t trained for.
- Do this at test time, with no extra training or changing the robot’s policy weights.
- Use a vision-LLM to understand what’s in the scene and what the instruction means, then guide the robot’s action planning to fit the new situation.
- Show that this works better than other existing “steering” methods in both simulations and real-world tests.
How the method works
Think of the robot’s policy like a driver who knows many maneuvers (turn left, brake, park) but often gets thrown off when the road layout changes. VLS acts like a smart GPS that understands the scene and your instructions, then gently nudges the driver to choose the right maneuvers for the current road.
Here are the key steps, explained with everyday ideas:
Step 1: Understand the scene and instruction
- The system uses a vision-LLM (VLM) to identify important objects (like “red cup,” “coaster,” “basket”) from the camera image and the instruction text.
- It makes clean object masks (with tools like SAM) and extracts features (with DINOv2) to find a few important 3D points in the scene—like the cup’s location or the basket’s center. Think of these as “key spots” the robot might need to move to or avoid.
Step 2: Write simple “reward rules” the robot can follow
- The VLM breaks the task into stages, like: 1) Move the gripper near the cup 2) Grasp the cup 3) Place the cup in the basket
- For each stage, the VLM produces a small scoring program (a “reward function”) that gives higher scores to action plans that do the right thing. For example, “stay close to the cup” gets a high score during stage 1.
- Importantly, these scores are differentiable—meaning the robot can tell how to adjust its plan to get a better score, like following the steepest slope up a hill.
Step 3: Steer the robot’s plan during generation
Many modern robot policies generate action sequences using “denoising” methods (diffusion or flow matching). You can picture this as starting with a blurry idea of an action plan and gradually sharpening it.
- Gradient nudges: VLS adds tiny pushes (gradients) that point the plan toward higher-scoring actions. This is like telling the driver, “a bit more to the left,” “slow down,” as the plan takes shape.
- Keep options diverse: At the start, VLS spreads out several candidate plans so they don’t all cluster in one bad choice. This is like trying multiple routes at once to see which looks best.
- Copy the promising ones: As plans improve, VLS periodically keeps and duplicates the best candidates and drops the poor ones (using a method inspired by the Feynman–Kac principle). It’s like keeping the routes with fewer delays and discarding the rest.
Step 4: Close the loop and switch stages smoothly
- Adaptive strength: If the robot is doing well in the current stage, VLS eases off the steering so the base policy can handle fine control. If it’s drifting off, VLS pushes harder.
- Robust stage switching: VLS uses two thresholds (a “hysteresis” trick, like a stable on/off switch) to decide when to move from one stage to the next. This avoids flip-flopping between stages when conditions are borderline.
What they found and why it matters
Here are the key results from tests in simulators and with a real robot arm:
- Stronger performance in simulations:
- CALVIN benchmark: VLS achieved about 94% average success on moving objects and 87% on articulated parts (like doors and drawers), beating other steering methods by 15–25 percentage points and far surpassing the unsteered base policy.
- LIBERO-PRO benchmark: VLS improved success rates by up to 13% on tasks with changed object positions or changed instructions, outperforming well-known vision-language-action baselines (like OpenVLA and variants of TT-0.5).
- Real-world robot success:
- On a Franka robot, VLS boosted average in-distribution success to around 69% and maintained much higher success under changes like new object appearances, swapped positions, or even replacing the target object with something never seen before. In the toughest case (a new, unseen mug), VLS still succeeded about 40% of the time, while the baseline failed.
- Why the parts matter:
- Removing the gradient nudges caused performance to collapse—these dense “pushes” are the main reason VLS works.
- Diversity and resampling make the system more stable and efficient, preventing early commitment to bad plans and helping it find good ones quicker.
- There’s a compute trade-off: more samples can improve success but increase latency.
What this means going forward
- Practical impact: VLS lets you take a powerful robot policy and adapt it to new scenes or instructions on the spot, without retraining. This could make robots more reliable in homes, warehouses, and labs where conditions change all the time.
- Conceptual shift: Instead of relearning skills for every new setup, you keep the skills and control their execution using a scene- and language-aware “steering wheel.”
- Limitations: The method adds some computation time due to multiple samples and resampling. Future work could make the scoring smarter and the steering faster so it runs with even lower delay.
- Big picture: Combining a robot’s learned motor skills with the broad understanding of vision-LLMs is a promising path to more flexible, trustworthy robot behavior in the real world.
Knowledge Gaps
Knowledge gaps, limitations, and open questions
Below is a single, concrete list of missing pieces, uncertainties, and unexplored directions that future researchers could address.
- Reliability of VLM-generated rewards: How accurate and faithful are the VLM-produced differentiable reward functions to task semantics and geometry, especially under ambiguous or complex natural language? Build benchmarks to quantify hallucination rates and misalignment, and develop automated verification or confidence measures for reward code.
- Safety and constraint compliance: The current reward shaping does not explicitly ensure collision avoidance or respect physical constraints near obstacles. Integrate safety shields (e.g., control barrier functions), formal constraint checks, or certified planners into the steering loop and evaluate safety metrics beyond success rate.
- Grounding pipeline robustness: The approach relies on RGB-D, SAM segmentation, DINO features, and clustering to produce 3D keypoints. Systematically evaluate failure modes under noisy depth, occlusions, poor lighting, truncation, and clutter; design robust alternatives (e.g., monocular depth, multi-view fusion, uncertainty-aware clustering) and quantify sensitivity.
- Stage decomposition correctness: VLM-based stage generation and switching may misidentify or reorder sub-tasks. Develop methods to detect incorrect stage proposals, learn or verify stage transitions from execution feedback, and automatically tune Schmitt-trigger thresholds for diverse tasks.
- Handling negative and relational constraints: Tasks like “pick up the cup without coaster” require excluding specific relations. Provide mechanisms to express and enforce negation, exclusion, and multi-object relational constraints in differentiable rewards, and test on compositional instructions with conflicting or conditional requirements.
- Gradient quality and stability: The reward gradient approximates a likelihood without guarantees. Analyze gradient noise, bias, and stability; explore smoothing, trust-region updates, line search, or second-order methods; and characterize conditions under which guided denoising converges to constraint-satisfying trajectories.
- Theoretical guarantees of steering: Provide formal analysis of convergence, sample efficiency, and preservation of the expert manifold under gradient-guided denoising plus FK resampling, including bounds on deviation from the base policy and conditions for improvement versus degradation.
- Hyperparameter sensitivity and auto-tuning: The framework depends on guidance scale, batch size, MCMC inner steps, RBF parameters, and resampling schedules. Quantify sensitivity and develop on-the-fly auto-tuning (e.g., Bayesian optimization or adaptive controllers) that balance performance and latency.
- Computational efficiency and real-time control: Batch sampling, MCMC, and FK resampling introduce latency that may be incompatible with high-frequency control. Investigate model distillation, gradient approximation, event-triggered guidance, hardware acceleration, and asynchronous steering to meet real-time constraints on edge devices.
- Generalization to contact-rich and dexterous manipulation: Current evaluations focus on relatively simple pick-and-place and articulated tasks. Test VLS on contact-rich, force-sensitive, deformable object manipulation, and integrate tactile/force feedback into reward functions.
- Robustness to severe OOD and dynamic scenes: Assess performance under large distribution shifts (novel categories, drastic layouts), moving targets, and non-stationary environments; add online updates to the keypoint scaffold P and rewards to handle time-varying constraints.
- Multi-modal perception beyond vision: Explore incorporating tactile, audio, proprioception, and state estimators into the grounded scaffold and reward generation, and measure gains under occlusion or poor visual conditions.
- Discrete actions and non-differentiable components: Gripper open/close and certain high-level actions are discrete. Develop surrogate differentiable relaxations or hybrid steering that can handle non-differentiable action dimensions without compromising control fidelity.
- Base policy coverage and policy class diversity: The method is demonstrated on diffusion and flow-matching policies. Evaluate transferability to other generative or non-generative controllers (e.g., transformer action regressors, model-based policies) and quantify when steering is effective versus detrimental.
- Comparative evaluation with value/critic-guided methods: The paper argues critic-guidance is “undesirable” but lacks direct empirical comparisons under identical OOD settings. Run controlled studies to validate claims and explore hybrid approaches combining VLM rewards with learned values.
- Keypoint scaffold design choices: Clarify and assess how many keypoints are needed, clustering strategy, feature weighting, and spatiotemporal updates. Provide ablations and guidelines for constructing P that minimize failure under varied object geometries and articulations.
- Stage-switching policy under uncertainty: The Schmitt-trigger parameters (Rhigh, Rlow) are task-specific. Investigate principled ways to set and adapt thresholds across tasks, and study oscillation and deadlock failure modes under noisy rewards.
- Reward code generation security and reliability: VLM-generated PyTorch code may contain bugs or unsafe operations. Establish a sandboxed, type-checked pipeline with unit tests, static analysis, and runtime guards; report compilation/runtime error rates and auto-recovery strategies.
- Evaluation breadth and reproducibility: Current results cover CALVIN, LIBERO-PRO, and one Franka setup. Test across more robots, grippers, sensors, and labs; release code, prompts, and reward libraries; and report metrics beyond success rate (e.g., interventions, collisions, path length).
- Language understanding limits: Assess VLS under long, compositional, referential, ambiguous, or multilingual instructions; devise prompt-robust reward generation, and measure brittleness to linguistic perturbations.
- OOD detection and fallback strategies: Introduce mechanisms to detect when rewards or grounding are unreliable, trigger conservative execution (base policy only), or request human/VLM clarification, and quantify the trade-off between safety and task completion.
- Resampling schedule design: FK resampling frequency and weighting affect mode exploration vs. exploitation. Provide principled schedules, analyze degeneracy and particle impoverishment, and study interactions with gradient guidance.
- Integration with formal task planners: Explore combining VLS with task-and-motion planning (TAMP) for long horizons, where VLM-generated rewards guide local executions while planners ensure global feasibility and constraint satisfaction.
- Metrics for “correctness” of spatial reasoning: Create datasets with ground-truth geometric relations and measure the fidelity of P and Rs to those relations; report calibration curves linking reward values to actual task satisfaction.
- Resource and deployment constraints: Quantify memory and compute footprints for VLS components (SAM, DINOv2, VLM), assess on-device/internet-free operation, and propose lightweight surrogates for resource-constrained robots.
Practical Applications
Immediate Applications
The following applications can be deployed now using the paper’s training-free, inference-time steering of frozen diffusion/flow-matching robot policies, VLM-based reward generation, and closed-loop stage control. Each item notes likely sectors, tools/workflows that could emerge, and feasibility dependencies.
- Robotics (manufacturing and warehousing): adaptive pick-and-place when bins, shelves, or workcells change layout or new SKUs appear; reduces re-training cycles after reconfiguration or mild clutter.
- Tools/workflows: a VLS middleware in ROS/ROS 2 that wraps existing diffusion/flow policies; “Reward Compiler” that converts VLM outputs into PyTorch-differentiable stage rewards; deployment playbooks for guidance scaling and Schmitt-trigger stage switching.
- Assumptions/dependencies: RGB-D sensing and reliable depth; accurate segmentation (e.g., SAM) and feature grounding (DINOv2); sufficiently capable frozen base policy; GPU budget for batch sampling and resampling.
- Robotics (lab automation): inference-time adaptation to new glassware positions, swapped racks, or instrument substitutions without retraining; robust multi-stage execution (e.g., grasp-transfer-place) under small spatial shifts.
- Tools/workflows: VLS “experiment templates” for common lab tasks with VLM prompts and stage rewards; FK-resampling for sample-efficient adaptation; automatic stage switching with hysteresis to avoid oscillations.
- Assumptions/dependencies: consistent 3D keypoint grounding; safe motion primitives in the base policy; controlled lighting to stabilize perception.
- Service and domestic robots: home assistants that handle clutter, moved objects, and novel household items (e.g., “put the mug on the green plate” when plates swap locations).
- Tools/workflows: on-device VLS runtime for consumer robots; household task libraries with stage-wise rewards generated by VLMs from natural language; adaptive guidance scaling to keep motions smooth.
- Assumptions/dependencies: consumer-grade RGB-D or multi-view cameras; robust SAM/DINO performance on household scenes; safety policies limiting guidance strength near humans.
- Education and prototyping (academia, startups): rapid experimentation with OOD adaptation on benchmark suites (CALVIN, LIBERO-PRO) and real robots (e.g., Franka); course modules on inference-time control vs. retraining.
- Tools/workflows: an open-source VLS SDK with example prompts, reward functions, and ablations (gradient guidance, FK resampling, RBF diversity); notebooks for teaching denoising steering and stage control.
- Assumptions/dependencies: access to pretrained policies; modest GPU resources; stable VLMs for reward synthesis.
- Robotics integration services (industry): “retrofit” packages to make existing imitation-learned policies resilient to layout changes and new instructions; reduction of downtime after workcell tweaks.
- Tools/workflows: assessment pipelines to profile compute-performance trade-offs (batch size vs. latency); automated calibration of guidance hyperparameters; OOD acceptance tests based on success rate and episode length.
- Assumptions/dependencies: contractual access to base policies; compatibility with customer hardware and ROS stacks; defined OOD scenarios within the policy’s motor capability.
- Safety and compliance practice (policy and operations): OOD-robustness testing added to deployment checklists; data-lean adaptation (no new training data) that can ease privacy/compliance burdens.
- Tools/workflows: OOD test batteries (position shift, object swap, instruction changes) using VLS; reporting on guidance strength, stage transitions, and failure modes; documentation templates for risk assessment.
- Assumptions/dependencies: safety interlocks to cap steering effects; clear operational boundaries for tasks; audit logs of VLM prompts and reward code.
Long-Term Applications
These applications require further research, scaling, or development to meet real-time, safety, and generalization requirements. They build on the paper’s VLM-generated differentiable rewards, gradient-guided denoising, diversity/resampling, and closed-loop stage logic.
- Healthcare robotics (clinical and assistive): adaptation for non-critical tasks (instrument handling, patient-room tidying) under frequent layout and device changes; eventual extension to semi-autonomous assistive manipulation.
- Tools/products: clinically certified VLS layer with formalized stage rewards and safe guidance bounds; integrated failure detectors and human-in-the-loop overrides; hospital-specific prompt libraries.
- Dependencies: rigorous safety certification; real-time latency guarantees; domain-robust segmentation/grounding; base policies with medically approved motion primitives.
- Mobile manipulation and multi-robot systems: OOD steering for robots that move, perceive, and manipulate across rooms or dynamic environments; coordination where stage rewards incorporate inter-robot constraints.
- Tools/workflows: distributed VLS services that share grounded keypoints and stage status; multi-agent reward synthesis and FK-resampling across particle pools; ROS 2 graph integrations for team-level stage switching.
- Dependencies: robust, scalable perception outdoors/indoors; networking and synchronization; stronger base policies for locomotion+manipulation; compute-efficient steering (edge accelerators).
- Formal verification and safety guarantees for inference-time steering: provable bounds on reward-guided denoising and stage transitions; certified “safe sets” around humans.
- Tools/products: verified reward DSLs (domain-specific languages) compiled to differentiable code with static checks; conformance tests and formal models of hysteresis-based switching; supervisory controllers integrating VLS with control barrier functions.
- Dependencies: theoretical advances in guided diffusion/flow safety; standardized test suites for OOD with safety metrics; regulator engagement.
- Sector-agnostic automation (software, RPA with physical and virtual agents): extending VLS-like steering to GUI agents and hybrid physical-digital workflows where VLMs generate task-stage rewards over action sequences.
- Tools/workflows: generalized “vision-language steering” for UI actions (pointer movement, click sequences) with differentiable proxies; cross-domain stage switching (e.g., fetch item physically, then file digital record).
- Dependencies: appropriate differentiable scoring for non-physical actions; integration with enterprise IT and audit requirements; robust VLM reasoning over complex instructions.
- Energy and industrial maintenance robotics: adaptive inspection/maintenance under plant reconfigurations, clutter, and novel parts; test-time steering to respect spatial constraints around hazardous regions.
- Tools/products: VLS packs with plant-specific keypoint grounding and risk-aware rewards; scheduling systems that modulate guidance strength by hazard proximity; asset-change documentation tied to updated VLM prompts.
- Dependencies: high-reliability perception in harsh environments; safety barriers and remote oversight; compute acceleration for real-time response.
- Standardization and benchmarking (policy, academia, consortia): OOD-resilience standards for robot deployment; new public benchmarks that measure inference-time control effectiveness across spatial and semantic shifts.
- Tools/workflows: community datasets and leaderboards for steering methods (gradient guidance, resampling, diversity) across tasks and sensors; procurement standards specifying OOD test coverage and latency budgets.
- Dependencies: multi-stakeholder collaboration; reproducible evaluation protocols; openness around pretrained policies and reward synthesis prompts.
- Hardware-software co-design for real-time VLS: accelerators and perception pipelines that cut latency of batch sampling, MCMC refinement, and FK resampling to meet sub-100 ms control loops.
- Tools/products: GPU/ASIC kernels optimized for guided denoising and gradient aggregation; fast SAM/DINO variants or task-specific segmenters; on-robot VLM distillations for low-latency reward compilation.
- Dependencies: engineering investment; careful trade-offs between diversity and speed; maintenance of accuracy under compression/distillation.
- Curriculum-style reward generation and autonomous task decomposition: VLMs that learn to produce progressively shaped, stage-aware rewards for complex, long-horizon tasks with minimal human prompt tuning.
- Tools/workflows: meta-learning pipelines that adapt reward programs to new domains; repositories of reusable stage schemas; automatic detection of when to transition stages or retry strategies.
- Dependencies: improved VLM grounding and consistency; mechanisms to avoid reward hacking; robust feedback from execution metrics.
Notes across all applications:
- Compute-performance trade-offs are central: larger particle batches and inner refinement steps improve success but raise latency; deployment must tune guidance strength and batch size to meet real-time constraints.
- Success hinges on the base policy’s motor competence; VLS steers existing skills but cannot create skills absent from training.
- Perception robustness (segmentation, depth, feature grounding) and prompt design critically affect reward quality and safety.
Glossary
- Action chunk: A contiguous block of planned actions over a fixed horizon used as the policy’s prediction unit. "an action chunk at:t+T with chunk horizon T"
- Articulated parts: Objects with joints or movable components (e.g., doors, drawers) requiring specific manipulation strategies. "articulated parts (drawer, switch, button, door)"
- Classifier Guidance: A technique for diffusion models that injects gradients from a classifier to bias generation toward a condition without retraining. "we leverage Classifier Guidance [13] to steer the sampling process of the base policy."
- Closed-loop execution control: A feedback-driven mechanism that adjusts guidance and stage transitions during execution for robustness. "VLS incorporates a closed-loop execution control mechanism."
- Covariate shift: A change in input distribution between training and deployment that can degrade performance. "trajectories remain within an expert manifold under covariate shift"
- Denoising path: The sequence of intermediate states followed by a generative model as it transforms noise into a sample. "by correcting the denoising path."
- DINOv2: A self-supervised vision transformer used to extract dense visual features aligned with semantics. "extract semantically aligned dense visual features using DINOv2 [8]"
- Expert manifold: The subset of state-action trajectories represented in expert demonstrations, used as a prior for behavior. "so trajectories remain within an expert manifold"
- Feynman-Kac steering: A particle-based resampling method that weights samples by reward-derived potentials to guide generation. "VLS employs a gradient-free resampling mechanism based on Feynman-Kac (FK) steering [12, 14, 44]."
- Flow matching: A generative modeling approach that learns a continuous-time velocity field to transform noise into data. "flow matching simplifies the denoising process by learning a continuous velocity field v."
- Guidance scale (hyperparameter): A scalar that controls the strength of injected guidance during sampling. "where X is the guidance scale hypermeter to control the guidance strength."
- Imitation learning: Learning a policy from expert demonstrations by maximizing the likelihood of demonstrated actions. "Imitation learning aims to learn a policy Te from an expert demonstration dataset"
- Inference-time steering: Modulating a pretrained model’s sampling process at test time to satisfy new conditions without updating parameters. "a training-free framework for inference-time steering of frozen generative robot policies."
- Interacting particle system: A set of parallel samples that evolve with weighting and resampling, used to explore solution space. "We interpret the batch of action proposals as an interacting particle system"
- Keypoints (3D keypoints): Compact spatial anchors extracted from scenes to represent task-relevant geometry for reward evaluation. "a compact geometric scaffold P of task-relevant 3D keypoints"
- MCMC-based guidance: Refinement using multiple inner stochastic updates per step to better explore and optimize under guidance. "analogous to MCMC-based guidance [16, 49, 17]."
- Multinomial resampling: A procedure that resamples particles according to normalized weights to focus on high-reward proposals. "multinomial resampling is applied to the particle set."
- Noise schedule coefficients: Parameters controlling the variance and mean in each diffusion step during forward/reverse processes. "ak and ok are noise schedule coefficients"
- Ordinary Differential Equation (ODE): A continuous-time differential equation used here to model flow-matching dynamics. "Flow matching models the transition of distribution as an Ordinary Differential Equation (ODE):"
- Out-of-distribution (OOD): Inputs or conditions that differ from the training distribution, often causing model brittleness. "out-of-distribution (OOD) observation-language inputs"
- Potential field: A scalar field over actions that encodes preferences or constraints; its gradients guide optimization. "Rs defines a stage-specific potential field over the action space"
- Proprioception: Internal robot state signals (e.g., joint positions, gripper state) used as part of observations. "typically RGB images and robot proprioception"
- Programmatic reward functions: Differentiable, code-defined scoring functions (e.g., in PyTorch) synthesized to evaluate action proposals. "synthesize stage-aware, programmatic reward functions"
- RBF repulsion: A diversity-promoting mechanism using radial basis function terms to push samples apart. "incorporating RBF [24] repulsion terms"
- Repulsive gradient: A gradient term that increases pairwise distances among samples to prevent mode collapse. "we define a repulsive gradient based on pairwise distances:"
- RGB-D: Combined color and depth sensing used for scene observation in robotic tasks. "given RGB-D observation ot"
- Schmitt trigger: A hysteresis-based switching mechanism preventing oscillations by using separate thresholds for activation/deactivation. "Schmitt-trigger [43] switching logic"
- Segment Anything Model (SAM): A foundation segmentation model used to extract object masks from images. "Segment Anything Model (SAM [29])"
- Stage-aware reward functions: Rewards tailored to sequential task phases to provide context-specific guidance during execution. "stage-aware, programmatic reward functions"
- Trajectory-differentiable reward functions: Rewards that are differentiable with respect to entire action sequences, enabling gradient guidance. "synthesize trajectory- differentiable reward functions"
- Transition kernel: The stochastic rule defining how samples evolve during generation; can be tilted toward target distributions. "tilts the transition kernel of the generative process"
- Value/Q model: A learned critic that estimates future return (value or Q) and can inject gradients to bias generation. "injects gradients from a learned value/Q model into de- noising"
- Velocity field: A vector field specifying instantaneous motion in flow matching for transforming noise into data. "learning a continuous velocity field v."
- Vision-Language Action (VLA) policies: Policies conditioned on visual and language inputs that output robot actions. "frozen VLA policies"
- Vision-LLM (VLM): A model that jointly processes visual inputs and language to provide semantic understanding. "vision-LLMs (VLMs)"
- Vision-Language Steering (VLS): The proposed framework that uses VLM-synthesized differentiable rewards to steer pretrained robot policies at test time. "We present Vision-Language Steering (VLS), a training-free framework"
Collections
Sign up for free to add this paper to one or more collections.