MobileWorldBench: Semantic Mobile GUI Benchmark
- MobileWorldBench is a benchmark suite that systematically evaluates semantic world modeling for mobile GUI agents using natural-language state predictions.
- It comprises 1.4M annotated transitions from real mobile interactions, offering rigorous datasets and metrics for next-state generation and yes/no QA tasks.
- Integration with model-based planning pipelines shows significant improvements in task success rates and decision-making over traditional pixel-based models.
MobileWorldBench is a benchmark and dataset suite that systematically evaluates semantic world modeling for mobile GUI agents, focusing on the capacity of vision–LLMs (VLMs) to predict and reason over future mobile app states in abstract, decision-relevant natural language, rather than low-level pixel outputs. Conceived to address the limitations of pixel-space world models in complex GUI environments, MobileWorldBench and its associated MobileWorld dataset offer comprehensive empirical infrastructure for advancing model-based planning in mobile interaction scenarios, scaling to 1.4 million annotated transitions and integrated with concrete agent planning frameworks (Li et al., 16 Dec 2025).
1. Benchmark Definition and Objectives
The fundamental goal of MobileWorldBench is to enable rigorous, large-scale evaluation of a VLM's ability to model state transitions in mobile GUIs at the semantic level. The benchmark is built on two primary tasks:
- Next-State-Generation: Given a current screen and an action , the model generates a natural-language description of the resulting state change.
- Next-State-QA: Given , the model answers a set of yes/no questions about the predicted next state .
This formulation is motivated by the impracticality of pixel-level world models in GUI settings, where the diversity of layouts, textual content, and icons makes faithful pixel-wise prediction intractable and unnecessary for decision-making. Instead, semantic world models abstract the dynamics into high-level, goal-relevant descriptions—sufficient for downstream planning frameworks.
2. Dataset Composition and Annotation Pipeline
MobileWorld, the dataset underpinning MobileWorldBench, consists of 1.4 million triplets. The data is minimally curated from human interaction trajectories recorded in large-scale mobile environments such as Android in the Wild (AiTW) and AndroidControl. Each transition is annotated with:
- Three candidate natural-language state-change descriptions, with best-of-three human selection ().
- Approximately eight candidate yes/no QA pairs per transition, filtered for relevance and correctness, yielding 540,000 high-quality pairs.
The annotation pipeline is highly structured:
- Trajectory Collection: Automated recording of user actions and resulting GUIs.
- VLM Annotation: High-level actions derived from low-level gestures and visual overlays, prompting a VLM to propose candidate descriptions and questions.
- Automated Filtering: VLM self-checking (answering its own questions using ), with irrelevant pairs pruned.
- Human Verification: Manual curation to ensure correctness, focus, and non-ambiguity, culminating in a finalized pool of 1,787 human-verified QA items for benchmark evaluation.
A typical sample is:
3. Semantic World Model Formalism
MobileWorldBench frames semantic world modeling as a latent variable generative process:
where represents the latent semantic change. The VLM is trained to model
and supports two operational modes:
- Text Generation:
- Yes/No QA: For each ,
The training objective is a joint cross-entropy loss over generation and QA heads:
4. Vision–LLM Architecture and Training
The core model is Qwen3-VL-8B-Instruct, a multimodal transformer with separate vision and language modules:
- Vision Encoder: Patch-based ViT encoder followed by projection and cross-attention integration into the text stack.
- LLM: Decoder-only transformer optimized for instruction-following.
- Input Representation: Encoded screenshot , natural-language action , and optional question with interleaved cross-modal attention.
- Optimization Regimen: Fine-tuned for 2 epochs on MobileWorld using 8 × NVIDIA A6000, AdamW optimizer, batch size 128, with learning rates (LLM) and (vision encoder).
5. Integration with Model-Based Planning Pipelines
MobileWorldBench is explicitly designed for integration into model-based agent planning. The canonical inference loop is as follows:
- At step , given goal and state , an action proposal model generates candidates .
- For each , the VLM-based semantic world model predicts .
- A free-form description is decoded.
- Each is scored via a value model .
- The agent selects .
This one-step lookahead is an approximate maximization of the expected-reward objective:
where is value-model-predicted reward.
6. Evaluation Metrics and Empirical Findings
Evaluation is performed on both text generation and QA axes:
- Next-State-Generation: VLM-judge metrics for accuracy (0–5), relevance (0–5), completeness (0–5), with an overall score in .
- Next-State-QA: Fraction of correctly answered yes/no questions (binary accuracy).
Benchmark splits:
- Generation: 250 held-out transitions.
- QA: 500 held-out transitions, 1,787 QA pairs.
Quantitative improvements from MobileWorld-finetuned VLM (vs. base Qwen3-VL-8B-Instruct):
| Metric | Baseline | Finetuned | Absolute Gain |
|---|---|---|---|
| Generation Overall (/15) | 11.84 | 12.39 | +4.7% |
| QA Accuracy (%) | 67.32 | 71.40 | +4.08 pts |
| Downstream Task SR (%) | 46.9 | 54.3 | +7.4 pts |
Ablation experiments indicate the VLM-annotator pipeline (VLM-generated training labels) outperforms off-the-shelf VLMs, supporting the validity of the data construction.
7. Implications and Research Significance
- Tractability: Semantic world modeling bypasses the intractable combinatorics of pixel forecasting in GUIs, allowing direct optimization for decision-relevant abstractions.
- Benchmark Scale and Quality: MobileWorldBench is the first large-scale, human-verified benchmark for semantic next-state prediction in mobile UIs, with 1.4 M annotated samples.
- Empirical Utility: Finetuned VLMs on MobileWorld produce significant gains in both state-change generation, question answering, and end-to-end AndroidWorld task success, illustrating the real-world benefits of semantic world modeling.
- Planning Integration: Model-based planning with a learned value model over semantic summaries yields a >7 percentage point increase in task completion, directly affecting agent robustness in realistic settings.
- Future Directions: This semantic abstraction approach supports further development of high-level reasoning in embodied and GUI agents, with expected impact on robust, sample-efficient task completion and transferability across diverse UIs.
MobileWorldBench is openly available, including dataset, benchmark splits, and model checkpoints, facilitating broad adoption and reproducibility in research on semantic world models for mobile agents (Li et al., 16 Dec 2025).