Papers
Topics
Authors
Recent
Search
2000 character limit reached

Mobile-Agent-v3.5: Multi-platform Fundamental GUI Agents

Published 15 Feb 2026 in cs.AI and cs.CL | (2602.16855v1)

Abstract: The paper introduces GUI-Owl-1.5, the latest native GUI agent model that features instruct/thinking variants in multiple sizes (2B/4B/8B/32B/235B) and supports a range of platforms (desktop, mobile, browser, and more) to enable cloud-edge collaboration and real-time interaction. GUI-Owl-1.5 achieves state-of-the-art results on more than 20+ GUI benchmarks on open-source models: (1) on GUI automation tasks, it obtains 56.5 on OSWorld, 71.6 on AndroidWorld, and 48.4 on WebArena; (2) on grounding tasks, it obtains 80.3 on ScreenSpotPro; (3) on tool-calling tasks, it obtains 47.6 on OSWorld-MCP, and 46.8 on MobileWorld; (4) on memory and knowledge tasks, it obtains 75.5 on GUI-Knowledge Bench. GUI-Owl-1.5 incorporates several key innovations: (1) Hybird Data Flywheel: we construct the data pipeline for UI understanding and trajectory generation based on a combination of simulated environments and cloud-based sandbox environments, in order to improve the efficiency and quality of data collection. (2) Unified Enhancement of Agent Capabilities: we use a unified thought-synthesis pipeline to enhance the model's reasoning capabilities, while placing particular emphasis on improving key agent abilities, including Tool/MCP use, memory and multi-agent adaptation; (3) Multi-platform Environment RL Scaling: We propose a new environment RL algorithm, MRPO, to address the challenges of multi-platform conflicts and the low training efficiency of long-horizon tasks. The GUI-Owl-1.5 models are open-sourced, and an online cloud-sandbox demo is available at https://github.com/X-PLUG/MobileAgent.

Summary

  • The paper introduces a unified GUI agent framework that enables multi-turn decision-making across desktop, mobile, browser, and edge platforms.
  • The paper proposes a hybrid data flywheel and MRPO-based reinforcement learning algorithm to enhance training efficiency and cross-platform generalization.
  • The paper demonstrates state-of-the-art performance on diverse GUI benchmarks, validating improvements in tool invocation, memory management, and multi-agent collaboration.

Mobile-Agent-v3.5: Multi-platform Fundamental GUI Agents

Motivation and Model Overview

"Mobile-Agent-v3.5: Multi-platform Fundamental GUI Agents" (2602.16855) presents GUI-Owl-1.5, a native multimodal agent model designed for autonomous operation across desktop, mobile, browser, and edge platforms. The system addresses the performance gap and practical usability bottlenecks in GUI agents by introducing extensive architectural, data, and training innovations. GUI-Owl-1.5 is instantiated in instruct/thinking variants ranging from 2B up to 235B parameters and supports cloud-edge collaboration and real-time interaction scenarios. Model variants are tailored to edge deployment (fast, lightweight instruct models) or cloud-based orchestration (reasoning-enriched thinking models), supporting distributed multi-agent coordination. Figure 1

Figure 1: Multi-platform architecture and interface abstraction of Mobile-Agent-v3.5, highlighting unified environment and agent capability.

The agent is formulated as a multi-turn decision-making policy. The input at each step encompasses a screenshot and a natural language command, while the output includes an explicit reasoning chain, a planned action summary, and a structured function call for execution. The closed-loop design enables interaction with dynamical environments, expanding the action space to support external tool invocation and multi-device orchestration. Figure 2

Figure 2: The interaction flow of GUI-Owl-1.5 depicting closed-loop reasoning, multimodal context, and action outputs.

Hybrid Data Flywheel: Data Construction and Trajectory Generation

A key challenge in GUI agent training is efficient, high-quality data acquisition across heterogeneous environments. The paper establishes a Hybrid Data Flywheel for both grounding and trajectory supervision:

  • Grounding data is synthesized via annotated UI element generation, multi-window spatial configurations, tutorial-based QA mining, trajectory extraction, and infeasible query construction, addressing diversity insufficiency in offline datasets. Figure 3

    Figure 3: The grounding pipeline integrates synthetic scenarios, knowledge extraction, and trajectory mining for high-quality GUI element alignment.

  • Trajectory corpora are generated using a DAG-based synthesis protocol supporting both real device and scalable virtual environments. Human annotation supplements unsolved or hard cases; automated rollouts benefit from precise sub-task validation and coverage expansion. Figure 4

    Figure 4: Unified trajectory data pipeline combining DAG synthesis, automated rollout, annotation, and virtual env feedback.

Virtual environments enable trajectory generation for scenarios disrupted by anti-bot mechanisms or lacking accurate feedback, allowing atomic operation scripting and precise progress supervision. Ablations demonstrate that removing virtual environment data severely degrades performance on atomic and complex operations.

Unified Agent Capability Enhancement

GUI-Owl-1.5 adopts a unified enhancement pipeline:

  • GUI knowledge injection leverages tutorial QA, web navigation data, and world modeling—training the model to anticipate interface state transitions post-action.
  • Chain-of-thought augmentation synthesizes step-wise reasoning for each trajectory, including memory management, progress reflection, tool invocation rationale, and screen observation. This instills long-horizon task planning and in-context retention.
  • Multi-agent collaboration data is collected with structured role decomposition (Manager, Worker, Reflector, Notetaker). This enables specialization within coordinated multi-agent setups and closed-loop feedback loops. Figure 5

    Figure 5: Agent capability enhancement pipeline: world modeling, CoT augmentation, memory management, and multi-agent collaboration.

Ablations confirm both virtual environment trajectories and CoT augmentation are critical for robust agent capabilities.

Multi-platform RL Scaling: MRPO Algorithm

To tackle instability and inefficiency in RL optimization for GUI agents, especially across device families, the paper proposes Multi-platform Reinforcement Policy Optimization (MRPO):

  • Device-conditioned unified policy learning, with alternating single-device-family optimization to mitigate cross-platform gradient interference and preserve generalization.
  • An online rollout buffer oversamples and subsamples trajectory groups to enhance diversity under outcome collapse, maintaining strict on-policy guarantees.
  • Token-ID transport is used to align log-probability computation between environment-side inference and training, avoiding artifacts due to tokenizer discrepancies.
  • RL training strategies (unstable-set filtering and interleaved platform optimization) achieve faster convergence and greater stability compared to naive approaches. Figure 6

    Figure 6: MRPO reinforcement learning pipeline with device-conditioned policy, rollout buffer, and token-ID alignment.

    Figure 7

    Figure 7: Ablation study showing benefits of unstable-task prioritization and interleaved multi-platform RL training.

Empirical Results and Task Benchmarking

GUI-Owl-1.5 exhibits strong performance across more than 20 GUI benchmarks, including OSWorld, AndroidWorld, WebArena, ScreenSpot-Pro, OSWorld-MCP, MobileWorld, MemGUI-Bench, and GUI Knowledge Bench. On GUI automation, agentic tool use, grounding, memory, and knowledge tasks, the model achieves state-of-the-art or competitive accuracy for both single-device and multi-platform scenarios:

  • OSWorld: 56.5 (32B-Instruct)
  • AndroidWorld: 71.6 (8B-Thinking)
  • WebArena: 48.4 (32B-Thinking)
  • ScreenSpot-Pro (grounding): 80.3 with crop refinement (32B-Instruct)
  • OSWorld-MCP and Mobile-World: 47.6/46.8 (32B-Instruct)
  • GUI Knowledge Bench: 75.45
  • MemGUI-Bench: 27.1 (32B), outperforming prior native agents Figure 8

    Figure 8: End-to-end performance summary for mainstream GUI task automation, grounding, and knowledge benchmarks.

The thinking variants outperform instruct counterparts, notably on long-horizon tasks such as WebVoyager and Online-Mind2Web. The multi-agent and memory abilities, validated on MemGUI-Bench and GUI-Knowledge, confirm the efficacy of the enhancement pipeline.

Case Studies: Mobile, Desktop, and Tool-Augmented Operations

GUI-Owl-1.5 demonstrates autonomous long-horizon task execution and robust memory management in both mobile and desktop environments:

  • Android: Search and aggregate user information from multiple apps, synthesizing summaries using memory retention and cross-application reasoning. Figure 9

    Figure 9: Android operation: agent searches and summarizes cross-platform user information, utilizing memory and reasoning.

  • Windows: Web search and notetaking, integrating task progress tracking and memorization for spreadsheet population. Figure 10

    Figure 10: Windows operation: agent performs multi-step web search and populates spreadsheet with retained information.

  • Desktop: Coordinated tool invocation and API calls for script editing, execution, and output verification. Figure 11

    Figure 11: Desktop operation: agent combines tool calls and GUI actions for autonomous code editing and execution.

Practical and Theoretical Implications

GUI-Owl-1.5 refines the paradigm for generalist GUI agents, demonstrating robust cross-device orchestration, scalable RL training, and structured agent capability enhancement. The explicit separation of roles and memory, pipeline-based data generation, and rigorous RL strategies provide a foundation for capacity scaling and real-world integration in device automation, tool use, and agentic workflows.

Practically, GUI-Owl-1.5 is suited for real-world deployment ranging from lightweight edge devices to distributed cloud-agent setups for automation and system administration use cases. The open-sourced models and cloud sandbox enable reproducibility and extension. Theoretically, the hybrid training paradigm and interleaved RL scheduling may inform future developments in agent architectures for embodied, multi-platform, and tool-augmented interaction.

Conclusion

Mobile-Agent-v3.5 introduces GUI-Owl-1.5, a native, multi-platform GUI agent model featuring explicit architectural, data, and RL innovations. The model achieves state-of-the-art results across fundamental GUI automation, grounding, tool calling, memory, and knowledge benchmarks, validating its generalized and scalable agentic capabilities. The structured approach to data synthesis, agent enhancement, and RL optimization sets a new standard for foundational GUI agents, underpinning future theoretical and practical advances in autonomous agent design and deployment.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 56 likes about this paper.