Papers
Topics
Authors
Recent
Search
2000 character limit reached

Articulated Object Estimation in the Wild

Published 1 Sep 2025 in cs.RO and cs.CV | (2509.01708v1)

Abstract: Understanding the 3D motion of articulated objects is essential in robotic scene understanding, mobile manipulation, and motion planning. Prior methods for articulation estimation have primarily focused on controlled settings, assuming either fixed camera viewpoints or direct observations of various object states, which tend to fail in more realistic unconstrained environments. In contrast, humans effortlessly infer articulation by watching others manipulate objects. Inspired by this, we introduce ArtiPoint, a novel estimation framework that can infer articulated object models under dynamic camera motion and partial observability. By combining deep point tracking with a factor graph optimization framework, ArtiPoint robustly estimates articulated part trajectories and articulation axes directly from raw RGB-D videos. To foster future research in this domain, we introduce Arti4D, the first ego-centric in-the-wild dataset that captures articulated object interactions at a scene level, accompanied by articulation labels and ground-truth camera poses. We benchmark ArtiPoint against a range of classical and learning-based baselines, demonstrating its superior performance on Arti4D. We make code and Arti4D publicly available at https://artipoint.cs.uni-freiburg.de.

Summary

  • The paper introduces ArtiPoint, which uses deep point tracking and factor graph optimization to estimate articulated object models in dynamic environments.
  • The methodology leverages ego-centric RGB-D video and a novel Arti4D dataset to overcome limitations of fixed-camera setups in practical, cluttered scenes.
  • Experimental results show lower angular and translational errors compared to existing approaches, ensuring high accuracy in revolute and prismatic joint estimation.

Articulated Object Estimation in the Wild

The paper "Articulated Object Estimation in the Wild" introduces a novel framework, ArtiPoint, capable of estimating articulated object models in dynamic environments from ego-centric RGB-D inputs under conditions of partial observability. This method addresses the limitations of previous approaches that are constrained to fixed camera positions and isolated objects, often failing in real-world, cluttered scenes.

Introduction

Robotic manipulation involves moving objects to achieve desired goals, and the complexity increases significantly when dealing with articulated objects. Previous solutions have predominantly leveraged stable environments with static cameras, which limits their applicability in practical settings such as domestic environments or workplaces where conditions are dynamic. ArtiPoint provides a solution by employing deep point tracking combined with factor graph optimization. This method allows for robust estimation using only raw video inputs without relying on pre-trained deep models that may overfit. Figure 1

Figure 1: We present ArtiPoint, our novel approach to articulated object estimation in the wild. ArtiPoint makes use of deep point tracking and factor graph optimization.

Methodology

ArtiPoint's methodology is centered around leveraging human interactions with objects as cues for estimating articulated motions. This is divided into four stages: interaction detection, point tracking, 3D trajectory lifting and filtering, and articulation model estimation.

  1. Interaction Interval Extraction: Using an ego-centric video feed, hand masks are extracted to identify segments of interaction. A moving average computes the visibility of hands, creating time intervals corresponding to object interaction scenarios.
  2. Point Tracking: CoTracker3, an advanced point tracking system, is employed to track points around detected hands, efficiently gathering data on the object's movement trajectory during interactions.
  3. 3D Trajectory Estimation and Filtering: Extracted 2D trajectories are converted to 3D using depth readings, and further local transformations are applied to compensate for the camera motion. Filtering techniques remove static points and occlusions, enhancing the accuracy of motion estimation.
  4. Articulation Model Estimation: Factor graph formulations are applied to the trajectory data, flexibly modeling complex joint types, such as revolute and prismatic joints, in 6D space by leveraging SE(3)SE(3) transformations. Figure 2

    Figure 2: Overview of the method: Interaction segments are identified and point trajectories are extracted and processed to estimate articulation models.

Dataset: Arti4D

The introduction of Arti4D, a dataset recorded with an ego-centric camera, provides the first real-world scene-level annotated interactions of articulated objects. This dataset includes interaction windows, axis labels, and camera pose ground truths, offering a comprehensive benchmark for future research. Arti4D covers various environments, capturing diverse and complex interactions, from kitchen scenes to robotics labs, under dynamic camera movements and real-world scene complications. Figure 3

Figure 3

Figure 3

Figure 3

Figure 3: Reconstructed scenes of the four Arti4D environments.

Experimental Results

The experimental evaluation demonstrates that ArtiPoint outperforms existing methods, including both classical factor graph solutions and contemporary deep learning models. Compared to alternatives like Ditto and ArtGS, ArtiPoint shows lower angular and translational errors across both revolute and prismatic joints.

Quantitative results detail ArtiPoint's robust performance, with metrics showing the lowest errors and highest type prediction accuracy in prismatic and revolute joint identification. The deployment of ArtiPoint using realistic camera odometry further confirms its stability in unpredictable environments. Figure 4

Figure 4: Qualitative results on Arti4D: Estimated joint axis and pose trajectory demonstrate ArtiPoint’s effectiveness in complex scenes.

Conclusion

ArtiPoint marks a significant step forward in articulated object estimation by enabling systems to understand complex interactions in dynamic settings. This work not only proposes an innovative method but also presents a challenging dataset to drive future research. As AI continues to integrate into real-world applications, methods like ArtiPoint are essential for enhancing robotic interaction capabilities. Future advancements could further refine these techniques, making autonomous systems even more adaptive and resilient in diverse environments. Figure 5

Figure 5: Scene-level prediction results showcasing the comprehensive output of the ArtiPoint method.

In summary, ArtiPoint offers a robust and scalable approach to articulated object estimation, setting a new standard for real-world applicability. Integrating this system into broader robotic platforms holds promise for transformative applications in everyday environments.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.