Papers
Topics
Authors
Recent
Search
2000 character limit reached

Augmented Reality Remote Operation of Dual Arm Manipulators in Hot Boxes

Published 28 Mar 2023 in cs.RO | (2303.16055v1)

Abstract: In nuclear isotope and chemistry laboratories, hot cells and gloveboxes provide scientists with a controlled and safe environment to perform experiments. Working on experiments in these isolated containment cells requires scientists to be physically present. For hot cell work today, scientists manipulate equipment and radioactive material inside through a bilateral mechanical control mechanism. Motions produced outside the cell with the master control levers are mechanically transferred to the internal grippers inside the shielded containment cell. There is a growing need to have the capability to conduct experiments within these cells remotely. A simple method to enable remote manipulations within hot cell and glovebox cells is to mount two robotic arms inside a box to mimic the motions of human hands. An AR application was built in this work to allow a user wearing a Microsoft HoloLens 2 headset to teleoperate dual arm manipulators by grasping robotic end-effector digital replicas in AR from a remote location. In addition to the real-time replica of the physical robotic arms in AR, the application enables users to view a live video stream attached to the robotic arms and parse a 3D point cloud of 3D objects in their remote AR environment for better situational awareness. This work also provides users with virtual fixture to assist in manipulation and other teleoperation tasks.

Citations (3)

Summary

  • The paper introduces an AR interface that enables remote control of dual-arm robotic manipulators in hazardous hot box environments.
  • The paper details a system architecture integrating HoloLens 2, ROS, and Kinova Gen3 arms for intuitive gesture-based manipulation.
  • Results demonstrate practical application with real-time updates through live video streams and 3D point cloud visualizations to enhance precision.

Augmented Reality Interface for Remote Hot Box Manipulation

This paper introduces an augmented reality (AR) interface designed to enable remote operation of dual-arm manipulators within hot box environments, addressing the challenges of conducting experiments in hazardous or isolated settings. The system leverages a Microsoft HoloLens 2 headset to provide users with an intuitive and natural control interface, enhancing situational awareness through real-time visualization and interactive elements.

System Architecture and Components

The system integrates a HoloLens 2, ROS packages, and Kinova Gen3 robotic arms to facilitate remote operation through natural controls. The Unity game engine, in conjunction with ROS package scripts, establishes a communication pipeline between the headset and the robotic arms. Users can manipulate the robotic arms by pinching their thumb and index finger on virtual arm grippers within the AR environment. The hand poses tracked by the HoloLens 2 are converted into serialized messages and transmitted to an Ubuntu machine functioning as a control station server. This server deserializes the messages and broadcasts them to specific scripts, which interpret the messages and convert them into twist commands for the robotic arms. Positional updates from the robotic arms are relayed back to the HoloLens 2, updating the virtual arm representations in real-time. Figure 1

Figure 1: [Left] New hot box design with two robotic arms mounted to the top manipulating chemistry tools. [Right] First-person-view of a HoloLens 2 user seeing both a virtual replica of the hot box atop the table and the real physical hotbox on the right.

Software Implementation

The AR application is built within the Unity game engine, providing HoloLens 2 users with virtual representations of the hot box and processing hand tracking data. The Unity application publishes hand positions and subscribes to the robotic arms' joint positions via ROS topics on a RosBridge server. Standard PoseStamped and JointState ROS messages are used for hand and robot joint positions, respectively. The messages are serialized into JSON format before transmission over the wireless network to ensure rapid data translation across the local router. A ROS package called RosBridge, running on a Linux server, receives and deserializes these messages, broadcasting them locally to ROS nodes on the Linux machine. The Kinova AR Teleop Package, a custom ROS package, converts the PoseStamped positions of the hands into linear and angular velocities (Twist messages) that the robots can understand. These Twist messages are then published to the Kinova ROS Kortex package, which sends the move commands to the arms, utilizing its internal inverse kinematic solver. The ROS Kortex package continuously publishes JointState information back to the Kinova AR Teleop Package, which packages, serializes, and sends the data back to the HoloLens 2 for updating the virtual arm positions. Figure 2

Figure 2: Images captured from the perspective of a HoloLens 2 user. [Left] User pinching the virtual end effector commanding the physical arm to move. [Middle] Robotic arm interacting with a virtual plane in AR used to assist in teleoperation. [Right] Virtual point cloud visualization of the physical tube.

Practical Applications and Testing

The AR application has been successfully implemented to control both Kinova dual-arm and ReThink Robotics Baxter dual-arm manipulators. Users can manipulate the robots remotely using pinch and grab gestures. The interface includes togglable features such as a live video stream, virtual fixtures, and point cloud visualizations. The virtual robots can also be scaled within the scene to facilitate precise manipulations or confined space arrangements. Figure 3

Figure 3: System overview of how the HoloLens 2 integrates with the robotic manipulators inside of the hot box.

Conclusion and Future Directions

The AR application enables scientists to remotely operate robotic manipulators within a hot box environment using hand gestures. The system incorporates a live video stream, 3D point cloud visualization, and virtual fixtures to enhance the user experience. Future research will focus on optimizing hand gesturing methods for controlling the robotic arms' position and orientation. The HoloLens 2 captures extensive sensor data, offering opportunities to explore the use of voice, eye gaze, and head position to further improve the remote operation experience.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.