Papers
Topics
Authors
Recent
Search
2000 character limit reached

Leveraging Multi-modal Sensing for Robotic Insertion Tasks in R&D Laboratories

Published 2 Jul 2023 in cs.RO | (2307.00671v1)

Abstract: Performing a large volume of experiments in Chemistry labs creates repetitive actions costing researchers time, automating these routines is highly desirable. Previous experiments in robotic chemistry have performed high numbers of experiments autonomously, however, these processes rely on automated machines in all stages from solid or liquid addition to analysis of the final product. In these systems every transition between machine requires the robotic chemist to pick and place glass vials, however, this is currently performed using open loop methods which require all equipment being used by the robot to be in well defined known locations. We seek to begin closing the loop in this vial handling process in a way which also fosters human-robot collaboration in the chemistry lab environment. To do this the robot must be able to detect valid placement positions for the vials it is collecting, and reliably insert them into the detected locations. We create a single modality visual method for estimating placement locations to provide a baseline before introducing two additional methods of feedback (force and tactile feedback). Our visual method uses a combination of classic computer vision methods and a CNN discriminator to detect possible insertion points, then a vial is grasped and positioned above an insertion point and the multi-modal methods guide the final insertion movements using an efficient search pattern. Through our experiments we show the baseline insertion rate of 48.78% improves to 89.55% with the addition of our "force and vision" multi-modal feedback method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. B. Burger, P. M. Maffettone, V. V. Gusev, C. M. Aitchison, Y. Bai, X. Wang, X. Li, B. M. Alston, B. Li, R. Clowes, N. Rankin, B. Harris, R. S. Sprick, and A. I. Cooper, “A mobile robotic chemist,” Nature, vol. 583, no. 7815, pp. 237–241, Jul. 2020, number: 7815 Publisher: Nature Publishing Group. [Online]. Available: https://www.nature.com/articles/s41586-020-2442-2
  2. H. Fakhruldeen, G. Pizzuto, J. Glowacki, and A. I. Cooper, “Archemist: Autonomous robotic chemistry system architecture,” IEEE International Conference on Robotics and Automation, 2022.
  3. P. Shiri, V. Lai, T. Zepel, D. Griffin, J. Reifman, S. Clark, S. Grunert, L. P. Yunker, S. Steiner, H. Situ, F. Yang, P. L. Prieto, and J. E. Hein, “Automated solubility screening platform using computer vision,” iScience, vol. 24, no. 3, p. 102176, 2021. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S2589004221001449
  4. M. Lambeta, P.-W. Chou, S. Tian, B. Yang, B. Maloon, V. R. Most, D. Stroud, R. Santos, A. Byagowi, G. Kammerer, D. Jayaraman, and R. Calandra, “DIGIT: A Novel Design for a Low-Cost Compact High-Resolution Tactile Sensor with Application to In-Hand Manipulation,” IEEE Robotics and Automation Letters, vol. 5, no. 3, pp. 3838–3845, Jul. 2020, arXiv:2005.14679 [cs, eess, stat]. [Online]. Available: http://arxiv.org/abs/2005.14679
  5. L. Pecyna, S. Dong, and S. Luo, “Visual-Tactile Multimodality for Following Deformable Linear Objects Using Reinforcement Learning,” Mar. 2022, arXiv:2204.00117 [cs]. [Online]. Available: http://arxiv.org/abs/2204.00117
  6. M. A. Lee, Y. Zhu, P. Zachares, M. Tan, K. Srinivasan, S. Savarese, L. Fei-Fei, A. Garg, and J. Bohg, “Making Sense of Vision and Touch: Learning Multimodal Representations for Contact-Rich Tasks,” IEEE Transactions on Robotics, vol. 36, no. 3, pp. 582–596, Jun. 2020, conference Name: IEEE Transactions on Robotics.
  7. G. Pizzuto, J. de Berardinis, L. Longley, H. Fakhruldeen, and A. I. Cooper, “Solis: Autonomous solubility screening using deep neural networks,” IEEE International Joint Conference on Neural Networks, 2022.
  8. J. X.-Y. Lim, D. Leow, Q.-C. Pham, and C.-H. Tan, “Development of a robotic system for automatic organic chemistry synthesis,” IEEE Transactions on Automation Science and Engineering, vol. 18, no. 4, pp. 2185–2190, 2021.
  9. H. Fleischer, L. A. Kroos, S. Joshi, T. Roddelkopf, R. Stoll, and K. Thurow, “Dual-arm robotic compound-oriented measurement system: Integration of a positive pressure solid phase extraction unit,” in 2021 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), 2021, pp. 1–6.
  10. H. Fleischer, R. R. Drews, J. Janson, B. R. C. Patlolla, X. Chu, M. Klos, and K. Thurow, “Application of a dual-arm robot in complex sample preparation and measurement processes,” Journal of Laboratory Automation, vol. 21, no. 5, pp. 671–681, 2016, pMID: 27000132. [Online]. Available: https://doi.org/10.1177/2211068216637352
  11. S. Joshi, X. Chu, H. Fleischer, T. Roddelkopf, M. Klos, and K. Thurow, “Analysis of measurement process design for a dual-arm robot using graphical user interface,” in 2019 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), 2019, pp. 1–6.
  12. A. M. Amen, K. W. Barry, J. M. Boyle, C. E. Brook, S. Choo, L. T. Cornmesser, D. J. Dilworth, J. A. Doudna, A. J. Ehrenberg, I. Fedrigo, S. E. Friedline, T. G. W. Graham, R. Green, J. R. Hamilton, A. Hirsh, M. L. Hochstrasser, D. Hockemeyer, N. Krishnappa, A. Lari, H. Li, E. Lin-Shiao, T. Lu, E. F. Lyons, K. G. Mark, L. A. Martell, A. R. O. Martins, S. L. McDevitt, P. S. Mitchell, E. A. Moehle, C. Naca, D. Nandakumar, E. O’Brien, D. J. Pappas, K. Pestal, D. L. Quach, B. E. Rubin, R. Sachdeva, E. C. Stahl, A. M. Syed, I.-L. Tan, A. L. Tollner, C. A. Tsuchida, C. K. Tsui, T. K. Turkalo, F. D. Urnov, M. B. Warf, O. N. Whitney, L. B. Witkowsky, and IGI Testing Consortium, “Blueprint for a pop-up SARS-CoV-2 testing lab,” Nature Biotechnology, vol. 38, no. 7, pp. 791–797, Jul. 2020, number: 7 Publisher: Nature Publishing Group. [Online]. Available: https://www.nature.com/articles/s41587-020-0583-3
  13. T. Pickles, C. Mustoe, C. Brown, and A. Florence, “Autonomous datafactory : high-throughput screening for large-scale data collection to inform medicine manufacture,” British Journal of Pharmacy, 2022. [Online]. Available: https://doi.org/10.5920/bjpharm.1128
  14. J. Jiang, G. Cao, A. Butterworth, T.-T. Do, and S. Luo, “Where Shall I Touch? Vision-Guided Tactile Poking for Transparent Object Grasping,” IEEE/ASME Transactions on Mechatronics, pp. 1–12, 2022, conference Name: IEEE/ASME Transactions on Mechatronics.
  15. J. Hansen, F. Hogan, D. Rivkin, D. Meger, M. Jenkin, and G. Dudek, “Visuotactile-RL: Learning Multimodal Manipulation Policies with Deep Reinforcement Learning,” in 2022 International Conference on Robotics and Automation (ICRA), May 2022, pp. 8298–8304.
  16. M. S. Nixon and A. S. Aguado, “5 - High-level feature extraction: fixed shape matching,” in Feature Extraction and Image Processing for Computer Vision (Fourth Edition), M. S. Nixon and A. S. Aguado, Eds.   Academic Press, Jan. 2020, pp. 223–290. [Online]. Available: https://www.sciencedirect.com/science/article/pii/B9780128149768000051
  17. S. Suzuki and K. Abe, “Topological structural analysis of digitized binary images by border following,” Computer Vision, Graphics, and Image Processing, vol. 30, no. 1, pp. 32–46, Apr. 1985. [Online]. Available: https://www.sciencedirect.com/science/article/pii/0734189X85900167
  18. Duane Q. Nykamp, “Using Green’s theorem to find area.” [Online]. Available: https://mathinsight.org/greens_theorem_find_area
  19. H. Fang, H.-S. Fang, S. Xu, and C. Lu, “Transcg: A large-scale real-world dataset for transparent object depth completion and a grasping baseline,” IEEE Robotics and Automation Letters, pp. 1–8, 2022.
  20. H. Xu, Y. R. Wang, S. Eppel, A. Aspuru-Guzik, F. Shkurti, and A. Garg, “Seeing glass: Joint point cloud and depth completion for transparent objects,” CoRR, vol. abs/2110.00087, 2021. [Online]. Available: https://arxiv.org/abs/2110.00087
  21. X. Liu, R. Jonschkowski, A. Angelova, and K. Konolige, “KeyPose: Multi-View 3D Labeling and Keypoint Estimation for Transparent Objects,” May 2020, arXiv:1912.02805 [cs]. [Online]. Available: http://arxiv.org/abs/1912.02805
  22. J. Jiang, G. Cao, T.-T. Do, and S. Luo, “A4T: Hierarchical Affordance Detection for Transparent Objects Depth Reconstruction and Manipulation,” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 9826–9833, Oct. 2022, conference Name: IEEE Robotics and Automation Letters.
Citations (3)

Summary

  • The paper presents multi-modal sensing methods that boost vial insertion success from 48.78% with vision-only to 89.55% with force feedback.
  • It employs a dual-stage approach combining the circular Hough Transform and CNN filtering to improve goal position detection in lab automation.
  • The study compares visual-only, force-vision, and tactile-vision approaches, highlighting improved error recovery and enhanced safety in robotic systems.

Multi-modal Sensing for Robotic Insertion Tasks

This paper explores the use of multi-modal sensory feedback to improve the reliability of robotic vial insertion, a common task in R&D laboratories. The authors compare a single-modality vision-based approach with two multi-modal approaches: vision and force feedback, and vision and tactile feedback. The experimental results demonstrate that the multi-modal approaches significantly improve the success rate of vial insertion compared to the visual baseline. The most successful method, combining visual and force feedback, achieves an 89.55% success rate, compared to 48.78% for the vision-only method. The paper highlights the potential of multi-modal sensing to enhance the robustness and safety of robotic systems in laboratory automation.

Goal Position Detection and Filtering

The vial insertion task is divided into two sub-problems: goal position detection and vial insertion. For goal position detection, the authors employ a combination of the circular Hough Transform (CHT) and a CNN classifier. The CHT is used to generate a large number of possible vial placement locations from a top-down image of the workspace. To address the issue of oblique rack views, the CHT detection parameters are set to be overly sensitive, which leads to the detection of many false positives. A CNN classifier is then used to filter these candidates, identifying vacant locations belonging to the target rack. Input data for the CNN is created by scaling the candidate's corresponding radius by a margin factor to 110% of the original and cropping this region from the image, centered on the candidate location. The network then predicts if the cropped image belongs to the rack and, if so, whether the rack slot is occupied. The target is selected from the filtered candidates based on the highest CNN classification score for "In Rack" and "Unoccupied."

Multi-Modal Vial Insertion Methods

Figure 1

Figure 1: Sequential actions of the robot to perform vial insertion, showing the robot detecting the rack, collecting the vial, positioning it above the target, and inserting it.

The second sub-task involves grasping a vial and inserting it into the detected location using one of three modalities:

  • Visual Baseline: This method involves capturing a second image after moving the camera closer to the rack. The image processing from the previous stage is repeated to refine the insertion point estimate. The vial is then aligned with the revised insertion point and moved down the z-axis until it is below the rack height, at which point the gripper release command is sent.
  • Force and Visual Feedback: This method utilizes the robotic arm's internal force sensors to detect contact between the vial and the rack. A FIFO buffer is used to record the static force experienced by the sensor. During insertion, any deviation of more than 20% from the recorded static value causes the robot to stop and assess the vial's state. The position of the vial at the tip of the gripper is then used to determine if the vial has impacted the top surface of the rack. If contact is detected, a search algorithm is initiated.
  • Tactile and Visual Feedback: This method uses a pair of DIGIT visual tactile sensors to provide feedback during insertion. Reference tactile images are captured with the gripper open and not in contact with any object. During the insertion attempt, the absolute difference between the current tactile image and each reference image is calculated. Contact regions are extracted from the resulting difference image, and their center points are used to track the vial's position. Deviation from a neutral state triggers the robot to stop and assess the vial's state.

Search Algorithm

Figure 2

Figure 2: A workflow diagram illustrating the three modalities used: visual, force and vision, and tactile and vision, emphasizing the different methods for assessing vial placement.

A key component of the multi-modal approaches is a search algorithm that allows the robot to recover from failed placement attempts. This algorithm leverages the knowledge of the initial target location and attempts to detect neighboring regions, disregarding occupancy. A bounding box is fitted around the selected target, and the search is conducted within this box. The search creates an envelope around the initial placement position at a distance represented by the search spacing. Trial locations are generated within this envelope, and the robot attempts to insert the vial into these locations. If the vial fails to be inserted after searching the current envelope, it is expanded by the search spacing. This process is repeated until the vial is successfully inserted or the search region is entirely beyond the bounds.

Experimental Results and Analysis

The experimental results, summarized in Table 1, demonstrate the effectiveness of the multi-modal approaches. The visual baseline achieved a success rate of 48.78%, while the force and visual feedback method achieved 89.55%, and the tactile and visual feedback method achieved 69.81%. The force and visual feedback method also exhibited a lower average runtime compared to the visual baseline, despite requiring multiple placement attempts.

The decrease in first-time placement accuracy for the force and visual feedback method, compared to the visual baseline, highlights the effectiveness of the second imaging step in the visual-only modality. However, the multi-modal methods demonstrate a greater capacity for error recovery, as shown in Figure 3, enabling them to achieve higher overall success rates. Figure 3

Figure 3: A bar chart illustrating the distribution of successful placement attempts by the number of attempts required for the multi-modal methods.

The authors note that the tactile feedback method, while improving upon the visual baseline, exhibits a lower success rate than the force feedback method. This is attributed to the Digit sensor's lower gripping force and the smoother contact surface of its silicon membrane, which can lead to the vial moving within the gripper during failed placement attempts. Despite this limitation, the tactile feedback method offers improved safety due to the lower forces exerted on the vial and surrounding surfaces, along with greater insight into the vial's orientation. Figure 4

Figure 4: A graph illustrating the cumulative probability of successful vial insertion across multiple attempts, showing that the multi-modal methods eventually surpass the visual baseline.

Conclusion and Future Directions

The authors conclude that multi-modal sensing offers a promising approach for improving the reliability and safety of robotic vial handling in laboratory automation. The combination of visual and force feedback demonstrates the highest placement reliability, while the inclusion of tactile sensors provides additional insights into the grasped object's orientation, enhancing safety during movement planning. The authors propose several avenues for future research, including exploring the combination of all three modalities and employing machine learning techniques to further improve placement success rates. They also suggest improving the search algorithm by dynamically adjusting the search spacing and accounting for previous search results to compensate for calibration errors.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.