Reinforcement Learning with Parameterized Manipulation Primitives for Robotic Assembly
Abstract: A common theme in robot assembly is the adoption of Manipulation Primitives as the atomic motion to compose assembly strategy, typically in the form of a state machine or a graph. While this approach has shown great performance and robustness in increasingly complex assembly tasks, the state machine has to be engineered manually in most cases. Such hard-coded strategies will fail to handle unexpected situations that are not considered in the design. To address this issue, we propose to find dynamics sequence of manipulation primitives through Reinforcement Learning. Leveraging parameterized manipulation primitives, the proposed method greatly improves both assembly performance and sample efficiency of Reinforcement Learning compared to a previous work using non-parameterized manipulation primitives. In practice, our method achieves good zero-shot sim-to-real performance on high-precision peg insertion tasks with different geometry, clearance, and material.
- U. Thomas, B. Finkemeyer, T. Kroger, and F. Wahl, “Error-tolerant execution of complex robot tasks based on skill primitives,” in 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422), vol. 3, Sep. 2003, pp. 3069–3075 vol.3.
- B. Finkemeyer, T. Kröger, and F. M. Wahl, “Executing assembly tasks specified by manipulation primitive nets,” Advanced Robotics, vol. 19, no. 5, pp. 591–611, Jan. 2005.
- F. Suárez-Ruiz and Q.-C. Pham, “A framework for fine robotic assembly,” in 2016 IEEE International Conference on Robotics and Automation (ICRA), May 2016, pp. 421–426.
- F. Suárez-Ruiz, X. Zhou, and Q.-C. Pham, “Can robots assemble an IKEA chair?” Science Robotics, vol. 3, no. 17, p. eaat6385, Apr. 2018.
- N. Vuong, H. Pham, and Q.-C. Pham, “Learning Sequences of Manipulation Primitives for Robotic Assembly,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), May 2021, pp. 4086–4092.
- T. Kroger, B. Finkemeyer, and F. Wahl, “A task frame formalism for practical implementations,” in IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA ’04. 2004, vol. 5, Apr. 2004, pp. 5218–5223.
- T. Kröger, B. Finkemeyer, and F. M. Wahl, “Manipulation Primitives — A Universal Interface between Sensor-Based Motion Control and Robot Programming,” in Robotic Systems for Handling and Assembly, ser. Springer Tracts in Advanced Robotics, D. Schütz and F. M. Wahl, Eds. Berlin, Heidelberg: Springer, 2011, pp. 293–313.
- L. Johannsmeier, M. Gerchow, and S. Haddadin, “A Framework for Robot Manipulation: Skill Formalism, Meta Learning and Adaptive Control,” in 2019 International Conference on Robotics and Automation (ICRA), May 2019, pp. 5844–5850.
- T. Inoue, G. De Magistris, A. Munawar, T. Yokoya, and R. Tachibana, “Deep reinforcement learning for high precision assembly tasks,” in 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). Vancouver, BC: IEEE, Sep. 2017, pp. 819–825.
- M. Hamaya, R. Lee, K. Tanaka, F. von Drigalski, C. Nakashima, Y. Shibata, and Y. Ijiri, “Learning Robotic Assembly Tasks with Lower Dimensional Systems by Leveraging Physical Softness and Environmental Constraints,” in 2020 IEEE International Conference on Robotics and Automation (ICRA), May 2020, pp. 7747–7753.
- T. Johannink, S. Bahl, A. Nair, J. Luo, A. Kumar, M. Loskyll, J. A. Ojea, E. Solowjow, and S. Levine, “Residual Reinforcement Learning for Robot Control,” in 2019 International Conference on Robotics and Automation (ICRA), May 2019, pp. 6023–6029.
- G. Schoettler, A. Nair, J. A. Ojea, S. Levine, and E. Solowjow, “Meta-Reinforcement Learning for Robotic Industrial Insertion Tasks,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Oct. 2020, pp. 9728–9735.
- Y. Ma, D. Xu, and F. Qin, “Efficient Insertion Control for Precision Assembly Based on Demonstration Learning and Reinforcement Learning,” IEEE Transactions on Industrial Informatics, vol. 17, no. 7, pp. 4492–4502, Jul. 2021.
- T. Davchev, K. S. Luck, M. Burke, F. Meier, S. Schaal, and S. Ramamoorthy, “Residual Learning from Demonstration: Adapting DMPs for Contact-rich Manipulation,” arXiv:2008.07682 [cs], Sep. 2021.
- J. Luo, E. Solowjow, C. Wen, J. A. Ojea, A. M. Agogino, A. Tamar, and P. Abbeel, “Reinforcement Learning on Variable Impedance Controller for High-Precision Robotic Assembly,” in 2019 International Conference on Robotics and Automation (ICRA), May 2019, pp. 3080–3087.
- E. Todorov and W. Li, “A generalized iterative LQG method for locally-optimal feedback control of constrained nonlinear stochastic systems,” in Proceedings of the 2005, American Control Conference, 2005., Jun. 2005, pp. 300–306 vol. 1.
- M. Kaspar, J. D. Muñoz Osorio, and J. Bock, “Sim2Real Transfer for Reinforcement Learning without Dynamics Randomization,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Oct. 2020, pp. 4383–4388.
- J. Tobin, R. Fong, A. Ray, J. Schneider, W. Zaremba, and P. Abbeel, “Domain Randomization for Transferring Deep Neural Networks from Simulation to the Real World,” arXiv:1703.06907 [cs], Mar. 2017.
- X. B. Peng, M. Andrychowicz, W. Zaremba, and P. Abbeel, “Sim-to-real transfer of robotic control with dynamics randomization,” in 2018 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2018, pp. 3803–3810.
- C. C. Beltran-Hernandez, D. Petit, I. G. Ramirez-Alpizar, and K. Harada, “Variable Compliance Control for Robotic Peg-in-Hole Assembly: A Deep-Reinforcement-Learning Approach,” Applied Sciences, vol. 10, no. 19, p. 6923, Jan. 2020.
- R. Martín-Martín, M. A. Lee, R. Gardner, S. Savarese, J. Bohg, and A. Garg, “Variable impedance control in end-effector space: An action space for reinforcement learning in contact-rich tasks,” arXiv preprint arXiv:1906.08880, 2019.
- M. Bogdanovic, M. Khadiv, and L. Righetti, “Learning Variable Impedance Control for Contact Sensitive Tasks,” arXiv:1907.07500 [cs], Jul. 2020.
- R. Chitnis, S. Tulsiani, S. Gupta, and A. Gupta, “Efficient bimanual manipulation using learned task schemas,” in 2020 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2020, pp. 1149–1155.
- M. Sharma, J. Liang, J. Zhao, A. LaGrassa, and O. Kroemer, “Learning to Compose Hierarchical Object-Centric Controllers for Robotic Manipulation,” arXiv:2011.04627 [cs], Nov. 2020.
- S. Nasiriany, H. Liu, and Y. Zhu, “Augmenting Reinforcement Learning with Behavior Primitives for Diverse Manipulation Tasks,” arXiv:2110.03655 [cs], Oct. 2021.
- X. Zhang, S. Jin, C. Wang, X. Zhu, and M. Tomizuka, “Learning Insertion Primitives with Discrete-Continuous Hybrid Action Space for Robotic Assembly Tasks,” Oct. 2021.
- M. Dalal, D. Pathak, and R. Salakhutdinov, “Accelerating Robotic Reinforcement Learning via Parameterized Action Primitives,” arXiv:2110.15360 [cs], Oct. 2021.
- O. Khatib, “A unified approach for motion and force control of robot manipulators: The operational space formulation,” IEEE Journal on Robotics and Automation, vol. 3, no. 1, pp. 43–53, Feb. 1987.
- W. Masson, P. Ranchod, and G. Konidaris, “Reinforcement learning with parameterized actions,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30, 2016.
- C. Yu, Z. Cai, H. Pham, and Q.-C. Pham, “Siamese Convolutional Neural Network for Sub-millimeter-accurate Camera Pose Estimation and Visual Servoing,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Nov. 2019, pp. 935–941.
- J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal Policy Optimization Algorithms,” arXiv:1707.06347 [cs], Aug. 2017.
- E. Todorov, T. Erez, and Y. Tassa, “MuJoCo: A physics engine for model-based control,” in 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Oct. 2012, pp. 5026–5033.
- G. Brockman, V. Cheung, L. Pettersson, J. Schneider, J. Schulman, J. Tang, and W. Zaremba, “OpenAI Gym,” arXiv:1606.01540 [cs], Jun. 2016.
- N. Hansen and A. Ostermeier, “Completely Derandomized Self-Adaptation in Evolution Strategies,” Evolutionary Computation, vol. 9, pp. 159–195, Jun. 2001.
- H. Pham and Q.-C. Pham, “Convex Controller Synthesis for Robot Contact,” arXiv:1909.04313 [cs], Jan. 2020.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.