Force-EvT: A Closer Look at Robotic Gripper Force Measurement with Event-based Vision Transformer
Abstract: Robotic grippers are receiving increasing attention in various industries as essential components of robots for interacting and manipulating objects. While significant progress has been made in the past, conventional rigid grippers still have limitations in handling irregular objects and can damage fragile objects. We have shown that soft grippers offer deformability to adapt to a variety of object shapes and maximize object protection. At the same time, dynamic vision sensors (e.g., event-based cameras) are capable of capturing small changes in brightness and streaming them asynchronously as events, unlike RGB cameras, which do not perform well in low-light and fast-moving environments. In this paper, a dynamic-vision-based algorithm is proposed to measure the force applied to the gripper. In particular, we first set up a DVXplorer Lite series event camera to capture twenty-five sets of event data. Second, motivated by the impressive performance of the Vision Transformer (ViT) algorithm in dense image prediction tasks, we propose a new approach that demonstrates the potential for real-time force estimation and meets the requirements of real-world scenarios. We extensively evaluate the proposed algorithm on a wide range of scenarios and settings, and show that it consistently outperforms recent approaches.
- Dynamic-vision-based force measurements using convolutional recurrent neural networks. Sensors, 20(16):4469, 2020.
- Simultaneous optical flow and intensity estimation from an event camera. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 884–892, 2016.
- An r-squared measure of goodness of fit for some common nonlinear regression models. Journal of econometrics, 77(2):329–342, 1997.
- T. Chai and R. R. Draxler. Root mean square error (rmse) or mean absolute error (mae)?–arguments against avoiding rmse in the literature. Geoscientific model development, 7(3):1247–1250, 2014.
- Vision transformer adapter for dense predictions. arXiv preprint arXiv:2205.08534, 2022.
- J. Dai and D. Kerr. A six-component contact force measurement device based on the stewart platform. Proceedings of the Institution of Mechanical Engineers, Part C: Journal of Mechanical Engineering Science, 214(5):687–697, 2000.
- Addressing unintended bias in toxicity detection: An lstm and attention-based approach. In 2023 5th International Conference on Artificial Intelligence and Computer Applications (ICAICA), pp. 375–379. IEEE, 2023.
- A review: State of the art of robotic grippers. Int Res J Eng Technol, 5(5):371–375, 2018.
- An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- Machine learning based deflection prediction and inverse design for discrete variable stiffness units. In International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, vol. 87363, p. V008T08A043. American Society of Mechanical Engineers, 2023.
- A variable stiffness robotic gripper based on parallel beam with vision-based force sensing for flexible grasping. Robotica, pp. 1–19, 2023.
- Event-based vision: A survey. IEEE transactions on pattern analysis and machine intelligence, 44(1):154–180, 2020.
- Q. Guo and J. Wen. Multi-level fusion based deep convolutional network for image quality assessment. In Pattern Recognition. ICPR International Workshops and Challenges: Virtual Event, January 10–15, 2021, Proceedings, Part VI, pp. 670–678. Springer, 2021.
- Dexterous anthropomorphic robot hand with distributed tactile sensor: Gifu hand ii. IEEE/ASME transactions on mechatronics, 7(3):296–303, 2002.
- Design and fabrication of a novel force sensor for robot grippers. IEEE Sensors Journal, 18(4):1410–1418, 2017.
- A survey of robots in healthcare. Technologies, 9(1):8, 2021.
- T. Lien. Gripper technologies for food industry robots. In Robotics and automation in the food industry, pp. 143–170. Elsevier, 2013.
- Y. Lu and G. Lu. Superthermal: Matching thermal as visible through thermal feature exploration. IEEE Robotics and Automation Letters, 6(2):2690–2697, 2021.
- Lidar-forest dataset: Lidar point cloud simulation dataset for forestry application. arXiv preprint arXiv:2402.04546, 2024.
- Transflow: Transformer as flow learner. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 18063–18073, 2023.
- Event-based vision meets deep learning on steering prediction for self-driving cars. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5419–5427, 2018.
- Moving object detection for event-based vision using graph spectral clustering. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 876–884, 2021.
- Neuromorphic event-based slip detection and suppression in robotic grasping and manipulation. IEEE Access, 8:153364–153384, 2020.
- A novel dynamic-vision-based approach for tactile sensing applications. IEEE Transactions on Instrumentation and Measurement, 69(5):1881–1893, 2019.
- Learning to detect objects with a 1 megapixel event camera. Advances in Neural Information Processing Systems, 33:16639–16652, 2020.
- A vision transformer approach for traffic congestion prediction in urban areas. IEEE Transactions on Intelligent Transportation Systems, 24(4):3922–3934, 2023.
- Long-term object tracking with a moving event camera. In Bmvc, p. 241, 2018.
- Vision transformers for dense prediction. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 12179–12188, 2021.
- Emvs: Event-based multi-view stereo—3d reconstruction with an event camera in real-time. International Journal of Computer Vision, 126(12):1394–1414, 2018.
- Image quality assessment through fsim, ssim, mse and psnr—a comparative study. Journal of Computer and Communications, 7(3):8–18, 2019.
- Fast image reconstruction with an event camera. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 156–163, 2020.
- Rethinking transformer-based set prediction for object detection. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 3611–3620, 2021.
- Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 568–578, 2021.
- Enhancing liver segmentation: A deep learning approach with eas feature extraction and multi-scale fusion. International Journal of Innovative Research in Computer Science & Technology, 12(1):26–34, 2024.
- J. Wen and Q. Guo. When distortion meets perceptual quality: A multi-task learning pipeline. In PRICAI 2021: Trends in Artificial Intelligence: 18th Pacific Rim International Conference on Artificial Intelligence, PRICAI 2021, Hanoi, Vietnam, November 8–12, 2021, Proceedings, Part III 18, pp. 353–365. Springer, 2021.
- Event-based video reconstruction using transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2563–2572, 2021.
- A compliant adaptive gripper and its intrinsic force sensing method. IEEE Transactions on Robotics, 37(5):1584–1603, 2021.
- Robot tactile sensing: Vision based tactile sensor for force perception. In 2018 IEEE 8th Annual International Conference on CYBER Technology in Automation, Control, and Intelligent Systems (CYBER), pp. 1360–1365. IEEE, 2018.
- Deep learning for event-based vision: A comprehensive survey and benchmarks. arXiv preprint arXiv:2302.08890, 2023.
- Flexure-based variable stiffness gripper for large-scale grasping force regulation with vision. In Intelligent Robotics and Applications: 12th International Conference, ICIRA 2019, Shenyang, China, August 8–11, 2019, Proceedings, Part I 12, pp. 346–357. Springer, 2019.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.