Submitted Successfully!
To reward your contribution, here is a gift for you: A free trial for our video production service.
Thank you for your contribution! You can also upload a video entry or images related to this topic.
Version Summary Created by Modification Content Size Created at Operation
1 -- 1666 2024-01-04 08:56:33 |
2 Reference format revised. Meta information modification 1666 2024-01-05 06:32:48 |

Video Upload Options

Do you have a full video?

Confirm

Are you sure to Delete?
Cite
If you have any further questions, please contact Encyclopedia Editorial Office.
Kana, S.; Gurnani, J.; Ramanathan, V.; Ariffin, M.Z.; Turlapati, S.H.; Campolo, D. Learning Compliant Box-in-Box Insertion through Haptic-Based Robotic Teleoperation. Encyclopedia. Available online: https://encyclopedia.pub/entry/53413 (accessed on 29 April 2024).
Kana S, Gurnani J, Ramanathan V, Ariffin MZ, Turlapati SH, Campolo D. Learning Compliant Box-in-Box Insertion through Haptic-Based Robotic Teleoperation. Encyclopedia. Available at: https://encyclopedia.pub/entry/53413. Accessed April 29, 2024.
Kana, Sreekanth, Juhi Gurnani, Vishal Ramanathan, Mohammad Zaidi Ariffin, Sri Harsha Turlapati, Domenico Campolo. "Learning Compliant Box-in-Box Insertion through Haptic-Based Robotic Teleoperation" Encyclopedia, https://encyclopedia.pub/entry/53413 (accessed April 29, 2024).
Kana, S., Gurnani, J., Ramanathan, V., Ariffin, M.Z., Turlapati, S.H., & Campolo, D. (2024, January 04). Learning Compliant Box-in-Box Insertion through Haptic-Based Robotic Teleoperation. In Encyclopedia. https://encyclopedia.pub/entry/53413
Kana, Sreekanth, et al. "Learning Compliant Box-in-Box Insertion through Haptic-Based Robotic Teleoperation." Encyclopedia. Web. 04 January, 2024.
Learning Compliant Box-in-Box Insertion through Haptic-Based Robotic Teleoperation
Edit

In modern logistics, the box-in-box insertion task is representative of a wide range of packaging applications, and automating compliant object insertion is difficult due to challenges in modelling the object deformation during insertion. Using Learning from Demonstration (LfD) paradigms, which are frequently used in robotics to facilitate skill transfer from humans to robots, can be one solution for complex tasks that are difficult to mathematically model. 

box-in-box insertion compliant insertion Learning from Demonstration

1. Introduction

Modern logistics require box-in-box insertion for a broad range of packaging tasks involving compliant, cardboard boxes. Industries such as electronics and cosmetics heavily rely on such box packaging. Typically, smaller boxes containing a product are stacked and placed into larger boxes for transportation. However, these tasks are often carried out manually on the factory floor, leading to drawbacks such as fatigue, limited operational time, and time consumption due to the repetitive and monotonous nature of the work. To overcome these challenges, there is a need for robot automation solutions to perform these tasks. Given that most industrial products are packaged in cuboid-shaped boxes, the problem at hand can be described as ‘developing an autonomous system that can perform compliant box-in-box insertion, wherein a cuboid box made of a flexible material is inserted into another larger box made of a similar deformable material’ (see Figure 1).
Figure 1. Example of a compliant box-in-box insertion where a folding carton (box B) containing an electric light bulb is to be inserted into a receptacle folding carton (box A). Manual insertion performed by a human (bottom).
Nonetheless, autonomous insertion is a heavily researched field in robotics, especially in the classical peg-in-hole assembly problem [1]. The existing approaches fall into contact model-based and contact model-free approaches [2]. While the former covers techniques based on the contact state modelling between the parts, the latter focuses on learning paradigms such as Learning from Demonstration and Reinforcement Learning. Contact state modelling is known to be sensitive to uncertainties such as the elasticity of the system. A model-free strategy like Learning from Demonstration looks more promising given that the task in question requires compliant objects because humans’ superior sensory systems and decision-making skills enable them to accomplish insertion tasks even under uncertain environments.
While various research has attempted automated assembly in the past, the focus has frequently been on peg-in-hole insertion dealing with rigid objects because it is the basis of many assembly operations. To avoid the damage or wear of such objects, however, some compliance has been introduced through the use of force-guided robotic systems [3] combined with various other techniques, such as the use of machine learning [4] and vision [5], for example. Another simple approach used to safely manipulate objects with ease is to operate the manipulator in velocity control mode [6], which can then be translated to joint torques using the computed torque control method, as performed in [7], to introduce the desired compliance during manipulation. While several works take advantage of active interaction controls such as impedance/admittance control [8], compliance control [9], and hybrid force/position control [10], some other works also focus on utilising grasp planning [11] and soft grippers [12][13] to mimic the compliance of human fingers to reduce the effects of localisation uncertainties [7].

2. Learning from Demonstration and Teleoperation

Learning from Demonstration (LfD) is a policy learning approach that involves a human demonstrator performing/demonstrating a task that is to be imitated by a robot [14]. This technique ends the need for a non-expert robot user to learn how to manually program for a task, as maybe required by other assembly task implementations such as in [15]. Learning from Demonstration can be implemented on robots in several ways, of which the most common ones are kinesthetic teaching [16], teleoperation [17][18], vision-sensing [19][20][21], and the use of wearable sensors.
The datasets gathered from these demonstrations consist of task execution skills acquired from humans. These skills are then extricated (learnt) with the help of different techniques for the task to be performed by the robot [16]. One such commonly used technique is the Hidden Markov Model (HMM) [22][23], which is a robust probabilistic method to encode the spatial and temporal variabilities of human motion across various demonstrations [24] as a sequence of states. These states are defined as separate Gaussian Mixture Models (GMM) to explain the input data [25]. Ref. [23] compared the use of the HMM and Gaussian Mixture Regressions (GMR) approach vs. another, more popular Dynamic Movement Primitive (DMP) technique to allow robots to acquire skills through imitation. DMPs are basically units of action in terms of attractor dynamics of non-linear differential equations that encode a desired movement trajectory [26]. This method allows one to learn a control policy for a task from the demonstrations provided. However, standard DMP learning is prone to existing noise in human demonstrations [27].
It was also concluded by [23] that in the context of separated learning and reproduction, the HMM was more systematic in generalising motion than DMP. It allowed the demonstrator to provide partial demonstrations for a specific segment of the task instead of repeating the whole task again. This is an important feature when the aim is to refine one part of the movement.
Another study by [28], on the robotic assembly of mixed deformable and rigid objects, leveraged the use of haptic feedback with position and velocity controlled robots to make them compliant without explicitly engaging joint torque control. This information was integrated into a reinforcement learning (RL) framework as the insertion hole was smaller than the peg in diameter and was deformable. Thus, the contact mechanics for it were unknown, making it difficult to design a feedback control law. However, the lack of a vision system in their setup caused the insertion to sometimes fail. Another limitation of this method was that it only worked when the peg was relatively close to the hole. Research about automatic control helps with designing adaptive control [29][30] in cases where the peg position is different. Several other works utilising RL to adapt to varying dynamics of the environment during contact-rich robotic manipulation of non-rigid, deformable objects were reviewed by [31].
While the HMM and RL are widely used techniques for encoding task trajectories. RL requires a large dataset of demonstrations for training, which is an issue when there are time constraints, and the HMM method works by interpolating between discrete sets which are unsuitable for the continuous process of encoding tasks [32]. Also, the HMM relies on the proper choice of gains for stability, which requires estimating perturbations and the range of initial positions that the system can handle in advance, which can lead to inaccuracies [23].
The GMM is a probabilistic method that is simple and robust enough for skill learning at the trajectory level. It enables the extraction of all trajectory constraints [16][21]. A combination of the GMM and GMR has been used to encode and generalise trajectories in previous independent works, such as [21][22][33]. The work by [22] presented a Learning by Demonstration framework for extracting different features of various manipulation tasks taught to a humanoid robot via kinesthetic teaching. The collected data were first projected onto a latent space for dimensionality reduction through the Principal Component Analysis (PCA) technique. Then, the signals were temporally aligned using the Dynamic Time Warping (DTW) approach for probabilistic encoding in GMMs. Ref. [21] made use of GMMs to encode upper-body gesture learning through the use of vision sensors and generalised the data through GMR.
While the GMM has been used to encode the training data for box-in-box insertion tasks, the proposed framework introduces a novel generalisation method based on GMR in Barycentric coordinates. Barycentric coordinates are commonly used in computer graphics and computational mechanics to represent a point inside a simplex as an affine combination of all of its vertices. This technique can be generalised to arbitrary polytopes [34] and is implemented in the generalisation approach of the training datasets. GMR is used to learn the robot trajectories that were demonstrated to the robot by the human operator and encoded by Gaussians. Regression makes it possible to retrieve a smooth signal analytically [22].
To perform human demonstrations of contact tasks like box-in-box assembly, it is important to consider how the task forces are being recorded and, more importantly, separated from the human forces acting on the system. Traditional kinesthetic teaching [35], i.e., by a human directly interacting with a robot and physically guiding it through the contact task, would result in the recording of the human forces coupled with the contact task forces. It is difficult to isolate the contact task forces only to perform force control later when generalising. One way to achieve this is to use two robots coupled through teleoperation control [36] so that the human interacts with the master robot and the slave robot deals with the contact task forces. Here, the term ‘Master-Slave’ simply means that the slave robot follows the motion of the master robot in the joint space. That is, to perform a task in a remote fashion, the human operator physically guides the master robot, thereby controlling the action of the slave robot which is in direct interaction with the environment. In fact, one of the first applications of teleoperation control was manipulation, specifically to handle radioactive materials from a safe distance in nuclear research [37].
For contact-rich tasks such as tight tolerance insertion, haptic feedback improves performance in the manipulation of remote objects [38]. Sensing in force-based teleoperation can be implemented through two possibilities: (a) a force/torque sensor attached to the flange of the robot to sense the end-effector forces and (b) a torque sensor built into every manipulator joint itself [39]. The latter approach was preferred due to the expensive nature of standalone six-axis F/T sensors. Instead, the robot construction was performed with strain gauges at every joint of an industrial robot manipulator [40]. Instead of using explicit force-control to track the contact task haptics, the teleoperation’s impedance characteristics are used to capture the forces that the slave should exert on the environment. This is inherent in the master kinematics and is a key aspect of the proposed approach, which allows for an open-loop playback of the assembly demonstrations given the task conditions do not change.

References

  1. Whitney, D.E. Mechanical Assemblies: Their Design, Manufacture, and Role in Product Development; Oxford University Press: New York, NY, USA, 2004; Volume 1.
  2. Xu, J.; Hou, Z.; Liu, Z.; Qiao, H. Compare contact model-based control and contact model-free learning: A survey of robotic peg-in-hole assembly strategies. arXiv 2019, arXiv:1904.05240.
  3. Chin, K.S.; Ratnam, M.M.; Mandava, R. Force-guided robot in automated assembly of mobile phone. Assem. Autom. 2003, 23, 75–86.
  4. Li, X.; Xiao, J.; Zhao, W.; Liu, H.; Wang, G. Multiple peg-in-hole compliant assembly based on a learning-accelerated deep deterministic policy gradient strategy. Ind. Robot. Int. J. Robot. Res. Appl. 2021, 49, 54–64.
  5. Song, R.; Li, F.; Fu, T.; Zhao, J. A robotic automatic assembly system based on vision. Appl. Sci. 2020, 10, 1157.
  6. Navarro-Alarcon, D.; Liu, Y.H.; Romero, J.G.; Li, P. On the visual deformation servoing of compliant objects: Uncalibrated control methods and experiments. Int. J. Robot. Res. 2014, 33, 1462–1480.
  7. Kazemi, M.; Valois, J.S.; Bagnell, J.A.; Pollard, N. Human-inspired force compliant grasping primitives. Auton. Robot. 2014, 37, 209–225.
  8. Hogan, N. The mechanics of multi-joint posture and movement control. Biol. Cybern. 1985, 52, 315–331.
  9. Mason, M.T. Compliance and force control for computer controlled manipulators. IEEE Trans. Syst. Man Cybern. 1981, 11, 418–432.
  10. Xing, D.; Liu, X.; Liu, F.; Xu, D. Efficient insertion strategy for precision assembly with uncertainties using a passive mechanism. IEEE Trans. Ind. Inform. 2020, 17, 1263–1273.
  11. Wermelinger, M.; Johns, R.; Gramazio, F.; Kohler, M.; Hutter, M. Grasping and Object Reorientation for Autonomous Construction of Stone Structures. IEEE Robot. Autom. Lett. 2021, 6, 5105–5112.
  12. Wang, T.; Joo, H.J.; Song, S.; Hu, W.; Keplinger, C.; Sitti, M. A versatile jellyfish-like robotic platform for effective underwater propulsion and manipulation. Sci. Adv. 2023, 9, eadg0292.
  13. Zhai, Y.; Boer, A.D.; Yan, J.; Shih, B.; Faber, M.; Speros, J.; Gupta, R.; Tolley, M.T. Desktop fabrication of monolithic soft robotic devices with embedded fluidic control circuits. Sci. Robot. 2023, 8, adg3792.
  14. Argall, B.D.; Chernova, S.; Veloso, M.; Browning, B. A survey of robot learning from demonstration. Robot. Auton. Syst. 2009, 57, 469–483.
  15. Polverini, M.P.; Zanchettin, A.M.; Rocco, P. A constraint-based programming approach for robotic assembly skills implementation. Robot. Comput. Integr. Manuf. 2019, 59, 69–81.
  16. Kyrarini, M.; Haseeb, M.A.; Ristić-Durrant, D.; Gräser, A. Robot learning of industrial assembly task via human demonstrations. Auton. Robot. 2019, 43, 239–257.
  17. Havoutis, I.; Calinon, S. Learning assistive teleoperation behaviors from demonstration. In Proceedings of the 2016 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), Lausanne, Switzerland, 23–27 October 2016; pp. 258–263.
  18. Pervez, A.; Ali, A.; Ryu, J.H.; Lee, D. Novel learning from demonstration approach for repetitive teleoperation tasks. In Proceedings of the 2017 IEEE World Haptics Conference (WHC), Munich, Germany, 6–9 June 2017; pp. 60–65.
  19. Huang, B.; Ye, M.; Lee, S.L.; Yang, G.Z. A Vision-Guided Multi-Robot Cooperation Framework for Learning-by-Demonstration and Task Reproduction. 2017. Available online: http://xxx.lanl.gov/abs/1706.00508 (accessed on 29 September 2023).
  20. Lin, L.; Yang, Y.; Song, Y.; Nemec, B.; Ude, A.; Rytz, J.; Buch, A.; Krüger, N.; Savarimuthu, T. Peg-in-Hole assembly under uncertain pose estimation. In Proceedings of the Proceeding of the 11th World Congress on Intelligent Control and Automation, Shenyang, China, 29 June–4 July 2014; pp. 2842–2847.
  21. Sabbaghi, E.; Bahrami, M.; Ghidary, S.S. Learning of gestures by imitation using a monocular vision system on a humanoid robot. In Proceedings of the 2014 Second RSI/ISM International Conference on Robotics and Mechatronics (ICRoM), Tehran, Iran, 15–17 October 2014; pp. 588–594.
  22. Calinon, S.; Guenter, F.; Billard, A. On learning, representing, and generalizing a task in a humanoid robot. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2007, 37, 286–298.
  23. Calinon, S.; Sauser, E.L.; Billard, A.G.; Caldwell, D.G. Evaluation of a probabilistic approach to learn and reproduce gestures by imitation. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK, USA, 3–7 May 2010; pp. 2671–2676.
  24. Lee, S.H.; Suh, I.H.; Calinon, S.; Johansson, R. Autonomous framework for segmenting robot trajectories of manipulation task. Auton. Robot. 2015, 38, 107–141.
  25. Brown, J.C.; Smaragdis, P. Hidden Markov and Gaussian mixture models for automatic call classification. J. Acoust. Soc. Am. 2009, 125, EL221–EL224.
  26. Meier, F.; Schaal, S. A probabilistic representation for dynamic movement primitives. arXiv 2016, arXiv:1612.05932.
  27. Ghalamzan E, A.M.; Paxton, C.; Hager, G.D.; Bascetta, L. An incremental approach to learning generalizable robot tasks from human demonstration. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 5616–5621.
  28. Luo, J.; Solowjow, E.; Wen, C.; Ojea, J.A.; Agogino, A.M. Deep reinforcement learning for robotic assembly of mixed deformable and rigid objects. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 2062–2069.
  29. Chi, R.; Li, H.; Shen, D.; Hou, Z.; Huang, B. Enhanced P-Type Control: Indirect Adaptive Learning From Set-Point Updates. IEEE Trans. Autom. Control 2023, 68, 1600–1613.
  30. Roman, R.C.; Precup, R.E.; Petriu, E.M. Hybrid data-driven fuzzy active disturbance rejection control for tower crane systems. Eur. J. Control 2021, 58, 373–387.
  31. Elguea-Aguinaco, Í.; Serrano-Muñoz, A.; Chrysostomou, D.; Inziarte-Hidalgo, I.; Bøgh, S.; Arana-Arexolaleiba, N. A review on reinforcement learning for contact-rich robotic manipulation tasks. Robot. Comput.-Integr. Manuf. 2023, 81, 102517.
  32. Luo, J.; Yang, C.; Li, Q.; Wang, M. A Task Learning Mechanism for the Telerobots. Int. J. Humanoid Robot. 2019, 16, 1950009.
  33. Calinon, S.; Billard, A. A framework integrating statistical and social cues to teach a humanoid robot new skills. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Workshop on Social Interaction with Intelligent Indoor Robots, Pasadena, CA, USA, 23–27 May 2008. number CONF.
  34. Hormann, K.; Sukumar, N. Generalized Barycentric Coordinates in Computer Graphics and Computational Mechanics; CRC Press: Boca Raton, FL, USA, 2017.
  35. Caccavale, R.; Saveriano, M.; Finzi, A.; Lee, D. Kinesthetic teaching and attentional supervision of structured tasks in human–robot interaction. Auton. Robot. 2019, 43, 1291–1307.
  36. Kuchenbecker, K.J.; Niemeyer, G. Induced master motion in force-reflecting teleoperation. J. Dyn. Sys. Meas. Control 2006, 128, 800–810.
  37. Niemeyer, G.; Preusche, C.; Stramigioli, S.; Lee, D. Telerobotics. In Springer Handbook of Robotics; Springer: Berlin/Heidelberg, Germany, 2016; pp. 1085–1108.
  38. Das, H.; Zak, H.; Kim, W.S.; Bejczy, A.K.; Schenker, P.S. Operator performance with alternative manual control modes in teleoperation. Presence Teleoperators Virtual Environ. 1992, 1, 201–218.
  39. Siciliano, B.; Sciavicco, L.; Villani, L.; Oriolo, G. Force Control; Springer: Berlin/Heidelberg, Germany, 2009.
  40. Luh, J.; Fisher, W.; Paul, R. Joint torque control by a direct feedback for industrial robots. IEEE Trans. Autom. Control 1983, 28, 153–161.
More
Information
Contributors MDPI registered users' name will be linked to their SciProfiles pages. To register with us, please refer to https://encyclopedia.pub/register : , , , , ,
View Times: 112
Revisions: 2 times (View History)
Update Date: 05 Jan 2024
1000/1000