Next Article in Journal / Special Issue
Trajectory Optimization of a Redundant Serial Robot Using Cartesian via Points and Kinematic Decoupling
Previous Article in Journal
On Using Inertial Measurement Units for Solving the Direct Kinematics Problem of Parallel Mechanisms
Previous Article in Special Issue
Quantifying Age-Related Differences of Ankle Mechanical Properties Using a Robotic Device

Robotics 2019, 8(4), 100;

Human–Robot Collaboration in Manufacturing Applications: A Review
Department of Mechanical Engineering, Imperial College London, London SW7 2AZ, UK
Department of Industrial Engineering, University of Padova, 35131 Padova, Italy
Department of Management and Engineering, University of Padova, 36100 Vicenza, Italy
Author to whom correspondence should be addressed.
Received: 30 September 2019 / Accepted: 3 December 2019 / Published: 6 December 2019


This paper provides an overview of collaborative robotics towards manufacturing applications. Over the last decade, the market has seen the introduction of a new category of robots—collaborative robots (or “cobots”)—designed to physically interact with humans in a shared environment, without the typical barriers or protective cages used in traditional robotics systems. Their potential is undisputed, especially regarding their flexible ability to make simple, quick, and cheap layout changes; however, it is necessary to have adequate knowledge of their correct uses and characteristics to obtain the advantages of this form of robotics, which can be a barrier for industry uptake. The paper starts with an introduction of human–robot collaboration, presenting the related standards and modes of operation. An extensive literature review of works published in this area is undertaken, with particular attention to the main industrial cases of application. The paper concludes with an analysis of the future trends in human–robot collaboration as determined by the authors.
collaborative robotics; cobots; human–robot interaction; literature review

1. Introduction

Traditional industrial robotic systems require heavy fence guarding and peripheral safety equipment that reduce flexibility while increasing costs and required space. The current market, however, asks for reduced lead times and mass customization, thus imposing flexible and multi-purpose assembly systems [1]. These needs are particularly common for small- and medium-sized enterprises (SMEs). Collaborative robots (or cobots [2]) represent a natural evolution that can solve existing challenges in manufacturing and assembly tasks, as they allow for a physical interaction with humans in a shared workspace; moreover, they are designed to be easily reprogrammed even by non-experts in order to be repurposed for different roles in a continuously evolving workflow [3]. Collaboration between humans and cobots is seen as a promising way to achieve increases in productivity while decreasing production costs, as it combines the ability of a human to judge, react, and plan with the repeatability and strength of a robot.
Several years have passed since the introduction of collaborative robots in industry, and cobots have now been applied in several different applications; furthermore, collaboration with traditional robots is considered in research, as it takes advantage of the devices’ power and performance. Therefore, we believe that it is the proper time to review the state of the art in this area, with a particular focus on industrial case studies and the economic convenience of these systems. A literature review is considered a suitable approach to identify the modern approaches towards Human–Robot Collaboration (HRC), in order to better understand the capabilities of the collaborative systems and highlight the possible existing gap on the basis of the presented future works.
The paper is organized as follows: After a brief overview of HRC methods, Section 2 provides an overview of the economic advantages of the collaborative systems, with a brief comparison with traditional systems. Our literature review analysis is presented in Section 3, and Section 4 contains a discussion of the collected data. Lastly, Section 5 concludes the work.


Despite their relatively recent spread, the concept of cobots was invented in 1996 by J. Edward Colgate and Michael Pashkin [2,4]. These devices were passive and operated by humans, and are quite different from modern cobots that are more represented by the likes of lightweight robots such as KUKA LBR iiwa, developed since the 1990s by KUKA Roboter GmbH and the Institute of Robotics and Mechatronics at the German Aerospace Center (DLR) [5], or the first commercial collaborative robot sold in 2008, which was a UR5 model produced by the Danish company Universal Robots [6].
First of all, we believe that it is important to distinguish the different ways of collaboration, since the term collaboration often generates misunderstandings in its definition.
Müller et al. [7] proposed a classification for the different methodologies in which humans and cobots can work together, as summarized in Figure 1, where the final state shows a collaborative environment.
  • Coexistence, when the human operator and cobot are in the same environment but generally do not interact with each other.
  • Synchronised, when the human operator and cobot work in the same workspace, but at different times.
  • Cooperation, when the human operator and cobot work in the same workspace at the same time, though each focuses on separate tasks.
  • Collaboration, when the human operator and the cobot must execute a task together; the action of the one has immediate consequences on the other, thanks to special sensors and vision systems.
It should be noted that neither this classification nor the terminology used are unique, and others may be found in the literature [8,9,10,11].
To provide definitions and guidelines for the safe and practical use of cobots in industry, several standards have been proposed. Collaborative applications are part of the general scope of machinery safety regulated by the Machinery Directive, which defines the RESS (Essential Health and Safety Requirements). For further documentation, we refer to [12].
The reference standards as reported in the Machinery Directive are:
  • UNI EN ISO 12100:2010 “Machine safety, general design principles, risk assessment, and risk reduction”.
  • UNI EN ISO 10218-2:2011 “Robots and equipment for robots, Safety requirements for industrial robots, Part 2: Systems and integration of robots”.
  • UNI EN ISO 10218-1:2012 “Robots and equipment for robots, Safety requirements for industrial robots, Part 1: Robots”.
In an international setting, the technical specification ISO/TS 15066:2016 “Robots and robotic devices, Collaborative Robots” is dedicated to the safety requirements of the collaborative methods envisaged by the Technical Standard UNI EN ISO 10218-2:2011.
According to the international standard UNI EN ISO 10218 1 and 2, and more widely explained in ISO/TS 15066:2016, four classes of safety requirements are defined for collaborative robots:
  • Safety-rated monitored stop (SMS) is used to cease robot motion in the collaborative workspace before an operator enters the collaborative workspace to interact with the robot system and complete a task. This mode is typically used when the cobot mostly works alone, but occasionally a human operator can enter its workspace.
  • Hand-guiding (HG), where an operator uses a hand-operated device, located at or near the robot end-effector, to transmit motion commands to the robot system.
  • Speed and separation monitoring (SSM), where the robot system and operator may move concurrently in the collaborative workspace. Risk reduction is achieved by maintaining at least the protective separation distance between operator and robot at all times. During robot motion, the robot system never gets closer to the operator than the protective separation distance. When the separation distance decreases to a value below the protective separation distance, the robot system stops. When the operator moves away from the robot system, the robot system can resume motion automatically according to the requirements of this clause. When the robot system reduces its speed, the protective separation distance decreases correspondingly.
  • Power and force limiting (PFL), where the robot system shall be designed to adequately reduce risks to an operator by not exceeding the applicable threshold limit values for quasi-static and transient contacts, as defined by the risk assessment.
Collaborative modes can be adopted even when using traditional industrial robots; however, several safety devices, e.g., laser sensors and vision systems, or controller alterations are required. Thus, a commercial cobot that does not require further hardware costs and setup can be a more attractive solution for industry.
Lastly, cobots are designed with particular features that distinguish them considerably from traditional robots, defined by Michalos et al. [13] as technological and ergonomic requirements. Furthermore, they should be equipped with additional features with respect to traditional robots, such as force and torque sensors, force limits, vision systems (cameras), laser systems, anti-collision systems, recognition of voice commands, and/or systems to coordinate the actions of human operators with their motion. For a more complete overview, we refer to [8,13]. Table A1 shows the characteristics of some of the most popular cobots, with a brief overview of some kinematic schemes in Table A2.

2. Convenience of Collaborative Robotics

The choice towards human–robot collaborative systems is mainly dictated by economic motivations, occupational health (ergonomics and human factors), and efficient use of factory space. Another advantage is the simplification in the robot programming for the actions necessary to perform a task [14]. In addition, learning by demonstration is a popular feature [15].
Furthermore, the greater convenience of collaborative systems is their flexibility: Theoretically, since collaborative cells do not require rigid safety systems, they could be allocated in other parts of plants more easily and more quickly; therefore, they could adapt well to those cases in which the production layout needs to change continuously [16]. However, it should be noted that high-risk applications have to be constrained as in any other traditional system, thus restricting the flexibility.
Collaborative systems can also achieve lower direct unit production costs: [17] observed that a higher degree of collaboration, called c % , has a high impact on throughput; moreover, depending on the assembly process considered, the throughput can be higher than in traditional systems.
Table 1 provides a comparison between collaborative and traditional systems for four different jobs: assembly (the act of attaching two or more components), placement (the act of positioning each part in the proper position), handling (the manipulation of the picked part), and picking (the act of taking from the feeding point). In order to adapt to market needs, a manual assembly system could be used, though this can lead to a decrease in productivity due to variations in quality and fluctuations in labor rates [18]. Comparing the human operator capabilities to automated systems, it is clear that the performance of manual assembly is greatly influenced by ergonomic factors, which restrict the product weight and the accuracy of the human operator [19]. Therefore, these restrictions limit the capabilities of human operators in the handling and picking tasks of heavy/bulky parts. These components can be manipulated with handling systems such as jib cranes: These devices could be considered as large workspace-serving robots [20], used for automated transportation of heavy parts. However, to the authors’ knowledge, there are no commercial end-effectors that allow these systems to carry out complex tasks, such as assembly or precise placing, since they are quite limited in terms of efficiency and precision [21].
Traditional robotic systems [22] bridge the presented gap, presenting manipulators with both high payload (e.g., FANUC M-2000 series with a payload of 2.3 t [23]) and high repeatability. However, the flexibility and dexterity required for complex assembly tasks could be too expensive, or even impossible, to achieve with traditional robotic systems [24]. This gap can be closed by collaborative systems, since they combine the capabilities of a traditional robot with the dexterity and flexibility of the human operator. Collaborative robots are especially advantageous for assembly tasks, particularly if the task is executed with a human operator. They are also suitable for pick and place applications, though the adoption of a traditional robot or a handling system can offer better results in terms of speed, precision, and payload.

3. Literature Review

This literature review analyses works from 2009–2018 that involved collaborative robots for manufacturing or assembly tasks. Reviewed papers needed to include a practical experiment involving a collaborative robot undertaking a manufacturing or assembly task; we ignored those that only considered the task in simulation. This criterion was implemented as, often, only practical experiments with real hardware can highlight both the challenges and advantages of cobots.
For this literature review, three search engines were used to collect papers over our time period that were selected using the following boolean string: ((collaborative AND robot) OR cobot OR cobotics) AND (manufacturing OR assembly). Our time period of 2009–2018 was chosen as the timeline for this literature review, as it is only in the last 10 years that we have seen the availability of collaborative robots in the market.
  • ScienceDirect returned 124 results, from which 26 were found to fit our literature review criteria after reading the title and abstract.
  • IEEExplore returned 234 results, from which 44 were found to fit our literature review criteria after reading the title and abstract.
  • Web of Science returned 302 results, from which 62 were found to fit our literature review criteria after reading the title and the abstract.
Of all these relevant results, 16 were duplicated results, leaving us with 113 papers to analyze. Upon a complete read-through of the papers, 41 papers were found to fully fit our criteria and have been included in this review. It should be noted that in the analysis regarding industry use cases, only 35 papers are referenced, as 6 papers were focused on the same case study as others and did not add extra information to our review.
The following parameters were studied: The robot used, control system, application, objectives, key findings, and suggested future work for all these studies, as summarized in Table A3. These were chosen for the following reasons. The robot choice is important, as it highlights which systems are successfully implemented for collaborative applications. The control system is interesting to analyze, as it dictates both safety and performance considerations of the task. Furthermore, when a human is in the control loop, the control system choice is specific to the manner of human–machine interaction— by seeing which methods are more popular and successfully implemented, we can identify trends and future directions. We characterized control systems as vision systems (such as cameras and laser sensors), position systems (such as encoders which are typical of traditional industrial robots), impedance control systems (through haptic interfaces), admittance control (taking advantage of the cobot torque sensors or voltage measurement), audio systems (related to voice command and used for voice/speech recognition), and other systems (that were not easily classified, or that were introduced only in one instance).
The application represents the task given to the cobot, which we believe allows a better understanding to be made regarding the capabilities of collaborative robots. These tasks were divided into assembly (when the cobot collaborates with the operator in an assembly process), human assistance (when the cobot acts as an ergonomic support for the operator, e.g., movable fixtures, quality control, based on vision systems), and lastly, machine tending (when the cobot performs loading/unloading operations).
Furthermore, we divided the objectives into three main topics: Productivity, representing the studies focused on task allocation, quality increase, and reduction of cycle time; safety, which includes not only strictly safety-related topics such as collision avoidance, but also an increase in human ergonomics and reduction of mental stress; and HRI (Human–Robot Interaction), which is focused on the development of new HRI methodologies, e.g., voice recognition. It should be noted that in no way is the proposed subdivision univocal; an interesting example could be [25,26,27]. These works were considered as safety because, even if the proposed solutions keep a high level of productivity, they operate on HRC safety.
The key findings were not grouped, since we believe they depend on the specific study and are too varied; however, they have been summarized in Table A3. Key findings were useful to present the capabilities of the collaborative systems and what HRC studies have achieved. They were included in our analysis in order to identify common solutions. Future work has been grouped into: HRI (works that focus on increasing HRI knowledge and design), safety (works that focusing on increasing the operator safety when working with the cobot), productivity (works focusing on increasing the task productivity in some manner), task complexity (works that focus on increasing the complexity of the task for a particular application), applicability (works that focus on increasing the scope of the work to be used for other industrial applications), and method (works that focus on enhancing the method of HRI via modeling, using alternative robots, or applying general rules and criteria to the design and evaluation process). From these groupings, we can identify ongoing challenges that still need to be solved in the field; by seeing what researchers identify as future work for industrial uptake, we can find trends across the industry in the direction research on which is focused. Our analysis of these parameters is presented in Section 4.

4. Discussion: Trends and Future Perspective

4.1. Trends in the Literature

By examining the literature as summarized in Table A3, we can identify several trends in the use of cobots in industrial settings. It should be noted that, for some of the considered studies in Table A3, we could not identify all of our parameters as specified in Section 3; thus, they are not considered in this specific discussion.
Although early researches utilized traditional industrial robots (Figure 2), the subsequent spread of cobots led to several studies based on the DLR LWR-III (2011), followed by the upgraded KUKA iiwa (from 2016 to 2018), ABB YuMi (also called FRIDA) in 2017 and 2018, and Universal Robots from 2014 to 2018. Several researchers applied the collaborative methods to industrial robots, usually due to their increased performance and widespread availability; however, the disadvantage of this choice is the increase in cost and complexity due to the inclusion of several external sensors and the limited HRC methodologies available. A relationship between the kinematics of the cobot and the application was not explicitly considered, since we believe that other parameters, such as the presence of force sensors in each axis, influenced the cobot choice made in these papers. However, it should be noted that the kinematics—precisely, the number of axes—was a feature considered in [28], whereas future works are focused on verifying their findings with kinematically redundant robots [29] or utilizing the redundancy for achieving better stiffness in hand-guiding [30].
Figure 3 presents the different control systems in the selected human–robot collaboration (HRC) studies. Position control systems were only used for traditional industrial robots, often using extra vision systems for safety reasons. Due to the inherent compliance of cobots, impedance control was more commonly chosen for these systems, though in many cases where an inherently compliant cobot was used, vision was also included for feedback [31,32,33]. Robot compliance can often be a trade-off with robot precision, so including a separate channel for feedback to monitor collisions and increase safety can be a useful method of maintaining manipulation performance. Vision is indeed the prevalent sensor used in HRC studies, also due to the flexibility and affordability of the systems, especially when using depth cameras such as Microsoft Kinect cameras. It is interesting to note that in recent years, Augmented Reality (AR) systems, such as the Microsoft Hololens, have been used more in HRC research, as they are able to provide information to the operator without obscuring their view of the assembly process. In one study, a sensitive skin was incorporated with the cobot to provide environmental information and maintain the operator’s safety. As these skins become more widely studied and developed, we could see this feedback control input become more common, though challenges such as response time must still be solved [34].
The considered studies used the aforementioned robots, both traditional industrial robots and cobots with different collaborative methodologies. Early studies were focused on SSM and PFL methodologies; we believe this focus is due to the need for safety and flexibility in traditional robotic systems and the early spread of cobots. Since 2016 and the introduction of ISO/TS 15066:2016, the considered research sample began to study other methodologies, especially the HG method, which, as shown in Figure 4, has become prevalent in recent years. The HG method is indeed a representative function of collaborative robots [30], since it allows even unskilled users to interact with and program the cobot, which can allow some degree of flexibility—even if the robot moves only on predefined directions—without the need for expensive algorithms [35]. It should be noted that the HG method could also be employed with traditional industrial robots, such as a COMAU NJ130 [36]: This allows one to take advantage of the robot’s characteristics, such as high speed and power, and increase the system’s flexibility.
As stated previously, the collaborative mode depends on the considered application. Figure 5 depicts the considered tasks over the last decade. The most studied task is assembly, likely due to the required flexibility in the task, which makes traditional robotic systems too expensive or difficult to implement. However, the task of production also requires flexibility, and could greatly benefit from collaborative applications. Likely, until the fundamental challenges of setting up collaborative workcells are solved for the easier tasks of assembly, we will not see many case studies targeting production.
In our review, 35 papers presented unique case studies of industrial applications. Two industries seem to drive this research—the automotive industry accounted for 22.85% of studies, and the electronics industry a further 17.14%. Interestingly, research for the automotive industry only began after 2015, and will likely continue to drive research in this area.
HRC studies present several objectives that can be grouped into three main topics. Figure 6 depicts the focus of HRC studies in the last decade. It is interesting to note that the first phase of HRC study [37,38,39,40,41] was more focused on increasing the production and safety aspects of HRC, at least in a manufacturing context. As the research progressed, an increasing number of studies were focused on HRI methodologies, becoming a predominant objective in 2017. The ostensible reduction in 2018 should not mislead us to believe that HRI studies were abandoned in that year: As stated before, the presented classification is not univocal, thus studies such as [42,43,44] could also be considered HRI studies.
The key findings of these studies highlight challenge areas that research has successfully addressed, or even solved, when cobots are used for industrial tasks. Multiple studies reported an increase in task performance—e.g., by reducing completion time and minimizing error [25,37,38,43]— as well as a better understanding of the operator space [29,31,32,41] and higher precision of workpiece manipulation [28,30,45]. Thematic areas of research intent can be identified, such as increasing and quantifying the trust of the operator in the robotic system [29,46,47], as well as improving safety by minimizing collisions [40].
The directions of future work identified in literature are summarized in Figure 7. Historically, researchers aimed to increase the HRI relevance of their work, also with a focus on higher safety requirements and more complex tasks. In recent years, the scope of future work has expanded, with researchers focusing on more complex methods that improve the performance of their systems— whether this is by applying their method to different application fields or more complex tasks. This is likely due to the prevalence of new cobots and sensing methodologies coming onto the market, maturing algorithms, and experience in designing collaborative workcells.
Many of the reviewed works highlight future work in terms of the method they used, whether it be by increasing the complexity of their modeling of the operator and/or environment [48], or using different metrics to evaluate performance [33,49,50] and task choice [51]. Others believe that expanding their research setup to other application areas is the next step [31,45,52]. In our view, these works can be achieved without any step change in existing technology or algorithms; rather, it requires more testing time. To increase safety, productivity, and task performance, researchers will need to improve planners, [39,53], environment and task understanding [28,40,54,55], operator intention understanding [38], and ergonomic cell setups [37,56]. To improve HRI systems, common future work focuses on increasing the robots’ and operators’ awareness of the task and environment by object recognition [44] and integrating multi-modal sensing in an intuitive manner for the operator [3,32,36].
In essence, this future direction focuses on having better understanding of the scene—whether this is what the operator intends to do, what is happening in the environment, or the status of the task. Researchers propose solving this by using more sensors and advanced algorithms, and fusing this information in a way that is easy to use and intuitive for the operator to understand. These systems will inherently lead to better safety, as unexpected motions will be minimized, leading consequently to more trust and uptake. We can expect that many of these advances can come from other areas of robotics research, such as learning by demonstration through hand-guiding or simulation techniques that make it easy to teach a robot a task, and advances in computer vision and machine learning for object recognition and semantic mapping. Other reviews, such as [8], identify similar trends, namely those of improved modeling and understanding, better task planning, and adaptive learning. It will be very interesting to see how this technology is incorporated into the industrial setting to take full advantage of the mechanics and control of cobots and the HRI methodologies of task collaboration.

4.2. Trend of the Market

We believe that the current market should also be presented in order to better place our literature review in the manufacturing context. According to [57], the overall collaborative robot market is estimated to grow from 710 million USD in 2018 to 12,303 million USD by 2025 at a compounded annual growth rate (CAGR) of 50.31% during the forecasted period. However, the International Federation of Robotics (IFR), acknowledging an increase in the robot adoption with over 66% of new sales in 2016, expects that market adoption may proceed at a somewhat slower pace over the forecasted timeframe [58]. However they suggest that the fall in robot prices [59] has led to a growing market for cobots, especially considering that small- and medium-sized enterprises (SMEs), which represent almost 70% of the global number of manufacturers [60] and could not afford robotic applications due to the high capital costs, are now adopting cobots, as they require less expertise and lower installation expenses, confirming a trend presented in scientific works [3].
Finally, [57] highlights that cobots, presenting different payloads, were preferred with up to 5 kg payload capacity; indeed, they held the largest market size in 2017, and a similar trend is expected to continue from 2018 to 2025. This preference of the market towards lightweight robots, which are safer but do not present the high speed and power typically connected with industrial robots [36,61], restrains the HRC possibilities in the current manufacturing scenario. However, we believe that without proper regulation, the current market will continue to mark a dividing line between heavy-duty tasks and HRC methods.

5. Conclusions

Human–robot collaboration is a new frontier for robotics, and the human–robot synergy will constitute a relevant factor in industry for improving production lines in terms of performances and flexibility. This will only be achieved with systems that are fundamentally safe for human operators, intuitive to use, and easy to set up. This paper has provided an overview of the current standards related to Human–Robot Collaboration, showing that it can be applied in a wide range of different modes. The state of the art was presented and the kinematics of several popular cobots were described. A literature analysis was carried out and 41 papers, presenting 35 unique industrial case studies, were reviewed.
Within the context of manufacturing applications, we focused on the control systems, the collaboration methodologies, and the tasks assigned to the cobots in HRC studies. From our analysis, we can identify that the research is largely driven by the electronics and automotive industries, but as cobots become cheaper and easier to integrate into workcells, we can expect SMEs from a wide range of industrial applications to lead their adoption. Objective, key findings and future research directions are also identified, the latter highlighting ongoing challenges that still need to be solved. We can expect that many of the advances needed in the identified directions could come from other areas of robotics research; how these will be incorporated into the industrial setting will lead to new challenges in the future.

Author Contributions

Conceptualization, E.M. and G.R.; Methodology, E.M. and G.R.; Formal analysis, E.M., R.M., and E.G.G.Z.; Investigation, E.M., R.M. and E.G.G.Z.; Data curation, E.M., R.M., and E.G.G.Z.; Writing—original draft preparation, E.M., R.M., and E.G.G.Z.; Writing—review and editing, M.F., E.M., R.M., G.R., and E.G.G.Z.; Supervision, M.F. and G.R.; Project administration, M.F. and G.R.; Funding acquisition, G.R.


This research was funded by University of Padua—Program BIRD 2018—Project no. BIRD187930, and by Regione Veneto FSE Grant 2105-55-11-2018.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Tables

Table A1. List of characteristics of some of the most used cobots for different kinematics.
Table A1. List of characteristics of some of the most used cobots for different kinematics.
BrandProduct NamePayload (Range) [kg]Reach (Range) [mm]Repeatability [mm]Internal Sensors & Control Systemn° of AxisKinematics
ABB (CHE-SWE)YuMi (dual arm IRB 14000, single arm IRB 14050)0.5 for each arm500–559±0.02Position control (encoder) and voltage measurement7 for each arm//
Comau (ITA)AURA1702800±0.1Position control (encoder), sensitive skin, laser scanner (optional)6//
Fanuc (JPN)CR-4iA, CR-7iA, CR-7iA/L, CR-14iA/L, CR-15iA, CR-35iA4–35550–1813± 0.01–0.03Force sensor (base)6Spherical wrist
Kawasaki (JPN)duAro1, duAro22–3 for each arm760–785±0.05Position controlled4 for each armDual arm scara (duAro1)
Kuka (DEU)iisy, iiwa 7 R800, iiwa 14 R8203–14600–820±0.1–0.15 (iisy not yet defined)Torque sensor in each axis and position control (iiwa)6 (iisy) 7 (iiwa)Spherical wrist (and shoulder for iiwa)
Omron (JAP)/ Techman (TWN)TM5 700, TM5 900, TM12, TM144–14700–1300±0.05 (TM5) ±0.12d wrist camera, position control (encoder), voltage measurement6Three parallel axes
Precise Automation (USA)PP100, PF3400, PAVP6, PAVS62–7432–770 or 685–1270±0.02–0.1Position control (encoder), voltage measurement4 (cartesian, scara) 6 (spherical wrist)Cartesian, scara, spherical wrist
Rethink Robotics (USA/DEU)Baxter, Sawyer4 (2 for each arm for Baxter)1260 for each arm±0.1Torque sensors in each axis, position control7 for each arm//
StaübliTX2-40, TX2(touch)-60, TX2(touch)-60L, TX2(touch)-90, TX2(touch)-90L, TX2touch-90XL2.3–20515–1450±0.02–0.04Position control, sensitive skin (touch)6Spherical wrist
Universal Robots (DNK)UR3, UR3e, UR5, UR5e, UR10, UR10e, UR16e3–16500–1300±0.03–0.1Position control (encoder), voltage measuremen, Force torque multiaxis load cell (e-series)6Three parallel axes
Yaskawa (JPN)HC10, HC10DT101200±0.1Torque sensors in each axis, position control6//
Table A2. Denavit–Hartenberg parameters and singularity configurations for the considered kinematic schemes.
Table A2. Denavit–Hartenberg parameters and singularity configurations for the considered kinematic schemes.
Kinematic SchemeDenavit–Hartenberg ParametersSingularity Configurations
Six axes with spherical wrist
T α a θ d
T 10 00 q 1 d 1
T 21 −90 a 1 - 90 + q 2 0
T 32 0 a 2 q 3 0
T 43 −90 a 3 q 4 d 4
T 54 900 q 5 0
T 65 −900 q 6 d 6
Wrist: J4 and J6 aligned Robotics 08 00100 i001
Shoulder: Wrist aligned with J1 Robotics 08 00100 i002
Elbow: Wrist coplanar with J2 and J3 Robotics 08 00100 i003
Six axes with three parallel axes
T α a θ d
T 10 00 q 1 d 1
T 21 −900 - 90 + q 2 d 2
T 32 0 a 2 q 3 - d 3
T 43 0 a 3 - 90 + q 4 d 4
T 54 −900 180 + q 5 d 5
T 65 −900 q 6 d 6
Wrist: J6 // J4 Robotics 08 00100 i004
Shoulder: Intersection of J5 and J6 coplanar with J1 and J2 Robotics 08 00100 i005
Elbow: J2, J3, and J4 coplanar Robotics 08 00100 i006
Six axes with offset wrist
T α a θ d
T 10 00 q 1 d 1
T 21 900 90 + q 2 - d 2
T 32 180 a 2 90 + q 3 - d 3
T 43 900 180 + q 4 - d 4
T 54 900 180 + q 5 - d 5
T 65 900 q 6 - d 6
Wrist: J5 ≈ 0° or 180° ± 15° Robotics 08 00100 i007
Shoulder: Wrist point near the yellow column (300 mm of radius) Robotics 08 00100 i008
Elbow: J3 ≈ 0° or 180° ±15° Robotics 08 00100 i009
Seven axes with spherical joints
T α a θ d
T 10 00 q 1 d 1
T 21 −900 q 2 0
T 32 900 q 3 d 3
T 43 900 q 4 0
T 54 −900 q 5 d 5
T 65 −900 q 6 0
T 76 900 q 7 d 7
Wrist motion: J6 = 0 & J4 = 90° Robotics 08 00100 i010
Shoulder Motion: J2 = 0 & J3 = ± 90° Robotics 08 00100 i011
Elbow Motion: J5 = ± 90° & J6 = 0 Robotics 08 00100 i012
Seven axes without spherical joints
T α a θ d
T 10 00 q 1 d 1
T 21 −90 - a 1 q 2 0
T 32 90 a 2 q 3 d 3
T 43 90 a 3 90 + q 4 0
T 54 90 a 4 q 5 d 5
T 65 90 - a 5 q 6 0
T 76 −90 a 6 180 + q 7 d 7
Wrist: J5 // J7 Robotics 08 00100 i013
Shoulder: Wrist point near J1 direction Robotics 08 00100 i014
Elbow: J3 // J5 Robotics 08 00100 i015
Table A3. Literature review analysis.
Table A3. Literature review analysis.
AuthorYearRobot UsedControl SystemCollaboration MethodsApplicationObjective of the StudyKey FindingsFuture Works
J. T. C. Tan et al. [37]2009Industrial robotAssemblyProductivity: Information display for operator supportDevelopment and implementation of an information support system, which leads to a reduction in assembly time.HRI: Studies on the position of the LCD TV (information source) in order to improve information reception.
T. Arai et al. [38]2009Industrial robotVisionPower and Force limitingAssemblyProductivity: Improve efficiency through devices that support the operatorHRC doubles productivity in comparison with manual assembly; reduction of human errors up to no defectsSafety: Monitoring operator’s fatigue and intention
J. T. C. Tan et al. [39]2009Industrial robotVisionPower and Force limitingAssemblyProductivity: Optimize working efficiency, improving quality and productivitySafety analysis lead to an increased distance and a reduction in the robot speed; mental strain due to working with robots reducedTask complexity: Analysis of the improvements obtained with HRC to order tasks.
C. Lenz et al. [40]2009Industrial robotPositionSpeed and separation monitoringAssemblySafety: HRC with traditional industrial robots, with focus on safetyCollision avoidance method based on restricted robot movements and virtual force fieldsSafety: Estimation and tracking of human body pose to avoid collision without the use of any markers.
T. Ende et al. [41]2011DLR LWR-III 3Gesture recognitionHuman assistantProductivity: Gather gestures for HRC from humansEleven gestures present recognition rate over 80%; recognition problem when torso position is part of the gestureN/A
H. Ding et al. [25]2013ABB FRIDA (YuMi)VisionSpeed and separation monitoringAssemblySafety: Collaborative behavior with operator safety and without productivity losses due to emergency stopsSpeed reduction applied based on the distance between human arm and Kinect position avoids emergency stopsN/A
H. Ding et al. [26]2013ABB FRIDA (YuMi)VisionSpeed and separation monitoringAssemblySafety: Multiple operators in collaborative behavior with operators safety concern and without productivity lossesDevelopment of a finite state automaton; speed reduction improves the uptime while respecting safety constraintsN/A
A. M. Zanchettin et al. [27]2013Industrial robotPositionSpeed and separation monitoringQuality controlSafety: Compromise between safety and productivity; Adaptable robot speedDevelopment of a safety-oriented path-constrained motion planning, tracking operator, and reducing robot speedN/A
K. P. Hawkins et al. [32]2014Universal Robots UR10VisionPower and Force limitingAssemblyProductivity: Robots need to anticipate human actions even with task or sensor ambiguityCompromises between human wait times and confidence in the human action detectionHRI: impact of the system on the operator’s sense of fluency, i.e., synchronization between cobot and operator
K. R. Guerin et al. [3]2015Universal Robots UR5ImpedanceMachine tendingProductivity: Robot assistant with set of capabilities for typical SMEsTest of machine tending: 82% of parts taken from machine (due to bad weld or bad grasp)HRI: Test ease of use and focus on HRC, gesture recognition for learning
I. D. Walker et al. [46]2015Industrial robotPositionHuman assistantHRI: Trust not considered in handoffs; derive model for robot trust on operatorThe robot pose changes accordingly to trust in human, reducing impact forces in case of low trustApplicability: Effectiveness of the approach in SME scenarios
T. Hamabe et al. [54]2015Kawada HIROVisionPower and Force limitingAssemblyHRI: Learn task from human demonstrationHuman and robot roles fixed, due to limits in robot’s manipulation capabilities; changing task order increases time due to recognition systemTask complexity: Complete set of scenarios assumed, cobot should be able to recognize new tasks autonomously
S. M. M. Rahman et al. [29]2016Kinova MICO 2-fingerVisionAssemblyHRI: Derive model for robot trust on human; trust-based motion planning for handover taskspHRI gets better with trust for contextual information transparent to human; increase: 20% safety, 30% handover success, 6.73% efficiencyMethod: Apply the proposed method with kinematically redundant robot
S. M. M. Rahman et al. [49]2016Rethink BaxterVisionAssemblyProductivity: Autonomous error detection with human’s interventionRegret based method leads to improvement of fluency, due to an increase in synchronization, reduction in mean cognitive workload and increase in human trust compared to a Bayesian approachMethod: Different objective criteria for the regret-based approach to evaluate HRC performance
L. Rozo et al. [48]2016WAM robot/ KUKA LWR iiwaImpedance/ admittanceHand-guidingAssemblyHRI: Robotic assistant needs to be easily reprogrammed, thus programming by demonstrationModel adapts to changes in starting and ending point, task, and control mode; with high compliance, the robot can not follow trajectoryMethod: Estimation of the damping matrix for the spring damper model; study how interaction forces can change the robot behaviors
A. M. Zanchettin et al. [62]2016ABB FRIDA (YuMi)VisionSpeed and separation monitoringAssemblySafety: Collision avoidance strategy: Decrease the speed of the cobotSpeed reduction method based on minimum distance; distance threshold adaptable to the programmed speed; continuous speed scalingN/A
A. Cherubini et al. [63]2016KUKA LWR4+AdmittanceSafety rated monitored stopAssemblySafety: Collaborative human–robot application to assembly a car homokinetic jointFramework that integrates many state-of-the-art robotics components, applied in real industrial scenariosProductivity: Deploying the proposed methodologies on mobile manipulator robots to increase flexibility
H. Fakhruldeen et al. [51]2016Rethink BaxterVision and audio recognitionPower and Force limitingAssemblyHRI: Implementation of a self-built planner in a cooperative task where the cobot actively collaboratesAPI development combining an object-oriented programming scheme with a Prolog meta interpreter to create these plans and execute themMethod: add cost to action evaluation, add non-productive actions, action completion percentage should be considered
S. Makris et al. [64]2016Industrial robotAR systemHand-guidingHuman assistantSafety: Development of an AR system in aid of operators in human–robot collaborative environment to reduce mental stressThe proposed system minimizes the time required for the operator to access information and send feedback; it decreases the stoppage and enhances the training processTask complexity: Application in other industrial environments
B. Whitsell et al. [28]2017KUKA LBR iiwaImpedance / admittanceHand-guiding/ Power and Force LimitingAssemblyHRI: Cooperate in everyday environment; robots need to adapt to human; haptics should adapt to the operator ways100% correct placement of a block in 1440 trials; robot can control a DOF if the operator does not control it (95.6%); lessening the human responsibility by letting the robot control an axis reduces the completion timeApplicability: Adapt variables to environment, e.g. task and robot coordinate system not aligned
J. Bös et al. [65]2017ABB YuMiAdmittancePower and Force limitingAssemblyProductivity: Increase assembly speed without reducing the flexibility or increasing contact forces using iterative learning control (ILC)Increase acceleration by applying Dynamic Movement Primitives to an ILC, reduce contact forces by adding a learning controller. Stochastic disturbances do not have a long term effect; task duration decreases by 40%, required contact force by 50%Method: Study and theoretical proof on stability in the long term
M. Wojtynek et al. [33]2017KUKA LBR iiwaVisionHand-guidingAssemblyProductivity: Create a modular and flexible system; abstraction of any equipmentEasy reconfiguration, without complex programmingMethod: Introduce metrics for quantitative measurement of HRC
B. Sadrfaridpour et al. [47]2017Rethink BaxterHuman tracking systemPower and Force limitingAssemblyHRI: Combination of pHRI and sHRI in order to predict human behavior and choose robot path and speedAugmenting physical/social capabilities increases one subjective measure (trust, workload, usability); assembly time does not changeN/A
I. El Makrini et al. [56]2017Rethink BaxterVisionPower and Force limitingAssemblyHRI: HRC based on natural communication; framework for the cobot to communicateThe framework is validated; more intuitive HRITask Complexity: Adapt robot to user; adjust parts position based on user’s height
P. J. Koch et al. [52]2017KUKA LWR4+AdmittanceScrewing for maintenanceHRI: Cobot development: Focus on intuitive human–robot interfaceHR interface, simple for user reconfiguration. Steps in order to transform a mobile manipulator into a cobotApplicability: Expand to several industrial maintenance tasks
M. Haage et al. [66]2017ABB YuMiVisionAssemblyHRI: Reduce the time and required expertise to setup a robotized assembly stationA web-based HRI for assisting human instructors to teach assembly tasks in a straightforward and intuitive mannerN/A
P. Gustavsson et al. [50]2017Universal Robots UR3Impedance and audio recognitionHand-guiding/ Power and Force LimitingAssemblyHRI: Joint Speech recognition and a haptic control in order to obtain an intuitive HRCDeveloped a simplified HRI responsive to vocal commands, that guides the user in the progress of the task with hapticsMethod: Test if haptic control can be used to move the robot with linear motions; an automatic way of logging the accuracy
M. Safeea et al. [30]2017KUKA LBR iiwaAdmittanceHand-guiding/ Safety rated Monitored StopAssemblySafety: Precise and intuitive hand guidingPossible to hand-guide the robot with accuracy, with no vibration, and in a natural and intuitive wayMethod: Utilizing the redundancy of iiwa to achieve better stiffness in hand-guiding
W. Wang et al. [45]2018Industrial robotGesture recognitionHand-guiding/ Power and Force LimitingAssemblyProductivity: Easier reconfiguration of the cobot using a teaching-by-demonstration model95% of accuracy (higher than previous methods based on vision); Lower inefficient timeApplicability: Use multimodal information to investigate different applications
N. Mendes et al. [31]2018KUKA LBR iiwaVisionSpeed and separation monitoringAssemblyHRI: Need for a flexible system with simple and fast interfaceGesture is intuitive but delays process; constrained flexibilityApplicability: Expand use case to several industrial fields
K. Darvish et al. [42]2018Rethink BaxterVisionPower and Force limitingAssemblyProductivity: Increase robot adaptability integrated in the FlexHRC architecture by an online task plannerPlanning and task representation require little time (less than 1% of total); the simulation follows the real data very wellN/A
A. Zanchettin et al. [43]2018ABB YuMiVisionPower and Force limitingAssemblyProductivity: Predict human behavior in order to increase the robot adaptabilityDecrease of task time equal to 17%;N/A
G. Pang et al. [34]2018ABB YuMiSensitive skinSafety rated monitored stop/ Speed and separation monitoringTest of collisionSafety: Cobot perceives stimulus only by its torque sensors, not guaranteeing collision avoidanceIntegration on cobot of sensitive skin; delay in the system reactionMethod: Reduce contact area of the sensitive skin; test with multisensing systems
V. V. Unhelkar et al. [53]2018Universal Robots UR10VisionSpeed and separation monitoringHuman assistantProductivity: Cobots can be successful, but they have restricted range: Mobile cobots for delivering partsPrediction of long motion (16 s) with a prediction time horizon up to 6 s; in simulation, reduced safety rated monitored stops, increasing task efficiencySafety: Recognize unmodeled motion and incremental planners
V. Tlach et al. [55]2018Industrial robotAdmittanceHand-guiding/ Safety rated Monitored StopAssemblyProductivity: Design of collaborative tasks in an applicationThe method is flexible to the type of productProductivity: Improve methods for recognizing objects
S. Heydaryan et al. [35]2018KUKA LBR iiwa 14 R820Vision / admittanceHand-guiding/ Safety rated Monitored StopAssemblySafety: Task allocation to ensure safety of the operator, increase productivity by increasing ergonomicsAssembly time of 203 s in SMS, but the robot obstructs the access to some screws; proposed a hand-guided solution (210s)N/A
G. Michalos et al. [36]2018Industrial robotAdmittanceHand-guidingAssemblySafety: Implementation of a robotic system for HRC assemblyDevelopment of HRC assembly cell with high payload industrial robots and human operators.HRI: Improve human immersion in the cell. Integrate all the sensing and interaction equipment
V. Gopinath et al. [61]2018Industrial robotAR systemAssemblySafety: Development of a collaborative assembly cell with large industrial robotsDevelopment of two work stationsN/A
A. Blaga et al. [44]2018Rethink BaxterVision and audio recognitionPower and Force limitingAssemblyProductivity: Improve the possibilities of the integration of Augmented Reality in collaborative tasks to shorten lead timesAR and HRC were integrated into an unitary system, meant to ease a worker’s daily tasks regarding the visualization of the next possible assembly stepHRI: Using object recognition combined with 3D printing, along with the latest HMD devices


  1. Barbazza, L.; Faccio, M.; Oscari, F.; Rosati, G. Agility in assembly systems: A comparison model. Assem. Autom. 2017, 37, 411–421. [Google Scholar] [CrossRef]
  2. Colgate, J.E.; Edward, J.; Peshkin, M.A.; Wannasuphoprasit, W. Cobots: Robots for Collaboration with Human Operators. In Proceedings of the 1996 ASME International Mechanical Engineering Congress and Exposition, Atlanta, GA, USA, 17–22 November 1996; pp. 433–439. [Google Scholar]
  3. Guerin, K.R.; Lea, C.; Paxton, C.; Hager, G.D. A framework for end-user instruction of a robot assistant for manufacturing. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 6167–6174. [Google Scholar]
  4. Peshkin, M.A.; Colgate, J.E.; Wannasuphoprasit, W.; Moore, C.A.; Gillespie, R.B.; Akella, P. Cobot architecture. IEEE Trans. Robot. Autom. 2001, 17, 377–390. [Google Scholar] [CrossRef]
  5. DLR—Institute of Robotics and Mechatronics. History of the DLR LWR. Available online: (accessed on 30 November 2019).
  6. Universal Robots. Low Cost and Easy Programming Made the UR5 a Winner. Available online: (accessed on 30 November 2019).
  7. Müller, R.; Vette, M.; Geenen, A. Skill-based dynamic task allocation in Human-Robot-Cooperation with the example of welding application. Procedia Manuf. 2017, 11, 13–21. [Google Scholar] [CrossRef]
  8. Wang, L.; Gao, R.; Váncza, J.; Krüger, J.; Wang, X.V.; Makris, S.; Chryssolouris, G. Symbiotic human-robot collaborative assembly. CIRP Ann. 2019, 68, 701–726. [Google Scholar] [CrossRef]
  9. Müller, R.; Vette, M.; Mailahn, O. Process-oriented task assignment for assembly processes with human-robot interaction. Procedia CIRP 2016, 44, 210–215. [Google Scholar] [CrossRef]
  10. Wang, X.V.; Kemény, Z.; Váncza, J.; Wang, L. Human–robot collaborative assembly in cyber-physical production: Classification framework and implementation. CIRP Ann. 2017, 66, 5–8. [Google Scholar] [CrossRef]
  11. Krüger, J.; Lien, T.K.; Verl, A. Cooperation of human and machines in assembly lines. CIRP Ann. 2009, 58, 628–646. [Google Scholar]
  12. Gaskill, S.; Went, S. Safety issues in modern applications of robots. Reliab. Eng. Syst. Saf. 1996, 53, 301–307. [Google Scholar] [CrossRef]
  13. Michalos, G.; Makris, S.; Tsarouchi, P.; Guasch, T.; Kontovrakis, D.; Chryssolouris, G. Design considerations for safe human-robot collaborative workplaces. Procedia CIrP 2015, 37, 248–253. [Google Scholar] [CrossRef]
  14. Gravel, D.P.; Newman, W.S. Flexible robotic assembly efforts at Ford Motor Company. In Proceeding of the 2001 IEEE International Symposium on Intelligent Control (ISIC’01) (Cat. No. 01CH37206), Mexico City, Mexico, 5–7 September 2001; pp. 173–182. [Google Scholar]
  15. Zhu, Z.; Hu, H. Robot learning from demonstration in robotic assembly: A survey. Robotics 2018, 7, 17. [Google Scholar]
  16. Fechter, M.; Foith-Förster, P.; Pfeiffer, M.S.; Bauernhansl, T. Axiomatic design approach for human-robot collaboration in flexibly linked assembly layouts. Procedia CIRP 2016, 50, 629–634. [Google Scholar] [CrossRef]
  17. Faccio, M.; Bottin, M.; Rosati, G. Collaborative and traditional robotic assembly: A comparison model. Int. J. Adv. Manuf. Technol. 2019, 102, 1355–1372. [Google Scholar] [CrossRef]
  18. Edmondson, N.; Redford, A. Generic flexible assembly system design. Assem. Autom. 2002, 22, 139–152. [Google Scholar] [CrossRef]
  19. Battini, D.; Faccio, M.; Persona, A.; Sgarbossa, F. New methodological framework to improve productivity and ergonomics in assembly system design. Int. J. Ind. Ergon. 2011, 41, 30–42. [Google Scholar] [CrossRef]
  20. Sawodny, O.; Aschemann, H.; Lahres, S. An automated gantry crane as a large workspace robot. Control Eng. Pract. 2002, 10, 1323–1338. [Google Scholar] [CrossRef]
  21. Krüger, J.; Bernhardt, R.; Surdilovic, D.; Spur, G. Intelligent assist systems for flexible assembly. CIRP Ann. 2006, 55, 29–32. [Google Scholar] [CrossRef]
  22. Rosati, G.; Faccio, M.; Carli, A.; Rossi, A. Fully flexible assembly systems (F-FAS): A new concept in flexible automation. Assem. Autom. 2013, 33, 8–21. [Google Scholar] [CrossRef]
  23. FANUC Italia, S.r.l. M-2000—The Strongest Heavy Duty Industrial Robot in the Marker. Available online: (accessed on 30 November 2019).
  24. Hägele, M.; Schaaf, W.; Helms, E. Robot assistants at manual workplaces: Effective co-operation and safety aspects. In Proceedings of the 33rd ISR (International Symposium on Robotics), Stockholm, Sweden, 7–11 October 2002; Volume 7. [Google Scholar]
  25. Ding, H.; Heyn, J.; Matthias, B.; Staab, H. Structured collaborative behavior of industrial robots in mixed human-robot environments. In Proceedings of the 2013 IEEE International Conference on Automation Science and Engineering (CASE), Madison, WI, USA, 17–20 August 2013; pp. 1101–1106. [Google Scholar]
  26. Ding, H.; Schipper, M.; Matthias, B. Collaborative behavior design of industrial robots for multiple human-robot collaboration. In Proceedings of the IEEE ISR 2013, Seoul, Korea, 24–26 October 2013; pp. 1–6. [Google Scholar]
  27. Zanchettin, A.M.; Rocco, P. Path-consistent safety in mixed human-robot collaborative manufacturing environments. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 1131–1136. [Google Scholar]
  28. Whitsell, B.; Artemiadis, P. Physical human–robot interaction (pHRI) in 6 DOF with asymmetric cooperation. IEEE Access 2017, 5, 10834–10845. [Google Scholar] [CrossRef]
  29. Rahman, S.M.; Wang, Y.; Walker, I.D.; Mears, L.; Pak, R.; Remy, S. Trust-based compliant robot-human handovers of payloads in collaborative assembly in flexible manufacturing. In Proceedings of the 2016 IEEE International Conference on Automation Science and Engineering (CASE), Fort Worth, TX, USA, 21–25 August 2016; pp. 355–360. [Google Scholar]
  30. Safeea, M.; Bearee, R.; Neto, P. End-effector precise hand-guiding for collaborative robots. In Iberian Robotics Conference; Springer: Berlin, Germany, 2017; pp. 595–605. [Google Scholar]
  31. Mendes, N.; Safeea, M.; Neto, P. Flexible programming and orchestration of collaborative robotic manufacturing systems. In Proceedings of the 2018 IEEE 16th International Conference on Industrial Informatics (INDIN), Porto, Portugal, 18–20 July 2018; pp. 913–918. [Google Scholar]
  32. Hawkins, K.P.; Bansal, S.; Vo, N.N.; Bobick, A.F. Anticipating human actions for collaboration in the presence of task and sensor uncertainty. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 2215–2222. [Google Scholar]
  33. Wojtynek, M.; Oestreich, H.; Beyer, O.; Wrede, S. Collaborative and robot-based plug & produce for rapid reconfiguration of modular production systems. In Proceedings of the 2017 IEEE/SICE International Symposium on System Integration (SII), Taipei, Taiwan, 11–14 December 2017; pp. 1067–1073. [Google Scholar]
  34. Pang, G.; Deng, J.; Wang, F.; Zhang, J.; Pang, Z.; Yang, G. Development of flexible robot skin for safe and natural human–robot collaboration. Micromachines 2018, 9, 576. [Google Scholar] [CrossRef]
  35. Heydaryan, S.; Suaza Bedolla, J.; Belingardi, G. Safety design and development of a human-robot collaboration assembly process in the automotive industry. Appl. Sci. 2018, 8, 344. [Google Scholar] [CrossRef]
  36. Michalos, G.; Kousi, N.; Karagiannis, P.; Gkournelos, C.; Dimoulas, K.; Koukas, S.; Mparis, K.; Papavasileiou, A.; Makris, S. Seamless human robot collaborative assembly—An automotive case study. Mechatronics 2018, 55, 194–211. [Google Scholar] [CrossRef]
  37. Tan, J.T.C.; Zhang, Y.; Duan, F.; Watanabe, K.; Kato, R.; Arai, T. Human factors studies in information support development for human-robot collaborative cellular manufacturing system. In Proceedings of the RO-MAN 2009—The 18th IEEE International Symposium on Robot and Human Interactive Communication, Toyama, Japan, 27 September–2 October 2009; pp. 334–339. [Google Scholar]
  38. Arai, T.; Duan, F.; Kato, R.; Tan, J.T.C.; Fujita, M.; Morioka, M.; Sakakibara, S. A new cell production assembly system with twin manipulators on mobile base. In Proceedings of the 2009 IEEE International Symposium on Assembly and Manufacturing, Suwon, Korea, 17–20 November 2009; pp. 149–154. [Google Scholar]
  39. Tan, J.T.C.; Duan, F.; Zhang, Y.; Watanabe, K.; Kato, R.; Arai, T. Human-robot collaboration in cellular manufacturing: Design and development. In Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, Saint Louis, MO, USA, 10–15 October 2009; pp. 29–34. [Google Scholar]
  40. Lenz, C.; Rickert, M.; Panin, G.; Knoll, A. Constraint task-based control in industrial settings. In Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, Saint Louis, MO, USA, 10–15 October 2009; pp. 3058–3063. [Google Scholar]
  41. Ende, T.; Haddadin, S.; Parusel, S.; Wüsthoff, T.; Hassenzahl, M.; Albu-Schäffer, A. A human-centered approach to robot gesture based communication within collaborative working processes. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 3367–3374. [Google Scholar]
  42. Darvish, K.; Bruno, B.; Simetti, E.; Mastrogiovanni, F.; Casalino, G. Interleaved Online Task Planning, Simulation, Task Allocation and Motion Control for Flexible Human-Robot Cooperation. In Proceedings of the 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), Nanjing, China, 27–31 August 2018; pp. 58–65. [Google Scholar]
  43. Zanchettin, A.; Casalino, A.; Piroddi, L.; Rocco, P. Prediction of human activity patterns for human-robot collaborative assembly tasks. IEEE Trans. Ind. Inf. 2018, 15, 3934–3942. [Google Scholar] [CrossRef]
  44. Blaga, A.; Tamas, L. Augmented Reality for Digital Manufacturing. In Proceedings of the 2018 26th Mediterranean Conference on Control and Automation (MED), Akko, Israel, 1–4 July 2018; pp. 173–178. [Google Scholar]
  45. Wang, W.; Li, R.; Diekel, Z.M.; Chen, Y.; Zhang, Z.; Jia, Y. Controlling Object Hand-Over in Human–Robot Collaboration Via Natural Wearable Sensing. IEEE Trans. Human-Mach. Syst. 2018, 49, 59–71. [Google Scholar] [CrossRef]
  46. Walker, I.D.; Mears, L.; Mizanoor, R.S.; Pak, R.; Remy, S.; Wang, Y. Robot-human handovers based on trust. In Proceedings of the 2015 IEEE Second International Conference on Mathematics and Computers in Sciences and in Industry (MCSI), Sliema, Malta, 17 August 2015; pp. 119–124. [Google Scholar]
  47. Sadrfaridpour, B.; Wang, Y. Collaborative assembly in hybrid manufacturing cells: An integrated framework for human–robot interaction. IEEE Trans. Autom. Sci. Eng. 2017, 15, 1178–1192. [Google Scholar] [CrossRef]
  48. Rozo, L.; Calinon, S.; Caldwell, D.G.; Jimenez, P.; Torras, C. Learning physical collaborative robot behaviors from human demonstrations. IEEE Trans. Robot. 2016, 32, 513–527. [Google Scholar] [CrossRef]
  49. Rahman, S.M.; Liao, Z.; Jiang, L.; Wang, Y. A regret-based autonomy allocation scheme for human-robot shared vision systems in collaborative assembly in manufacturing. In Proceedings of the 2016 IEEE International Conference on Automation Science and Engineering (CASE), Fort Worth, TX, USA, 21–25 August 2016; pp. 897–902. [Google Scholar]
  50. Gustavsson, P.; Syberfeldt, A.; Brewster, R.; Wang, L. Human-robot collaboration demonstrator combining speech recognition and haptic control. Procedia CIRP 2017, 63, 396–401. [Google Scholar] [CrossRef]
  51. Fakhruldeen, H.; Maheshwari, P.; Lenz, A.; Dailami, F.; Pipe, A.G. Human robot cooperation planner using plans embedded in objects. IFAC-PapersOnLine 2016, 49, 668–674. [Google Scholar] [CrossRef]
  52. Koch, P.J.; van Amstel, M.K.; Dȩbska, P.; Thormann, M.A.; Tetzlaff, A.J.; Bøgh, S.; Chrysostomou, D. A skill-based robot co-worker for industrial maintenance tasks. Procedia Manuf. 2017, 11, 83–90. [Google Scholar] [CrossRef]
  53. Unhelkar, V.V.; Lasota, P.A.; Tyroller, Q.; Buhai, R.D.; Marceau, L.; Deml, B.; Shah, J.A. Human-aware robotic assistant for collaborative assembly: Integrating human motion prediction with planning in time. IEEE Robot. Autom. Lett. 2018, 3, 2394–2401. [Google Scholar] [CrossRef]
  54. Hamabe, T.; Goto, H.; Miura, J. A programming by demonstration system for human-robot collaborative assembly tasks. In Proceedings of the 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO), Zhuhai, China, 6–9 December 2015; pp. 1195–1201. [Google Scholar]
  55. Tlach, V.; Kuric, I.; Zajačko, I.; Kumičáková, D.; Rengevič, A. The design of method intended for implementation of collaborative assembly tasks. Adv. Sci. Technol. Res. J. 2018, 12, 244–250. [Google Scholar] [CrossRef]
  56. El Makrini, I.; Merckaert, K.; Lefeber, D.; Vanderborght, B. Design of a collaborative architecture for human-robot assembly tasks. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1624–1629. [Google Scholar]
  57. MarketsandMarkets™ Research Private Ltd. Collaborative Robots Market by Payload Capacity (Up to 5 kg, Up to 10 kg, Above 10 kg), Industry (Automotive, Electronics, Metals & Machining, Plastics & Polymer, Food & Agriculture, Healthcare), Application, and Geography—Global Forecast to 2023. Available online: (accessed on 30 November 2019).
  58. International Federation of Robotics (IFR). Robots and the Workplace of the Future. 2018. Available online: (accessed on 30 November 2019).
  59. Barclays Investment Bank. Technology’s Mixed Blessing. 2017. Available online: (accessed on 30 November 2019).
  60. Tobe, F. Why Co-Bots Will Be a Huge Innovation and Growth Driver for Robotics Industry. 2015. Available online: (accessed on 30 November 2019).
  61. Gopinath, V.; Ore, F.; Grahn, S.; Johansen, K. Safety-Focussed Design of Collaborative Assembly Station with Large Industrial Robots. Procedia Manuf. 2018, 25, 503–510. [Google Scholar] [CrossRef]
  62. Zanchettin, A.M.; Ceriani, N.M.; Rocco, P.; Ding, H.; Matthias, B. Safety in human-robot collaborative manufacturing environments: Metrics and control. IEEE Trans. Autom. Sci. Eng. 2015, 13, 882–893. [Google Scholar] [CrossRef]
  63. Cherubini, A.; Passama, R.; Crosnier, A.; Lasnier, A.; Fraisse, P. Collaborative manufacturing with physical human–robot interaction. Robot. Comput. Integr. Manuf. 2016, 40, 1–13. [Google Scholar] [CrossRef]
  64. Makris, S.; Karagiannis, P.; Koukas, S.; Matthaiakis, A.S. Augmented reality system for operator support in human–robot collaborative assembly. CIRP Ann. 2016, 65, 61–64. [Google Scholar] [CrossRef]
  65. Bös, J.; Wahrburg, A.; Listmann, K.D. Iteratively Learned and Temporally Scaled Force Control with application to robotic assembly in unstructured environments. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 3000–3007. [Google Scholar]
  66. Haage, M.; Piperagkas, G.; Papadopoulos, C.; Mariolis, I.; Malec, J.; Bekiroglu, Y.; Hedelind, M.; Tzovaras, D. Teaching assembly by demonstration using advanced human robot interaction and a knowledge integration framework. Procedia Manuf. 2017, 11, 164–173. [Google Scholar] [CrossRef]
Figure 1. Types of use of a collaborative robot.
Figure 1. Types of use of a collaborative robot.
Robotics 08 00100 g001
Figure 2. Robot usage in selected human–robot collaboration studies in the period 2009–2018.
Figure 2. Robot usage in selected human–robot collaboration studies in the period 2009–2018.
Robotics 08 00100 g002
Figure 3. Control systems used in selected human–robot collaboration studies in the period 2009–2018: In red, the number of vision systems; in orange, position-controlled systems (used especially for traditional industrial robots); in gray, the cases for impedance control (e.g., through haptic interfaces); in yellow, admittance control (e.g., through torque sensors); in blue, audio systems (for voice/speech recognition); and green for other systems.
Figure 3. Control systems used in selected human–robot collaboration studies in the period 2009–2018: In red, the number of vision systems; in orange, position-controlled systems (used especially for traditional industrial robots); in gray, the cases for impedance control (e.g., through haptic interfaces); in yellow, admittance control (e.g., through torque sensors); in blue, audio systems (for voice/speech recognition); and green for other systems.
Robotics 08 00100 g003
Figure 4. Collaboration methods used in selected human–robot collaboration in the period 2009–2018: In blue, hand guiding (HG); in orange, safety-rated monitored stop (SMS); in gray, speed and separation monitoring (SSM); in yellow, power and force limiting (PFL).
Figure 4. Collaboration methods used in selected human–robot collaboration in the period 2009–2018: In blue, hand guiding (HG); in orange, safety-rated monitored stop (SMS); in gray, speed and separation monitoring (SSM); in yellow, power and force limiting (PFL).
Robotics 08 00100 g004
Figure 5. Tasks assigned to the robot in selected collaborative applications in research in the period 2009–2018: In blue, assembly tasks; in orange, the tasks used to assist the operator, e.g., handover of parts, quality control tasks, or machine tending, i.e., loading and/or unloading.
Figure 5. Tasks assigned to the robot in selected collaborative applications in research in the period 2009–2018: In blue, assembly tasks; in orange, the tasks used to assist the operator, e.g., handover of parts, quality control tasks, or machine tending, i.e., loading and/or unloading.
Robotics 08 00100 g005
Figure 6. Main topics or objectives in HRC studies. The objectives were divided into productivity studies (blue), safety studies (orange), e.g., ergonomics and collision avoidance, HRI (Human–Robot Interaction) studies (gray), e.g., development or improvement of HRI methodologies.
Figure 6. Main topics or objectives in HRC studies. The objectives were divided into productivity studies (blue), safety studies (orange), e.g., ergonomics and collision avoidance, HRI (Human–Robot Interaction) studies (gray), e.g., development or improvement of HRI methodologies.
Robotics 08 00100 g006
Figure 7. Future work topics from HRC studies. The work was divided into directions of HRI (dark blue), safety (orange), task complexity (gray), applicability (yellow), method (light blue), and productivity (green).
Figure 7. Future work topics from HRC studies. The work was divided into directions of HRI (dark blue), safety (orange), task complexity (gray), applicability (yellow), method (light blue), and productivity (green).
Robotics 08 00100 g007
Table 1. Qualitative evaluation of the most suitable solutions for the main industry tasks.
Table 1. Qualitative evaluation of the most suitable solutions for the main industry tasks.
Human OperatorCollaborative SystemsTraditional RobotHandling Systems
AssemblyHigh dexterity and flexibilityCombines human dexterity with robot capabilities [24]Dexterity/flexibility could be unreachable [24]No complex tasks with commercial end-effectors [21]
PlacementHigh dexterityCommercial cobots have lower repeatabilityHigh repeatability and payloadHigh payload
HandlingProduct weight restricted [19]Typical cobots have low payloadHigh payload and speed [23]High payload
PickingProduct weight restricted [19]Typical cobots have low payloadHigh payload and repeatability [23]Bin picking difficult due to size
Back to TopTop