A Survey of Robots in Healthcare
A Survey of Robots in Healthcare
A Survey of Robots in Healthcare
net/publication/348654628
CITATIONS READS
44 1,433
9 authors, including:
Some of the authors of this publication are also working on these related projects:
Designing a Socially Assistive Robot for Adaptive and Personalized Training View project
Rewind-Remind: Investigating Associative Memory in Younger and Older Adults View project
All content following this page was uploaded by Maria Kyrarini on 22 January 2021.
The Heracleia Human Centered Computing Laboratory, Department of Computer Science and Engineering,
The University of Texas at Arlington, Arlington, TX 76019, USA; fotios.lygerakis@mavs.uta.edu (F.L.);
akilesh.rajavenkatanarayanan@mavs.uta.edu (A.R.); christos.sevastopoulos@mavs.uta.edu (C.S.);
harishram.nambiappan@mavs.uta.edu (H.R.N.); kck8298@mavs.uta.edu (K.K.C.);
ashwin.rameshbabu@mavs.uta.edu (A.R.B.); joanne.mathew@mavs.uta.edu (J.M.); makedon@uta.edu (F.M.)
* Correspondence: maria.kyrarini@uta.edu
Abstract: In recent years, with the current advancements in Robotics and Artificial Intelligence (AI),
robots have the potential to support the field of healthcare. Robotic systems are often introduced
in the care of the elderly, children, and persons with disabilities, in hospitals, in rehabilitation and
walking assistance, and other healthcare situations. In this survey paper, the recent advances in
robotic technology applied in the healthcare domain are discussed. The paper provides detailed
information about state-of-the-art research in care, hospital, assistive, rehabilitation, and walking
assisting robots. The paper also discusses the open challenges healthcare robots face to be integrated
into our society.
Keywords: healthcare; robotics; care robots; nursing robots; hospital robots; assistive robots; rehabili-
tation robots; walking assisting robots
(i) present the recent state-of-the-art in this field and (ii) identify the open challenges and
future research directions. In this paper, robots have been grouped into the following five
categories; hospital, care, assistive, rehabilitation, and walking-assistant robots. For each
category of robots, the most recent work is presented and analyzed, both for commercially
available and research robots. The purpose of this review is to present the effectiveness of
the available healthcare robots, to address the open challenges they face, and to discuss
the future of robotic technology in healthcare. This article seeks to answer the following
questions:
1. What research has been performed towards developing robotics for healthcare?
2. How commercially available robots are used in healthcare?
3. What are the challenges the robots are facing in the -world environments?
These questions guide us to identify the current technology readiness level of health-
care robots and to identify the potential future research and development needed to
integrate robots into human-centric environments. The paper is organized as follows:
Section 2 discusses care robots, Section 3 presents the recent advances in hospital robots,
Section 4 discusses assistive robots, and Sections 5 and 6 describe the advances in rehabili-
tation and walking assisting robots, respectively. Section 7 discusses the open challenges
for robots in healthcare and concludes the paper.
2. Methodology
The selection of the papers consisted of three steps: (i) initial search in digital li-
braries, (ii) filtering based on defined criteria, and (iii) the final selection of research papers.
In the first step, a systematic search of IEEE Explore, ACM Digital Library, ScienceDirect,
and Google Scholar were performed to identify research papers that discuss robotic sys-
tems in healthcare. These databases were selected as they include a collection of indexed
publications, conference proceedings, and journals associated with robotics and they are
accessible within the library of the University of Texas at Arlington (UTA). The search was
restricted to publications in English between the years 2015 and 2020. These databases were
searched using the following search terms: robot or robotic system combined in all possible
ways with the words health, care, assistance, rehabilitation, or healthcare. Several robotic
platforms were discussed in more than one paper. After removing duplicates, the initial
list included 105 robotic platforms.
In the second step, this list of potentially relevant articles was reviewed by the authors
based to the following criteria: (1) the robotic platform should not be a surgical robot,
(2) the physical prototype of the robot exists, and (2) an evaluation with at least one human
subject has been conducted. The final set of 30 robotic platforms were selected, which
were categorized according to their application. Some robotic platforms were presented in
more than one category. Moreover, the COVID-19 pandemic [11] accelerated research in
hospital robotic platforms [12]. Therefore, online news were also searched to ensure the
latest robotic systems are included in this survey. Eight robotic platforms were added to
the final set. The final number of robotic platforms discussed in this paper is 38.
3. Care Robots
As is discussed by Wynsberghe [13], one cannot define exclusively the characteristics
of a Care Robot, as a robot can be named as such only by the way it is used, i.e., for
providing or assisting people in the process of patient care. Thus, this obscure definition
can include a variety of robots in this category with different hardware specifications
and capabilities. The vast majority of applications and surveys of Care Robots [14–21]
are oriented towards monitoring and assisting older adults both mentally (reminding,
supporting emotionally, motivating, etc.) and physically (handing over objects, delivering
items or assisting in dining) or diagnosing and assisting in the education of children with
mental disorders, such as autism.
Pepper [22] and Nao from Softbank Robotics (formerly Aldebaran Robotics) are social
robots with a potential application in care [23]. Pepper in Figure 1 is a four-foot semi-
Technologies 2021, 9, 8 3 of 26
humanoid robot with a wheeled base (instead of legs) on which sonar, laser, and bumper
sensors are mounted. There is a 10.1-inch touch display on its torso, and it has a total of
twenty Degrees of Freedom (DOF), including six DOF for each hand, two each for the
head and hips, one in the knees, and three in the base. The head hosts two RGB cameras,
a depth camera, a microphone, and a tactile sensor to perceive the world, and two speakers
where the ears would be on a human. A six-axis Inertial Measurement Unit (IMU) in
the base and two tactile sensors on its hands conclude the sensors Pepper is equipped
with. Pepper has been deployed successfully as a teaching assistant for children [24], as a
companion for elderly people [25], and as a coach to guide elderly people with psychiatric
disorders through rehabilitation recreational activities [26]. Recently, Carros et al. [27]
employed Pepper in a group setting scenario in an institutional care facility with older
adults for ten weeks and twenty sessions. The finding shows that the older adults enjoyed
the robotic interaction and their engagement was high during the sessions. However,
the study participants made it clear that they do not want robots to replace caregivers.
Moreover, for all participants, it was not always easy to understand what the robot was
doing. There were also some malfunctions of the robot, for example, robot applications did
not load in the intended timeframe, or the touchscreen sensitivity was unsatisfactory.
Nao, shown in Figure 3, is a 22.8-inch humanoid robot and the latest version (6th gen-
eration) has twenty-five DOF, with eleven of them in the legs and pelvis and the rest in the
upper body, similar to the upper body DOF of Pepper. Nao has two RGB cameras, nine
tactile sensors on its head and in its hands, four microphones, a sonar range finder, two
infrared emitters and receivers, one inertial board, and eight pressure sensors. Both robots
use the NAOqi operating system, which is open-source and supports many programming
languages, including Python and C++. There are numerous applications that these two
robots have been used for and many scenarios in which their interactions with humans
have been studied. Nao has been deployed as an assistant tutor for autistic children [31],
a physiotherapeutic assistive trainer for the elderly [32], a cognitive trainer [33], and a
healthcare assistant [34]. Recently, Qidwai et al. [35] employed NAO in a short study as a
teaching assistant for children with autism. The robot was programmed for a number of
teaching and therapeutic behaviors, such as singing, exercise, explaining, and playing with
children. The participants of the study were fifteen children with autism (ages 7–11) in a
school for children with special needs. The findings of the study were very encouraging
and show the potential of robots to enhance the learning process for children with autism.
One interesting observation from the study was that children who were afraid of the robot
from the very beginning did not perform well. In contrast, those who were fascinated by
the robot from the beginning performed the same activities easily and fluently. The authors
did not discuss any technical issues that may encounter with the robot. However, they
mention that the cost of the robot is an obstacle for its more frequent use by more children.
Zorabots [37] provides a universal software, called ZBOS, that bridges commercially
available robots with a platform-specific implementation for each robot. The range of
robotic platforms they support is wide, varying between android, mascot, mechanical
and non-humanoid robots [21]. Some of the robotic platforms that Zorabots supports are
Pepper, Nao, JAMES (Figure 4), and others. Therefore, there are a plethora of different
configurations and resulting DOF values, depending on the robotic platform that is under
consideration. In [38] the use of Nao is considered a robotic platform for therapeutic
and educational purposes for rehabilitation and special education in children with severe
physical disabilities. After a series of intervention sessions using Zorabots’ Nao (Zora
Nao for short), its high contribution was observed in achieving patients’ goals as set by
professionals regarding their movement and communication skills, and there was evidence
of positive impact in cognitive skills and attention. The value of Zora Nao was also
confirmed in a two-year study in fourteen nursing care organizations [39], where the role
of the care robot was to offer pleasure and entertainment or to stimulate the physical
activities of clients in residential care. In the first year of the study, the goal was to monitor
and evaluate how the Zora Nao robot is used daily. In the second year, the focus was on
evaluating whether the use of Zora Nao robots by care professionals can be extended to
other groups. For the evaluation of the robot, the authors collected data through interviews,
questionnaires, and observations. Several care professionals experienced several barriers
while using the robot. For example, long start-up time, software failures, short battery life,
and misunderstanding in speech recognition were some of the barriersmentioned.
that greets and directs customers, but to our knowledge, there is no evaluation study of its
use as a caregiver. Its predecessor, Care-O-Bot 3 [44] was evaluated as a caregiver robot
in two practical evaluations of one week each in a senior care facility [45]. Care-O-Bot 3
was used to bring water from a water cooler to the inhabitants and the inhabitants could
play the game “memory” on the robot’s tray. The overall experience of the personnel and
the inhabitants of the care facility towards the robot was very positive. However, the main
challenge for the robot during the delivery of water was the identification of the person
under poor lighting conditions and approaching them in a crowded sitting room with
several obstacles, such as chairs and walkers.
Lio [47] is another Care Robot developed by F&P Robotics, that has been tested in
nursing and retirement facilities, as well as for support at home. It consists of a four-
wheeled mobile base, on which there is mounted a P-Rob 3 [48] collaborative robotic
arm with six DOF and a payload weighing three to five kilograms. The base is equipped
with two RGB-D cameras and a wide fish-eye camera, two LiDARs, ultrasonic distance
sensors, and infrared sensors for floor detection. There are also different holders (for
bottles, cups, etc.), a screen, and speakers and omnidirectional microphones on the base,
as shown in Figure 6. Utilizing the aforementioned sensors in combination with state-of-
the-art mapping, localization, and perception algorithms, Lio can navigate freely in any
wheelchair-accessible room, recognize and greet people, identify objects, and receive voice
commands. The default gripper on Lio’s arm has two DOF and a camera mounted on it,
but more options are available, such as vacuum grippers or massage heads. Furthermore,
this Care Robot has many features that enhance its safe use while interacting with humans.
A compliant motion controller, fully covered in soft artificial leather material, collision
detection, and limited speed and force capabilities are the most important of these features.
In terms of privacy, Lio’s visual and navigation data are processed by integrated units
onboard and the data is anonymized and encoded before being stored. All the above
measures enable Lio to comply with ISO13482 standard safety requirements for personal
care robots. The use of Lio has already been evaluated in seven different health care
institutions in Germany and Switzerland. Lio took different roles throughout each day,
varying between delivering items like mail or blood samples, entertaining and motivating
patients, and reminding patients of important things by knocking on and opening their
doors to access their rooms. In another case, Lio was used at home to support a paraplegic
person with everyday tasks, e.g., opening and handing over bottles or assisting the person
with taking off a jacket for several weeks. Its functionality was hybrid, with some tasks
being carried out exclusively autonomously and other tasks with the control of the person.
Technologies 2021, 9, 8 7 of 26
Hobbit [50] (Figure 7), another care robot system designed for older adults, sets as its
main goals the detection and prevention of potential falls, as well as the proper handling of
emergency situations. It consists of a robotic platform equipped with a five-DOF robotic
arm and a gripper that can grasp objects of various shapes and structures. Regarding
its vision scheme, it has two cameras, a floor-parallel depth camera, and an RGB-Depth
camera (Microsoft Kinect), which facilitate the tasks of self-localization, obstacle detection
(juxtaposed with the use of eight ultrasound distance sensors), grasping and human-robot
interaction. Before executing any of its tasks, Hobbit’s operation relies heavily on the
success and precision of the following systems: Navigation, Human-Detection and Track-
ing, Gesture Recognition, Grasping, and Object Learning and Recognition. Subsequently,
Hobbit is able to achieve an array of actions such as “Call Hobbit”, cleaning the floor
from objects, learn/bring new objects to the user, and recognize a user’s instability. Hav-
ing been tested in a controlled environment, Hobbit’s functionality was characterized as
understandable and straightforward by the primary users involved.
We will next cover the RAMCIP [51] robot (Robotic Assistant for patients with Mild
Cognitive Impairment). RAMCIP [52] is a project intended to provide home assistance to
the elderly. With regards to the hardware integrated, the system encompasses a two-DOF
Mobile Platform, an elevation mechanism, a nine-DOF hand with a two-DOF wrist, and a
Technologies 2021, 9, 8 8 of 26
4. Hospital Robots
Recently, the industry has shown a growing interest in developing robots to assist
nurses in hospitals and clinics. The robotic nursing assistants are designed to function under
the direct control of nurses. A robotic nursing assistant will act as a teammate, helping
nurses by performing non-critical tasks, such as fetching supplies, giving nurses more time
to focus on critical tasks, such as caring for patients. Moreover, the COVID-19 pandemic
showed how vulnerable nurses are due to shortages of personal protective equipment [54].
Robots, on the other hand, are not vulnerable to viruses or other microorganisms and they
can assist during pandemics.
In recent years, there are a variety of commercially available robots that are currently
used in hospitals to help with transportation tasks. One example is Moxi [55] developed
by Diligent Robotics, which retrieves and brings supplies to hospital rooms and nursing
stations, delivers samples to laboratories, and removes soiled linen bags. Moxi, shown in
Figure 8, consists of a mobile base, a seven-DOF robotic arm with a two-finger gripper,
and sensors for environmental perception such as a camera and laser scanner. For the
last several years, Moxi has been tested in several hospitals around the state of Texas in
the USA [56]. Moxi manipulates objects known in advance, and the manipulation is very
structured. When Moxi is deployed in a hospital in the beginning, the robot learns the
locations of the supply rooms and the locations of where objects need to be delivered. Moxi
is able to navigate fully autonomously and safely by avoiding static and dynamic obstacles.
Technologies 2021, 9, 8 9 of 26
The trials at the Texas hospitals showed that Moxi was accepted not only by the nurses and
the clinical staff but also by the patients and their relatives [56]. However, Moxi did not
have any physical interactions with patients.
The robotics company ABB has demonstrated the concept of a dual-arm mobile
laboratory robot called YuMi to work alongside medical staff and lab workers [57]. Each
robotic arm of YuMi has seven DOF and it is equipped with a two-finger gripper. The Texas
Medical Center (TMC) Innovation Institute in Houston evaluates the YuMi robot in a
variety of logistic tasks, such as loading and unloading centrifuges, handling liquids,
preparing medicines, and picking up and sorting test tubes. The mobile robot TUG [58]
by Aethon helps with the delivery of medications, meals, supplies, tests, and waste. TUG
does not have any robotic arm, and it relies on the medical staff to load and unload
the objects it delivers. Similarly, the mobile robot Relay [59] by Swisslog Healthcare
delivers medications, lab samples, and other critical items. TUG and Relay are able
to navigate hallways autonomously and deliver objects from point A to point B while
avoiding obstacles.
However, there are some nursing robots whose main tasks are to assist patients.
In Japan, the RIKEN and Sumitomo Riko Company Limited have developed an experimen-
tal nursing robot, called ROBEAR [60], which is capable of lifting a patient from a bed into a
wheelchair or helping a patient to stand up. Moreover, Veebot Systems developed a needle
insertion robot [61]. The Veebot robot automates drawing blood and inserting Intravenous
therapy (IV). The Veebot robot can correctly identify the best vein with an accuracy of
83 percent [61]. Furthermore, the social robot Pepper [62] by SoftBank Robotics has been
used in several roles in healthcare, such as acting as receptionists in hospitals, conducting
survey research on patient satisfaction, and supporting staff in health monitoring [23].
Das et al. [63] proposed a sitter robot in a hospital environment that can monitor
patient vital signs, carry out conversational interactions with a patient, and inform the
nurse if abnormal patient conditions are detected. The authors developed a mobile app to
augment the verbal commands given to a robot through natural speech, camera, and other
native interfaces. The app enables the robot to assist the patient with decision-making
during pick and place tasks, monitor the patient’s health over time, and communicate
with the patient during sitting sessions. In the recent work of Das [64], an Adaptive
Robotic Nursing Assistant (ARNA) is presented. ARNA is a multipurpose robot that helps
nurses with day-to-day tasks, such as walking patients, fetching objects, and monitoring
patients’ health. ARNA was deployed and evaluated in a hospital environment at the
School of Nursing at the University of Louisville in Kentucky. Healthcare professionals
were involved in the evaluation of ARNA. Several metrics including completion time
and rate and level of user satisfaction were collected, and the results indicate an overall
positive response towards the use of a nursing robot. Moreover, ARNA was evaluated on
Technologies 2021, 9, 8 10 of 26
a cohort trial with 24 human subjects and results of this preliminary user study indicate
good usefulness and ease of use of the essential user sitter and walker characteristics
of the robot [65]. Recently, the Fraunhofer Institute for Manufacturing Engineering and
Automation developed a prototype robot, called DeKonBot, which consists of a mobile
base and a robotic arm. Its main task is to disinfect potentially contaminated surfaces,
such as door handles or elevator buttons [66]. However, both research platforms are still in
the development phase and have only been evaluated in laboratory settings.
5. Assistive Robots
People with paralysis have difficulties performing Activities of Daily Living (ADLs)
or working. In 2013, nearly 5.4 million people in the U.S (1.7% of the U.S. population) were
living with paralysis [67]. Stroke was the leading cause of paralysis, followed by Spinal
Cord Injury (SCI), Multiple Sclerosis (MS), and Cerebral Palsy (CP) [67]. In recent years,
with the advancements in robotics and Artificial Intelligence (AI), assistive robots have the
potential to provide care and support ADLs at home or at the workplace. In this paper,
we define assistive robots as robots that assist people with disabilities.
Various assistive robotic systems have been developed based on a Wheelchair Mounted
Robotic Arm (WARM). One example of an assistive robotic system is the FRIEND sys-
tem, which is an intelligent wheelchair-mounted manipulator [68,69]. The FRIEND has
passed through four generations; the first generations were focused on assisting people
with quadriplegia with ADLs, such as drinking and eating, while the fourth generation of
FRIEND focused on supporting these individuals in real-world environments. The FRIEND
IV [68] consists of a wheelchair platform, a seven-DOF robotic arm equipped with a two-
finger gripper and a hand camera, a chin joystick and head control interface, a stereo-camera
and a laser scanner. The FRIEND IV supported an individual with quadriplegia (end-user)
while the individual was working as a librarian retrospectively cataloging collections of
old books. The robotic system was responsible for autonomously recognizing and manipu-
lating the books, placing them in a specially designed book holder. The cataloging of the
books was done by the end-user using speech recognition. Moreover, whenever there was
a failure of the autonomous book manipulation, the end-user was able to take control using
an advanced Human-Machine Interface (HMI). With the end-user’s intervention, a success
rate of 95% was achieved.
However, most of the focus of assistive robots is to help people perform ADLs. Drink-
ing and eating tasks are considered highly prioritized tasks according to a survey of
potential end-users of robotic manipulators [70]. The Jaco 2 robotic arm from Kinova [71]
is a commercially assistive robot, which can be used for manipulation tasks by end-users.
It has two versions, one with six DOF and one with seven DOF, and it is equipped with
a two- or three-finger gripper. The Jaco 2 is widely used in research for assisting with
drinking and eating tasks as well as with manipulation tasks. Gordon et al. [72] developed
an adaptive robot-assisted feeding system. The system consists of a six-DOF Jaco 2 robotic
arm with a two-finger gripper. The gripper grabbed a fork equipped with a force-torque
sensor using a 3D-printed custom-built fork holder. The system uses an RGB-Depth camera
to identify the food on a plate, and then an online learning framework is developed for
successful bite acquisition. However, the algorithms converge after ten failures per food
item and all food items are discrete and solid. Bhattacharjee et al. [73] used the same robotic
system to explore the preferences of autonomy that users with mobility impairments have
in robot-assisted feeding. By evaluating the system with ten users with limited mobility,
the authors found that the users did not have a preference between partial or low autonomy
when controlling the robotic arm.
Goldau et al. [74] developed a system to autonomously assist people with severe
motor impairments during drinking. This work presents a robotic solution to enable
independent, straw-less drinking using a smart cup. The seven-DOF Jaco 2 robotic arm
with a three-finger gripper was used. An RGB-Depth camera was mounted on the robot’s
end-effector with a custom 3D-printed holder. The robot arm grasped the smart cup,
Technologies 2021, 9, 8 11 of 26
which consisted of a beak, two force sensors, and a Bluetooth module that transmitted
the force values to the system’s computer. The authors developed a vision-based robot
control system used to serve a drink, which handled the delivery of the grasped cup to
the user’s mouth based on camera input, and a robot force control system for the drinking
process, which enabled the user to control the process of tilting the cup based on the force
s/he applied. Two experimental studies have been conducted with mostly able-bodied
participants and one with quadriplegia, and the first results show a high user acceptance
rate and positive feedback. Shastha et al. [75] extended the robot force control system for
the drinking process by introducing Reinforcement Learning (RL) approaches and by using
a smart cup with one force sensor and without a beak. Five of the commonly used RL
algorithms were compared to find the best fit for the robotic drinking task, and the system
was tested in an experimental study. The preliminary results showed a high degree of
acceptance by the participants (mostly able-bodied and one with quadriplegia). However,
a user study with several participants suffering from quadriplegia is needed to evaluate
the proposed system.
The approaches discussed so far focus on partially autonomous approaches, where the
robot autonomously performs the task and the human is in the loop to provide high-level
inputs. In recent years, there is research on how to enable users with quadriplegia to
directly control the robotic arm by controlling the robot’s end-effector in three-dimensional
space. Several hands-free methods have been developed to enable direct robot control,
such as employing the use of head gestures [76,77], eye trackers [78], or Brain-Machine
Interfaces [79]. However, direct robot control is time-consuming. Kyrarini et al. [80]
developed a robot learning framework where a person with quadriplegia (end-user) was
able to ‘teach’ a seven-DOF Jaco 2 arm to serve a drink. The end-user was able to control the
translation and rotation of the robot’s end-effector and the gripper actuation through a state
machine and a hands-free Human-Robot Interface. The presented system was evaluated
by twelve able-bodied participants and one person suffering from quadriplegia, as shown
in Figure 9. The feedback from the end-user was very positive, because she was ‘in control’,
a feeling that she described as essential.
Figure 9. The Jaco 2 learns how to serve a drink from a person with quadriplegia via a head
gesture-based human–robot interface [80].
Besides drinking and eating tasks, assistive robots have also been used for other ADLs.
The Baxter humanoid robot from Rethink Robotics [81] has been used to assist a user with
dressing [82]. The proposed robot-assisted dressing system combines a tracking method
with hierarchical multitask control to minimize the force between the person and the robot.
The experimental results showed that the Baxter robot was able to provide personalized
dressing assistance in putting on a sleeveless jacket for users with simulated upper-body
Technologies 2021, 9, 8 12 of 26
impairments. Jevtic et al. in [83] developed a robotic assistant to help with shoe fitting.
The authors utilized Barrett’s seven-DOF WAM robotic arm [84] as a robotic assistant and
the user-provided verbal instructions to the robot. Additional modalities, such as users’
pointing gestures and adjustment of the robot’s shoe delivery position, were used to enable
a successful shoe fitting process. Other assistive robotic tasks that require a higher degree of
direct physical contact between the robot and the human have also been explored, such as
beard shaving [85], hair brushing [86], and bathing [87]. However, these robotic tasks are
still in a preliminary phase and more research is required to ensure safety and acceptability.
6. Rehabilitation Robots
Rehabilitation robots are a special robot type designed primarily for aiding humans
with physical impairments during the process of rehabilitation. There are many debilitating
motor disability diseases such as stroke, multiple sclerosis, injuries to the head and spinal
cord, and spina bifida [88,89]. Physiotherapy helps patients suffering from these infirmities,
as mentioned earlier, to regain their natural motor skills to the maximum extent possible.
Physiotherapists design a rehabilitation program for their patients where they try to
improve their patients’ motor skills [90]. Rehabilitation robots can assist physiotherapists
during the rehabilitation process. Therapists who use rehabilitation robots can enhance
gross motor skills better than conventional therapists. The robots can help patients perform
repetitive tasks and collect data from sensors for analysis [91,92]. In that way, the therapist
can readjust the rehabilitation process by better recognizing the different intricacies of
different patients. In this paper, we divide the rehabilitation robots into two subcategories,
upper-limb and lower-limb rehabilitation robots, which focus on helping humans with
upper body and lower body impairments, respectively.
focuses on elbow and forearm movement. Pang et al. [105] developed a tension mechanism
for an exoskeleton rehabilitation robot, shown in Figure 10, that utilizes the principle of
ensuring minimum driving torque.
Figure 10. Wearable upper limb rehabilitation robot with characteristics of the tension mecha-
nism [105].
In end-effector based rehabilitation robots, the work of research ranges from the
development of a new robot framework to the implementation and analysis of the re-
habilitation robot. Liu et al. [96] developed a two-link end-effector based robotic arm
which was tested by using it to perform line and circle tracking tasks for rehabilitation.
Ponomarenko et al. [106] designed an end-effector based rehabilitation robot that is elec-
tromagnetically powered and has five independent degrees of freedom in the shoulder
and elbow joints. There are also several other end-effector based upper body rehabilitation
robots, such as MIT MANUS which uses motor drive with impedance control, MIME of
Stanford University which uses motor drive with an EMG signal control and force con-
trol, and others [101]. Moreover, Barrett Medical has developed a commercially available
end-effector based rehabilitation robot called Burt [107]. Burt has multiple therapy modes
and enables game-based rehabilitation to engage patients and to increase the number of
repetitions reached in therapy sessions. A prior model of Burt Barrett Medical, called
the Barrett WAM arm [108], is used in research. It is a robotic arm available in two main
configurations, four-DOF and seven-DOF, and it can be controlled by force. The Barrett
WAM arm, shown in Figure 11, has been used as a part of systems for game-based upper
limb rehabilitation [109–111].
Apart from exoskeleton and end-effector based robots, there are also other forms of
implementation with respect to upper-limb rehabilitation robots. Mohamaddan et al. [95]
developed an upper limb rehabilitation robot for a home setting that uses an armrest and
Technologies 2021, 9, 8 14 of 26
a scissors lift mechanism. Ding et al. [98] proposed a robotic arm skate for upper body
rehabilitation equipped with a user interface that provides instructions and visual and
auditory feedback.
Although there are many research works in developing exoskeleton based rehabilita-
tion robots, many of the exoskeleton robots are either in the prototype stage or have not
yet been put up to evaluation with patients [97,101]. In addition to that, Rehmat et al. [103]
stated that only a few of the exoskeleton robots have been subjected to clinical trials.
While comparing between end-effector and exoskeleton upper body rehabilitation robots,
Meng et al. [100] states, both robot types have certain deficiencies such as the end-effector
robot having less degree of freedom, and the exoskeleton robot being heavy and not easy
to carry. The authors have also stated concerns about improving safety while dealing with
such robots and also improving gravity support in robots while patients are continuing
recovery [100].
Figure 12. MOPASS gait rehabilitation system from frontal (a) and back (b) sides [117].
Matjacic et al. [118] developed a Balance Assessment Robot for Treadmill walking
(BART) that assesses the balancing abilities of a patient during walking. BART consists of
a wide instrumented treadmill and an actuated pelvic link with a pelvis brace to deliver
perturbing force impulses at the pelvis level during treadmill walking [119]. Evaluation of
BART in a study of forty-one post-stroke and forty-three healthy subjects showed that a
substantial number of post-stroke subjects had reduced balancing capabilities. Therefore,
further rehabilitation for improving their balance is needed.
There are multiple tests to measure the effects of robot rehabilitation on patients
in comparison to traditional physiotherapy. Some of these tests include the Timed Up-
and-Go Test, the Fatigue Severity Scale (FSS), the Six-Minute Walk Test, and the Berg
Balance Score (BBS) to measure balance during the standing and sitting postures of patients.
Straudi et al. [120] performed rehabilitation using the Lokomat robot with sixteen subjects.
The subjects were divided into two groups of eight, and their Expanded Disability Status
Scale (EDSS) was calculated. The first group had an EDSS score of 5.8 ± 0.8, and the
second group had an EDSS score of 5.7 ± 0.7. The subjects who underwent twelve RAGT
sessions had an improvement in their gait speeds (0.07 m/s) and walking endurance
measurements (33.2 m), compared to the control group (−0.01 m/s and −0.7 m). Walking
endurance is a constant-load exercise test that measures the participant’s ability to sustain
a given sub-maximal exercise capacity, such as walking in patients’ cases. In another
recent study conducted by Zheng et al. [121] RAGT using Lokomat had a positive effect on
balance, as shown by the BBS score. The mean difference between the scores of patients
with RAGT and without RAGT is 4.25. In conclusion, robot-assisted rehabilitation with a
physiotherapist’s oversight yields better outcomes than traditional physiotherapy.
of the most commonly researched topics in medical robots. In this section, we will see some
of the common approaches to designing a robotic walking assistant.
Traditionally IWA design follows two standard approaches: cane-based walker or
walking-frames-based walker. The IWA’s standard functionalities are obstacle avoidance
and guidance, user intent prediction, user gait estimation, user velocity estimation, and fall
detection and prevention. These robotic devices are usually equipped with several sensors,
such as force-torque (FT) sensors, Laser Rangefinders (LRF), acoustic sensors, Charge-
Coupled Device (CCD) cameras for localization, and RGB-Depth cameras to be able to
perform these activities.
Cane-based walking assistants (CWA) closely resemble a cane, and owing to their
relatively small size, they are easily usable indoors and outdoors. The CWAs, even though
small, provide the required stability for people that are confident enough in their walking
and only need little to no support while doing so. Di et al. [127] proposed a novel intelligent
cane robot, which consists of an omnidirectional base, a universal joint connecting the cane
to the base, an LRF sensor, an FT sensor, and a touch panel. They propose a fall detection
and prevention scheme which uses an insole load sensor to be worn by the users inside
their shoes. The load sensor data, in combination with the LRF and FT sensor data, can
provide an impedance-based fall prevention controller for the robot. The above works do
not, however, discuss any user gait estimation techniques for the CWAs.
Yan et al. [128] proposed a different design, where the handle of the CWA is attached
to a ball joint to improve human-robot stability during walking and falling/stumbling
scenarios. The user must wear a sensor on their torso while using the walker to measure
posture for the human-robot system’s stability measurement. An LRF at the front is used
for obstacle detection and avoidance, and an LRF at the back is used for motion control and
fall detection. In their paper, Van Lam and Fujimoto [129] proposed a slim new CWA design
consisting of a single omnidirectional wheel. The robotic cane can maintain the user’s
balance by moving in the appropriate direction of the fall. Here, however, the authors
fail to consider important safety features like obstacle avoidance, guidance, and user
gait estimation.
Another type of IWA is the Walking-Frame based Walking Assistants (WWA), as shown
in Figure 13, which closely resembles a four-legged walker and provides more stability and
support. These walkers are usually large and are helpful in indoor scenarios. In addition
to the features provided by canes, walkers incorporate additional assistive features like
sit-to-stand assistance and rehabilitation exercise monitoring, to name a few. These addi-
tional features are the reasons why these walkers are more suited for long-term assistance
for the elderly in an assisted living facility or for patients who require rehabilitation and
additional support after surgery. Xu et al. [130], proposed a robotic walking helper that
consists of frames for the user to place their arms in for support. A Force Sensing Resistor
(FSR) sensor is placed inside each frame’s handle, which is used in combination with an
LRF sensor to analyze the user’s motion. A Support Vector Machine Algorithm is trained
to classify the user’s motion state to either the falling mode or normal walking mode, based
on which the robot is controlled.
A user-following smart walker, UFES smart walker, is proposed by Cifuentes et al. [131].
One Inertial Measurement Unit (IMU) sensor is mounted on the robot while the second
IMU sensor is mounted on the user’s pelvis region to detect pelvis rotation for angular
velocity computation. An LRF sensor is mounted below the knee-level to compute gait
parameters for estimating the robot’s linear velocity. The robot uses a combination of the
resulting data sets to follow in front of the user’s gait while supporting the user’s walking.
While the robot system considers the user’s gait for robot motion control, safety features
such as fall detection and prevention and obstacle avoidance are not discussed.
Technologies 2021, 9, 8 17 of 26
Figure 13. Walker frame based walking assistants (WWA)—the walking-aid robot [130].
An intelligent shared-control robot, Walbot, is proposed by Jiang et al. [132]. The robot
consists of an omnidirectional mobile base and a handrail designed to give maximum
motion capacity and support to the user. It consists of an LRF sensor in the front to localize
obstacles and an IMU sensor for measuring the slope of the ground. The user’s intent is
detected using a sensor-less force impedance controller by using an external disturbance
observer. The robot’s velocity is then computed to be compliant with the user’s intent.
The robot behaves passively by allowing the user to control the robot’s speed while actively
monitoring obstacles. The system does not include other features like gait parameter
estimation and fall detection and prevention.
Advanced assistive robotic systems have been proposed with state-of-the-art tech-
nologies and novel sensors in recent times. For instance, the assistive robotic system iWalk
that consists of an RGB-D RealSense camera for gait tracking, gait stability, and mobility
assessment, has been proposed by Chalvatzaki et al. [133,134]. iWalk employs an LRF
sensor for gait phase estimation and microphones and speakers for speech recognition and
voice feedback. The system is also able to monitor human activity during exercises and pro-
vide fall detection. The only drawback is that the system does not provide fall prevention.
Song et al. [135] propose a walking assistant robot that uses a passive-compliant control
to move the robot and an active obstacle avoidance controller. The robot consists of two
six-DOF robotic arms interlocked using a special mechanism to form a handrail. Two LRF
sensors are present on the robot. One sensor on the front estimates the gait parameters
while the other is used in the rear for obstacle location and avoidance. The robot’s motion
is computed using the user’s motion intentions, which are estimated using two sensors.
The gait parameters estimated using the front laser are used to determine the robot’s linear
velocity, while the angular velocity is determined using the FT sensor on the robot’s arm
and is based on the user’s steering intentions. Fall prevention is provided in case of freezing
gait, where the user may fall forward on the robot. During this scenario, the robot can bear
the user’s gait to prevent falls. Nevertheless, this system does not consider all fall cases
where the fall direction cannot be determined. Falls due to freezing gait are the only fall
types considered.
are expensive and the cost does not allow the ubiquitous use of robots. According to
ARK’s research [138], the cost of industrial robots has been declined in the last 15 years.
As machine learning and computer vision advance in recent years, this decline in costs
may cause an inflection point in the demand for robots. However, there are no statistics
about the cost of robots in healthcare, but a similar trend will follow as for industrial robots,
depending on the maturity of the technology and the demand for such products.
Table 1. Cont.
However, the maturity and readiness of the technology is still an open challenge. Wan
et al. [139] provided a recent review on the technological advantages in human–robot
interfaces, environmental perception and user monitoring, navigation, robust communica-
tion, Internet-of-Things, and Artificial Intelligence for mobile healthcare robot. Moreover,
Wan et al. identified several open research issues in intelligent communication, biosensors,
AI, and state-of-the-art deep learning algorithms that need to be addressed to achieve
robustness and safety. Moreover, safety standards require to be updated. For example,
the ISO 13482:2014 [140], which defines the safety requirements for personal care robots has
not been updated since 2014. Villaronga [141] discussed in detail the confusions that may
arise from ISO 13482:2014. Villaronga has included the following statement [141]: “ISO
13482:2014 classifies personal care robots from a technical perspective. That might be very useful
to create new robot applications, but not to give protection to consumers in legal terms. In fact,
compliance with some technical safety requirements does not necessarily imply compliance with the
entire existing legal framework.” The law requires to provide guidance not only for safety,
but also data protection, responsibility, transparency, autonomy, and dignity [142].
In terms of attribution of liability issues, the mechanical nature of healthcare robots
makes it impossible to assign them liability in case of malfunctioning or any other adverse
consequence related to their usage. This could make it extremely complicated to attribute
civil and criminal liability to natural and legal persons in relation to a harmful event caused
by the robot. By having AI involved along with this, accountability is made even more
complex and ensuring safety is one of the main challenges [143]. As robots and AI advance
at a fast speed, lawmakers should be more concerned to fill in the legal gaps.
One important concern is the privacy of the patients. Nursing care robots are usually
equipped with cameras that are capable of monitoring their patients, recording related
data, and communicating information wirelessly. Although such a feature can be useful
in safeguarding elderly patients, establishing virtual proximity with their family and
caregivers, it could also lead to a violation of the patients’ privacy. Without adequate
regulations, responsible corporate policies and protocols, these robots’ capabilities can
become a threat to the private lives of patients [143].
Moreover, there is the question of acceptability by the patients. Most of the robotic
systems discussed in this paper are evaluated by volunteers willing to interact with robots.
However, in a critical health-related situation, there is not only a question of whether the
patients would accept robotic assistance but also of whether their families and caregivers
would accept them. Caregivers see robots with a fear that they may replace them, as in other
industries robots are replacing humans. Therefore, the issue of employment is a widespread
concern. However, human-social contact is very crucial, especially when a patient needs
care. It should not be believed that robots can fulfill the emotional and physical needs of the
patients, especially the elderly [144]. For example, the study by Carros et al. [27] showed
clearly that the participants do not want robots to replace caregivers. Taylor et al. [145]
Technologies 2021, 9, 8 20 of 26
Author Contributions: Conceptualization, M.K.; materials related to hospital and assistive robots,
M.K.; materials related to care robots, F.L., C.S. and M.K.; materials related to walking assistant
robots A.R.; rehabilitation robots, H.R.N., K.K.C. and M.K.; investigation of the open challenges,
A.R.B. and M.K.; writing—original draft preparation, M.K., F.L., A.R., C.S., H.R.N., K.K.C. and A.R.B.;
writing—review and editing, M.K., J.M. and F.M.; Project administration, F.M.; Funding acquisition,
F.M. All authors have read and agreed to the published version of the manuscript.
Funding: This work is based upon research supported by the NSF under award numbers NSF-CHS
1565328 and NSF-PFI 1719031.
Institutional Review Board Statement: Not Applicable.
Informed Consent Statement: Not Applicable.
Data Availability Statement: Not Applicable.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Sander, M.; Oxlund, B.; Jespersen, A.; Krasnik, A.; Mortensen, E.L.; Westendorp, R.G.J.; Rasmussen, L.J. The challenges of human
population ageing. Age Ageing 2015, 44, 185–187. [CrossRef]
2. Chew, C.M. Caregiver Shortage Reaches Critical Stage. Provider (Wash. DC) 2017, 43, 14–16.
3. Poghosyan, L.; Clarke, S.P.; Finlayson, M.; Aiken, L.H. Nurse burnout and quality of care: Cross-national investigation in six
countries. Res. Nurs. Health 2010, 33, 288–298. [CrossRef]
Technologies 2021, 9, 8 21 of 26
4. Hall, L.H.; Johnson, J.; Watt, I.; Tsipa, A.; O’Connor, D.B. Healthcare staff wellbeing, burnout, and patient safety: A systematic
review. PLoS ONE 2016, 11, e0159015. [CrossRef]
5. Gross, J. Interviewing Roomba: A posthuman study of humans and robot vacuum cleaners. Explor. Media Ecol. 2020, 19, 285–297.
[CrossRef]
6. Caudwell, C.; Lacey, C. What do home robots want? The ambivalent power of cuteness in robotic relationships. Convergence
2020, 26, 956–968. [CrossRef]
7. Smids, J.; Nyholm, S.; Berkers, H. Robots in the Workplace: A Threat to—Or Opportunity for—Meaningful Work? Philos. Technol.
2020, 33, 503–522. [CrossRef]
8. Evjemo, L.D.; Gjerstad, T.; Grøtli, E.I.; Sziebig, G. Trends in Smart Manufacturing: Role of Humans and Industrial Robots in
Smart Factories. Curr. Robot. Rep. 2020, 1, 35–41. [CrossRef]
9. Leoste, J.; Heidmets, M. The impact of educational robots as learning tools on mathematics learning outcomes in basic education.
In Digital Turn in Schools—Research, Policy, Practice; Springer: Singapore, 2019; pp. 203–217.
10. Kwoh, Y.S.; Hou, J.; Jonckheere, E.A.; Hayati, S. A robot with improved absolute positioning accuracy for CT guided stereotactic
brain surgery. IEEE Trans. Biomed. Eng. 1988, 35, 153–160. [CrossRef]
11. Cucinotta, D.; Vanelli, M. WHO declares COVID-19 a pandemic. Acta Bio Medica Atenei Parm. 2020, 91, 157.
12. Khan, Z.H.; Siddique, A.; Lee, C.W. Robotics Utilization for Healthcare Digitization in Global COVID-19 Management. Int. J.
Environ. Res. Public Health 2020, 17, 3819. [CrossRef] [PubMed]
13. Wynsberghe, A. Designing Robots for Care: Care Centered Value-Sensitive Design. Sci. Eng. Ethics 2013, 19, 407–433.
Electronics 2020, 9, 367. [CrossRef] [PubMed]
14. Johansson-Pajala, R.M.; Thommes, K.; Hoppe, J.; Tuisku, O.; Hennala, L.; Pekkarinen, S.; Melkas, H.; Gustafsson, C. Care Robot
Orientation: What, Who and How? Potential Users’ Perceptions. Int. J. Soc. Robot. 2020. [CrossRef]
15. Vandemeulebroucke, T.; Dierckx de Casterlé, B.; Gastmans, C. The use of care robots in aged care: A systematic review of
argument-based ethics literature. Arch. Gerontol. Geriatr. 2018, 74, 15–25. [CrossRef] [PubMed]
16. Mordoch, E.; Osterreicher, A.; Guse, L.; Roger, K.; Thompson, G. Use of social commitment robots in the care of elderly people
with dementia: A literature review. Maturitas 2012, 74. [CrossRef] [PubMed]
17. Hashim, R.; Yussof, H. Feasibility of care robots for children with special needs: A review. In Proceedings of the 2017 IEEE
International Symposium on Robotics and Intelligent Sensors (IRIS), Ottawa, ON, Canada, 5–7 October 2017; pp. 379–382.
18. Dawe, J.; Sutherland, C.; Barco, A.; Broadbent, E. Can social robots help children in healthcare contexts? A scoping review. BMJ
Paediatr. Open 2019, 3. [CrossRef]
19. Logan, D.E.; Breazeal, C.; Goodwin, M.S.; Jeong, S.; O’Connell, B.; Smith-Freedman, D.; Heathers, J.; Weinstock, P. Social Robots
for Hospitalized Children. Pediatrics 2019, 144. [CrossRef]
20. Abou Allaban, A.; Wang, M.; Padır, T. A Systematic Review of Robotics Research in Support of In-Home Care for Older Adults.
Information 2020, 11, 75. [CrossRef]
21. Martinez-Martin, E.; Escalona, F.; Cazorla, M. Socially Assistive Robots for Older Adults and People with Autism: An Overview.
Electronics 2020, 9, 367. [CrossRef]
22. Pandey, A.K.; Gelin, R. A Mass-Produced Sociable Humanoid Robot: Pepper: The First Machine of Its Kind. IEEE Robot. Autom.
Mag. 2018, 25, 40–48. [CrossRef]
23. Pepper in Healthcare. Available online: https://www.softbankrobotics.com/emea/en/pepper-healthcare-ga (accessed on 5
January 2021).
24. Tanaka, F.; Isshiki, K.; Takahashi, F.; Uekusa, M.; Sei, R.; Hayashi, K. Pepper learns together with children: Development of an
educational application. In Proceedings of the 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids),
Seoul, Korea, 3–5 November 2015; pp. 270–275.
25. Yang, C.; Lu, M.; Tseng, S.; Fu, L. A companion robot for daily care of elders based on homeostasis. In Proceedings of the 2017
56th Annual Conference of the Society of Instrument and Control Engineers of Japan (SICE), Kanazawa, Japan, 19–22 September
2017; pp. 1401–1406.
26. Sato, M.; Yasuhara, Y.; Osaka, K.; Ito, H.; Dino, M.J.S.; Ong, I.L.; Zhao, Y.; Tanioka, T. Rehabilitation care with Pepper humanoid
robot: A qualitative case study of older patients with schizophrenia and/or dementia in Japan. Enferm. Clín. 2020, 30, 32–36.
[CrossRef] [PubMed]
27. Carros, F.; Meurer, J.; Löffler, D.; Unbehaun, D.; Matthies, S.; Koch, I.; Wieching, R.; Randall, D.; Hassenzahl, M.; Wulf, V.
Exploring Human-Robot Interaction with the Elderly: Results from a Ten-Week Case Study in a Care Home. In Proceedings of
the 2020 CHI Conference on Human Factors in Computing Systems (CHI ’20), Honolulu, HI, USA, 25–30 April 2020; pp. 1–12.
[CrossRef]
28. IIT NCSR Demokritos Presentation at Thessaloniki International Fair. Available online: https://www.iit.demokritos.gr/ (accessed
on 12 January 2020).
29. Costa, A.; Martinez-Martin, E.; Cazorla, M.; Julian, V. PHAROS—PHysical assistant RObot system. Sensors 2018, 18, 2633.
[CrossRef] [PubMed]
30. Martinez-Martin, E.; Costa, A.; Cazorla, M. PHAROS 2.0—A PHysical Assistant RObot System Improved. Sensors 2019, 19, 4531.
[CrossRef] [PubMed]
31. Yousif, J.; Yousif, M. Humanoid Robot as Assistant Tutor for Autistic Children. Int. J. Comput. Appl. Sci. 2020, 8, 8–13.
Technologies 2021, 9, 8 22 of 26
32. Bhuvaneswari, P.T.V.; Vignesh, S.; Papitha, S.; Dharmarajan, R.S. Humanoid robot based physiotherapeutic assistive trainer for
elderly health care. In Proceedings of the 2013 International Conference on Recent Trends in Information Technology (ICRTIT),
Chennai, India, 25–27 July 2013; pp. 163–168.
33. Tsiakas, K.; Abujelala, M.; Makedon, F. Task engagement as personalization feedback for socially-assistive robots and cognitive
training. Technologies 2018, 6, 49. [CrossRef]
34. Vital, J.P.M.; Couceiro, M.S.; Rodrigues, N.M.M.; Figueiredo, C.M.; Ferreira, N.M.F. Fostering the NAO platform as an elderly care
robot. In Proceedings of the 2013 IEEE 2nd International Conference on Serious Games and Applications for Health (SeGAH),
Vilamoura, Portugal, 2–3 May 2013; pp. 1–5.
35. Qidwai, U.; Kashem, S.B.A.; Conor, O. Humanoid Robot as a Teacher’s Assistant: Helping Children with Autism to Learn Social
and Academic Skills. J. Intell. Robot. Syst. 2019, 98, 759–770. [CrossRef]
36. Softbank Robotics Europe/CC BY-SA. Available online: https://creativecommons.org/licenses/by-sa/4.0 (accessed on 12
January 2020).
37. Zora Robots. Available online: https://www.zorarobotics.be (accessed on 12 January 2020).
38. Van den Heuvel, R.; Lexis, M.; Witte, L. Robot ZORA in rehabilitation and special education for children with severe physical
disabilities: A pilot study. Int. J. Rehabil. Res. 2017, 40, 1. [CrossRef]
39. Huisman, C.; Kort, H.S.M. Two-Year Use of Care Robot Zora in Dutch Nursing Homes: An Evaluation Study. Healthcare 2019,
7, 31. [CrossRef]
40. Robots.nu. Available online: https://robots.nu/en/robot/james-robot (accessed on 14 January 2021).
41. Kittmann, R.; Fröhlich, T.; Schäfer, J.; Reiser, U.; Weisshardt, F.; Haug, A. Let me Introduce Myself: I am Care-O-bot 4, a Gentleman
Robot. In Mensch und Computer 2015—Proceedings; Diefenbach, S., Henze, N., Pielot, M., Eds.; De Gruyter Oldenbourg: Berlin,
Germany, 2015; pp. 223–232.
42. Kort, H.; Huisman, C. Care Robot ZORA in Dutch Nursing Homes; An Evaluation Study. Stud. Health Technol. Inform. 2017,
242, 527–534. [CrossRef]
43. Lüdtke, M. Mathias (Fraunhofer IPA): Service robots with CAN-driven devices. CAN Newsletters 2017, 1, 30–33.
44. Care-O-bot 3. Available online: https://www.care-o-bot.de/en/care-o-bot-3.html (accessed on 14 January 2021).
45. Jacobs, T.; Graf, B. Practical evaluation of service robots for support and routine tasks in an elderly care facility. In Proceedings of
the 2012 IEEE Workshop on Advanced Robotics and its Social Impacts (ARSO), Munich, Germany, 21–23 May 2012; pp. 46–49.
46. Care-O-Bot. Available online: https://www.care-o-bot.de/en/care-o-bot-4/download/images.html (accessed on 14 January
2021).
47. Miseikis, J.; Caroni, P.; Duchamp, P.; Gasser, A.; Marko, R.; Miseikiene, N.; Zwilling, F.; Castelbajac, C.; Eicher, L.; Frueh, M.;
et al. Lio—A Personal Robot Assistant for Human-Robot Interaction and Care Applications. IEEE Robot. Autom. Lett. 2020,
5, 5339–5346. [CrossRef]
48. P-Rob. Available online: https://www.fp-robotics.com/en/p-rob/ (accessed on 14 January 2021).
49. F&P Robotics. Available online: https://www.fp-robotics.com/ (accessed on 14 January 2021).
50. Fischinger, D.; Einramhof, P.; Papoutsakis, K.; Wohlkinger, W.; Mayer, P.; Panek, P.; Hofmann, S.; Koertner, T.; Weiss, A.; Argyros,
A.; et al. Hobbit, a care robot supporting independent living at home: First prototype and lessons learned. Robot. Auton. Syst.
2016, 75, 60–78. [CrossRef]
51. Kostavelis, I.; Giakoumis, D.; Malasiotis, S.; Tzovaras, D. Ramcip: Towards a robotic assistant to support elderly with mild
cognitive impairments at home. In International Symposium on Pervasive Computing Paradigms for Mental Health; Springer:
Berlin/Heidelberg, Germany, 2015; pp. 186–195.
52. RAMCIP—Robotic Assistant for MCI Patients at home. Available online: https://ramcip-project.eu/ (accessed on 15 October
2020).
53. Koceski, S.; Koceska, N. Evaluation of an assistive telepresence robot for elderly healthcare. J. Med. Syst. 2016, 40, 121. [CrossRef]
[PubMed]
54. Ranney, M.L.; Griffeth, V.; Jha, A.K. Critical supply shortages—The need for ventilators and personal protective equipment
during the Covid-19 pandemic. N. Engl. J. Med. 2020, 382, e41. [CrossRef] [PubMed]
55. Moxi Helps Hospitals and Clinical Staff. Available online: https://diligentrobots.com/moxi (accessed on 17 September 2020).
56. Ackerman, E. How Diligent’s Robots Are Making a Difference in Texas Hospitals. Available online: https://spectrum.ieee.org/
automaton/robotics/medical-robots/how-diligents-robots-are-making-a-difference-in-texas-hospitals (accessed on 9 October
2019).
57. ABB Demonstrates Concept of Mobile Laboratory Robot for Hospital of the Future. Available online: https://new.abb.com/
news/detail/37279/hospital-of-the-future (accessed on 31 March 2020).
58. Tug. Change Healthcare. Mobile Robots Improve Patient Care, Employee Satisfaction, Safety, Productivity and More. Available
online: https://aethon.com/mobile-robots-for-healthcare/ (accessed on 17 September 2020).
59. Relay Autonomous Service Robot for Hospitals. Available online: https://www.swisslog-healthcare.com/en-us/products-and-
services/transport-automation/relay-autonomous-service-robot/ (accessed on 17 September 2020).
60. The Strong Robot with the Gentle Touch. Available online: https://www.riken.jp/en/news_pubs/research_news/pr/2015/201
50223_2/ (accessed on 23 February 2015).
Technologies 2021, 9, 8 23 of 26
61. Perry, T.S. Profile: Veebot. Making a Robot That Can Draw Blood Faster and More Safely Than a Human Can. Available online:
https://spectrum.ieee.org/robotics/medical-robots/profile-veebot (accessed on 26 July 2013).
62. Healthcare. Embrace Robotics in e-Healthcare and Create Digital Patient-Enablement Journey. Available online: https://www.
softbankrobotics.com/emea/en/industries/healthcare (accessed on 17 September 2020).
63. Das, S.K.; Sahu, A.; Popa, D.O. Mobile app for human-interaction with sitter robots. In Smart Biomedical and Physiological Sensor
Technology XIV; International Society for Optics and Photonics: Anaheim, CA, USA, 2017; Volume 10216, p. 102160D.
64. Das, S.K. Adaptive Physical Human-Robot Interaction (PHRI) with a Robotic Nursing Assistant. Ph.D. Thesis, University of
Louisville, Louisville, Kentucky, 2019.
65. Abubakar, S.; Das, S.K.; Robinson, C.; Saadatzi, M.N.; Logsdon, M.C.; Mitchell, H.; Chlebowy, D.; Popa, D.O. ARNA, a Service
robot for Nursing Assistance: System Overview and User Acceptability. In Proceedings of the 2020 IEEE 16th International
Conference on Automation Science and Engineering (CASE), Hong Kong, China, 20–21 August 2020; pp. 1408–1414.
66. DeKonBot—Mobile Robot for Disinfecting Potentially Contaminated Surfaces. Available online: https://www.youtube.com/
watch?v=lBj6P9lfXDM&feature=emb_title&ab_channel=FraunhoferIPA (accessed on 26 September 2020).
67. Armour, B.S.; Courtney-Long, E.A.; Fox, M.H.; Fredine, H.; Cahill, A. Prevalence and causes of paralysis—United States, 2013.
Am. J. Public Health 2016, 106, 1855–1857. [CrossRef]
68. Graser, A.; Heyer, T.; Fotoohi, L.; Lange, U.; Kampe, H.; Enjarini, B.; Heyer, S.; Fragkopoulos, C.; Ristic-Durrant, D. A supportive
friend at work: Robotic workplace assistance for the disabled. IEEE Robot. Autom. Mag. 2013, 20, 148–159. [CrossRef]
69. Martens, C.; Prenzel, O.; Gräser, A. The rehabilitation robots FRIEND-I & II: Daily life independency through semi-autonomous
task-execution. Rehabil. Robot. 2007, 1, 137–162.
70. Chung, C.S.; Wang, H.; Cooper, R.A. Functional assessment and performance evaluation for assistive robotic manipulators:
Literature review. J. Spinal Cord Med. 2013, 36, 273–289. [CrossRef]
71. KINOVA JACO Assistive Robotic Arm. Available online: https://www.kinovarobotics.com/en/products/assistive-technologies/
kinova-jaco-assistive-robotic-arm (accessed on 8 October 2020).
72. Gordon, E.K.; Meng, X.; Barnes, M.; Bhattacharjee, T.; Srinivasa, S.S. Adaptive Robot-Assisted Feeding: An Online Learning
Framework for Acquiring Previously-Unseen Food Items. arXiv 2019, arXiv:1908.07088.
73. Bhattacharjee, T.; Gordon, E.K.; Scalise, R.; Cabrera, M.E.; Caspi, A.; Cakmak, M.; Srinivasa, S.S. Is More Autonomy Always
Better? Exploring Preferences of Users with Mobility Impairments in Robot-assisted Feeding. In Proceedings of the 2020
ACM/IEEE International Conference on Human-Robot Interaction, Cambridge, UK, 23–26 March 2020; pp. 181–190.
74. Goldau, F.F.; Shastha, T.K.; Kyrarini, M.; Gräser, A. Autonomous multi-sensory robotic assistant for a drinking task. In Proceedings
of the 2019 IEEE 16th International Conference on Rehabilitation Robotics (ICORR), Toronto, ON, Canada, 24–28 June 2019;
pp. 210–216.
75. Kumar Shastha, T.; Kyrarini, M.; Gräser, A. Application of Reinforcement Learning to a Robotic Drinking Assistant. Robotics
2020, 9, 1. [CrossRef]
76. Haseeb, M.A.; Kyrarini, M.; Jiang, S.; Ristic-Durrant, D.; Gräser, A. Head gesture-based control for assistive robots. In Proceedings
of the 11th PErvasive Technologies Related to Assistive Environments Conference, Corfu, Greece, 26–29 June 2018; pp. 379–383.
77. Rudigkeit, N.; Gebhard, M. AMiCUS 2.0—System Presentation and Demonstration of Adaptability to Personal Needs by the
Example of an Individual with Progressed Multiple Sclerosis. Sensors 2020, 20, 1194. [CrossRef] [PubMed]
78. Alsharif, S.; Kuzmicheva, O.; Gräser, A. Gaze gesture-based human robot interface. In Technische Unterstützungssysteme, Die Die
Menschen Wirklich Wollen; Helmut-Schmidt-Universität: Hamburg, Germany, 2016; p. 339.
79. Mindermann, B. Untersuchung Eines Hybriden Brain-Computer Interfaces (BCIs) zur Optimalen Auslegung als Mensch-
Maschine-Schnittstelle. Ph.D. Thesis, Universität Bremen, Bremen, Germany, 2018.
80. Kyrarini, M.; Zheng, Q.; Haseeb, M.A.; Gräser, A. Robot Learning of Assistive Manipulation Tasks by Demonstration via Head
Gesture-based Interface. In Proceedings of the 2019 IEEE 16th International Conference on Rehabilitation Robotics (ICORR),
Toronto, ON, Canada, 24–28 June 2019; pp. 1139–1146.
81. Rethink Robotics. Available online: https://www.rethinkrobotics.com/ (accessed on 8 October 2020).
82. Zhang, F.; Cully, A.; Demiris, Y. Probabilistic real-time user posture tracking for personalized robot-assisted dressing. IEEE Trans.
Robot. 2019, 35, 873–888. [CrossRef]
83. Jevtić, A.; Valle, A.F.; Alenyà, G.; Chance, G.; Caleb-Solly, P.; Dogramadzi, S.; Torras, C. Personalized robot assistant for support
in dressing. IEEE Trans. Cogn. Dev. Syst. 2018, 11, 363–374. [CrossRef]
84. Barrett Advanced Robotics—The WAM Arm. Available online: https://advanced.barrett.com/wam-arm-1 (accessed on 9
October 2020).
85. Oladayo, A.S.; Assal, F.S.; El-Hussieny, H. Towards Development of an Autonomous Robotic System for Beard Shaving Assistance
for Disabled People. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari,
Italy, 6–9 October 2019; pp. 3435–3440.
86. Kahn, J. Hair-Brushing Robot Shows How Artificial Intelligence May Help the Disabled. Available online: https://fortune.com/
2019/12/11/robot-hair-brushing-elderly/ (accessed on 11 December 2019).
87. Zlatintsi, A.; Dometios, A.; Kardaris, N.; Rodomagoulakis, I.; Koutras, P.; Papageorgiou, X.; Maragos, P.; Tzafestas, C.S.;
Vartholomeos, P.; Hauer, K.; et al. I-Support: A robotic platform of an assistive bathing robot for the elderly population. Robot.
Auton. Syst. 2020, 126, 103451. [CrossRef]
Technologies 2021, 9, 8 24 of 26
88. Rajavenkatanarayanan, A.; Kanal, V.; Tsiakas, K.; Calderon, D.; Papakostas, M.; Abujelala, M.; Galib, M.; Ford, J.; Wylie, G.;
Makedon, F. A Survey of Assistive Technologies for Assessment and Rehabilitation of Motor Impairments in Multiple Sclerosis.
Multimodal Technol. Interact. 2019, 3, 6. [CrossRef]
89. Ármannsdóttir, A.L.; Beckerle, P.; Moreno, J.C.; van Asseldonk, E.H.; Manrique-Sancho, M.T.; Del-Ama, A.J.; Veneman, J.F.;
Briem, K. Assessing the Involvement of Users During Development of Lower Limb Wearable Robotic Exoskeletons: A Survey
Study. Hum. Factors 2020, 62, 351–364. [CrossRef]
90. Wu, J.; Gao, J.; Song, R.; Li, R.; Li, Y.; Jiang, L. The design and control of a 3DOF lower limb rehabilitation robot. Mechatronics
2016, 33, 13–22. [CrossRef]
91. Eiammanussakul, T.; Sangveraphunsiri, V. A lower limb rehabilitation robot in sitting position with a review of training activities.
J. Healthc. Eng. 2018, 2018. [CrossRef]
92. Mohanta, J.K.; Mohan, S.; Deepasundar, P.; Kiruba-Shankar, R. Development and control of a new sitting-type lower limb
rehabilitation robot. Comput. Electr. Eng. 2018, 67, 330–347. [CrossRef]
93. Fikri, M.A.; Abdullah, S.C.; Ramli, M.H.M. Arm exoskeleton for rehabilitation following stroke by learning algorithm prediction.
Procedia Comput. Sci. 2014, 42, 357–364. [CrossRef]
94. Tao, J.; Yu, S. Developing Conceptual PSS Models of Upper Limb Exoskeleton based Post-stroke Rehabilitation in China. Procedia
CIRP 2019, 80, 750–755. [CrossRef]
95. Mohamaddan, S.; Jamali, A.; Abidin, A.S.Z.; Jamaludin, M.S.; Abd Majid, N.A.; Ashari, M.F.; Hazmi, H. Development of upper
limb rehabilitation robot device for home setting. Procedia Comput. Sci. 2015, 76, 376–380. [CrossRef]
96. Liu, Y.; Li, C.; Ji, L.; Bi, S.; Zhang, X.; Huo, J.; Ji, R. Development and implementation of an end-effector upper limb rehabilitation
robot for hemiplegic patients with line and circle tracking training. J. Healthc. Eng. 2017, 2017, 4931217. [CrossRef] [PubMed]
97. Maciejasz, P.; Eschweiler, J.; Gerlach-Hahn, K.; Jansen-Toy, A.; Leonhardt, S. A survey on robotic devices for upper limb
rehabilitation. J. Neuroeng. Rehabil. 2014, 11, 3. [CrossRef] [PubMed]
98. Ding, Y.; Tay, E.H. An Interactive Training System for Upper Limb Rehabilitation Using Visual and Auditory Feedback.
In Proceedings of the 2019 5th International Conference on Robotics and Artificial Intelligence, Singapore, 22–24 November 2019;
pp. 54–58.
99. Rijanto, E.; Adiwiguna, E.; Rozaqi, L.; Sadono, A.P.; Nugraha, M.H. Experimental Performance Evaluation of Computer Vision
for an Upper Limbs Rehabilitation Robot. In Proceedings of the 2019 International Conference on Computer, Control, Informatics
and its Applications (IC3INA), Tangerang, Indonesia, 23–24 October 2019; pp. 59–63.
100. Meng, Q.; Xie, Q.; Yu, H. Upper-Limb Rehabilitation Robot: State of the Art and Existing Problems. In Proceedings of the
12th International Convention on Rehabilitation Engineering and Assistive Technology, Białystok, Poland, 28–30 June 2018;
pp. 155–158.
101. Zhang, K.; Chen, X.; Liu, F.; Tang, H.; Wang, J.; Wen, W. System framework of robotics in upper limb rehabilitation on poststroke
motor recovery. Behav. Neurol. 2018, 2018, 6737056. [CrossRef]
102. Lee, S.H.; Park, G.; Cho, D.Y.; Kim, H.Y.; Lee, J.Y.; Kim, S.; Park, S.B.; Shin, J.H. Comparisons between end-effector and exoskeleton
rehabilitation robots regarding upper extremity function among chronic stroke patients with moderate-to-severe upper limb
impairment. Sci. Rep. 2020, 10, 1806. [CrossRef]
103. Rehmat, N.; Zuo, J.; Meng, W.; Liu, Q.; Xie, S.Q.; Liang, H. Upper limb rehabilitation using robotic exoskeleton systems:
A systematic review. Int. J. Intell. Robot. Appl. 2018, 2, 283–295. [CrossRef]
104. Bouteraa, Y.; Abdallah, I.B. Exoskeleton robots for upper-limb rehabilitation. In Proceedings of the 2016 13th International
Multi-Conference on Systems, Signals & Devices (SSD), Leipzig, Germany, 21–24 March 2016; pp. 1–6.
105. Pang, Z.; Wang, T.; Wang, Z.; Yu, J.; Sun, Z.; Liu, S. Design and Analysis of a Wearable Upper Limb Rehabilitation Robot with
Characteristics of Tension Mechanism. Appl. Sci. 2020, 10, 2101. [CrossRef]
106. Ponomarenko, Y.; Aubakir, B.; Hussain, S.; Shintemirov, A. An end-effector based upper-limb rehabilitation robot: Pre-
liminary mechanism design. In Proceedings of the 2014 10th France-Japan/8th Europe-Asia Congress on Mecatronics
(MECATRONICS2014-Tokyo), Tokyo, Japan, 27–29 November 2014; pp. 168–172.
107. Burt—The User-Friendly Robot. Available online: https://medical.barrett.com/home (accessed on 18 October 2020).
108. The WAM® Arm. Available online: https://advanced.barrett.com/wam-arm-1#:~:text=The%20WAM%C2%AE%20Arm%20is,
mechanical%20force%20or%20torque%20sensors (accessed on 20 October 2020).
109. Lioulemes, A.; Theofanidis, M.; Kanal, V.; Tsiakas, K.; Abujelala, M.; Collander, C.; Townsend, W.B.; Boisselle, A.; Makedon, F.
MAGNI Dynamics: A vision-based kinematic and dynamic upper-limb model for intelligent robotic rehabilitation. Int. J. Biomed.
Biol. Eng. World Acad. Sci. Eng. Technol. 2017, 11, 158–167.
110. Rajavenkatanarayanan, A.; Kanal, V.; Tsiakas, K.; Brady, J.; Calderon, D.; Wylie, G.; Makedon, F. Towards a robot-based
multimodal framework to assess the impact of fatigue on user behavior and performance: A pilot study. In Proceedings of the
12th ACM International Conference on PErvasive Technologies Related to Assistive Environments, Rhodes, Greece, 5–7 June
2019; pp. 493–498.
111. Kanal, V.; Brady, J.; Nambiappan, H.; Kyrarini, M.; Wylie, G.; Makedon, F. Towards a serious game based human-robot framework
for fatigue assessment. In Proceedings of the 13th ACM International Conference on PErvasive Technologies Related to Assistive
Environments, Corfu, Greece, 30 June–3 July 2020; pp. 1–6.
Technologies 2021, 9, 8 25 of 26
112. Hobbs, B.; Artemiadis, P. A Review of Robot-Assisted Lower-Limb Stroke Therapy: Unexplored Paths and Future Directions in
Gait Rehabilitation. Front. Neurorobot. 2020, 14, 1–16. [CrossRef]
113. Zeilig, G.; Weingarden, H.; Zwecker, M.; Dudkiewicz, I.; Bloch, A.; Esquenazi, A. Safety and tolerance of the ReWalk™ exoskeleton
suit for ambulation by people with complete spinal cord injury: A pilot study. J. Spinal Cord Med. 2012, 35, 96–101. [CrossRef]
[PubMed]
114. ReWalk More Than Walking. Available online: https://rewalk.com/ (accessed on 8 December 2020).
115. Lokomat. Available online: https://www.hocoma.com/ (accessed on 14 January 2021).
116. Kuzmicheva, O.; Focke Martinez, S.; Krebs, U.; Spranger, M.; Moosburner, S.; Wagner, B.; Graser, A. Overground robot based
gait rehabilitation system MOPASS—Overview and first results from usability testing. In Proceedings of the IEEE International
Conference on Robotics and Automation, Stockholm, Sweden, 16–21 May 2016; pp. 3756–3763. [CrossRef]
117. Wang, X. Machine Learning for Gait Classification. Ph.D. Thesis, Universität Bremen, Bremen, Germany, 2017.
118. Matjačić, Z.; Zadravec, M.; Olenšek, A. Feasibility of robot-based perturbed-balance training during treadmill walking in a
high-functioning chronic stroke subject: A case-control study. J. Neuroeng. Rehabil. 2018, 15, 32. [CrossRef] [PubMed]
119. Zadravec, M.; Olenšek, A.; Rudolf, M.; Bizovičar, N.; Goljar, N.; Matjacic, Z. Assessment of dynamic balancing responses
following perturbations during slow walking in relation to clinical outcome measures for high-functioning post-stroke subjects.
Rev. Sq. 2020. [CrossRef] [PubMed]
120. Straudi, S.; Benedetti, M.G.; Venturini, E.; Manca, M.; Foti, C.; Basaglia, N. Does robot-assisted gait training ameliorate gait
abnormalities in multiple sclerosis? A pilot randomized-control trial. NeuroRehabilitation 2013, 33, 555–563. [CrossRef] [PubMed]
121. Zheng, Q.X.; Ge, L.; Wang, C.C.; Ma, Q.S.; Liao, Y.T.; Huang, P.P.; Wang, G.D.; Xie, Q.L.; Rask, M. Robot-assisted therapy for
balance function rehabilitation after stroke: A systematic review and meta-analysis. Int. J. Nurs. Stud. 2019, 95, 7–18. [CrossRef]
[PubMed]
122. Viswanathan, P.; Zambalde, E.P.; Foley, G.; Graham, J.L.; Wang, R.H.; Adhikari, B.; Mackworth, A.K.; Mihailidis, A.; Miller, W.C.;
Mitchell, I.M. Intelligent wheelchair control strategies for older adults with cognitive impairment: User attitudes, needs, and
preferences. Auton. Robot. 2017, 41, 539–554. [CrossRef]
123. Moon, I.; Lee, M.; Ryu, J.; Mun, M. Intelligent robotic wheelchair with EMG-, gesture-, and voice-based interfaces. In Proceedings
of the 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No. 03CH37453), Las Vegas,
NV, USA, 27–31 October 2003; Volume 4, pp. 3453–3458.
124. Huo, W.; Mohammed, S.; Moreno, J.C.; Amirat, Y. Lower limb wearable robots for assistance and rehabilitation: A state of the art.
IEEE Syst. J. 2014, 10, 1068–1081. [CrossRef]
125. Varol, H.A.; Sup, F.; Goldfarb, M. Multiclass real-time intent recognition of a powered lower limb prosthesis. IEEE Trans. Biomed.
Eng. 2009, 57, 542–551. [CrossRef]
126. Yan, Q.; Huang, J.; Tao, C.; Chen, X.; Xu, W. Intelligent mobile walking-aids: Perception, control and safety. Adv. Robot. 2020,
34, 2–18. [CrossRef]
127. Di, P.; Hasegawa, Y.; Nakagawa, S.; Sekiyama, K.; Fukuda, T.; Huang, J.; Huang, Q. Fall Detection and Prevention Control Using
Walking-Aid Cane Robot. IEEE/ASME Trans. Mechatron. 2016, 21, 625–637. [CrossRef]
128. Yan, Q.; Huang, J.; Luo, Z. Human-robot coordination stability for fall detection and prevention using cane robot. In Proceedings
of the 2016 International Symposium on Micro-NanoMechatronics and Human Science (MHS), Nagoya, Japan, 28–30 November
2016; pp. 1–7.
129. Van Lam, P.; Fujimoto, Y. A Robotic Cane for Balance Maintenance Assistance. IEEE Trans. Ind. Inform. 2019, 15, 3998–4009.
[CrossRef]
130. Xu, W.; Huang, J.; Cheng, L. A novel coordinated motion fusion-based walking-aid robot system. Sensors 2018, 18, 2761.
[CrossRef] [PubMed]
131. Cifuentes, C.A.; Rodriguez, C.; Frizera-Neto, A.; Bastos-Filho, T.F.; Carelli, R. Multimodal Human–Robot Interaction for
Walker-Assisted Gait. IEEE Syst. J. 2016, 10, 933–943. [CrossRef]
132. Jiang, S.; Lin, C.; Huang, K.; Song, K. Shared Control Design of a Walking-Assistant Robot. IEEE Trans. Control Syst. Technol. 2017,
25, 2143–2150. [CrossRef]
133. Chalvatzaki, G.; Koutras, P.; Hadfield, J.; Papageorgiou, X.S.; Tzafestas, C.S.; Maragos, P. Lstm-based network for human
gait stability prediction in an intelligent robotic rollator. In Proceedings of the 2019 International Conference on Robotics and
Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 4225–4232.
134. Chalvatzaki, G.; Papageorgiou, X.S.; Tzafestas, C.S. Towards a user-adaptive context-aware robotic walker with a pathological
gait assessment system: First experimental study. In Proceedings of the 2017 IEEE/RSJ international conference on intelligent
robots and systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5037–5042.
135. Song, K.T.; Jiang, S.Y.; Wu, S.Y. Safe guidance for a walking-assistant robot using gait estimation and obstacle avoidance.
IEEE/ASME Trans. Mechatron. 2017, 22, 2070–2078. [CrossRef]
136. Lokomat Therapy—Robotic Treadmill May Enable Some Patients with Spinal Cord Injuries or Stroke to Learn to Walk Again.
Available online: https://www.umms.org/rehab/health-services/therapeutic-technology/lokomat-therapy (accessed on 4
January 2021).
Technologies 2021, 9, 8 26 of 26
137. Leading German Private Health Insurer Enters Contract to Provide ReWalk Exoskeletons to Individuals with Spinal Cord In-
juries. Available online: https://www.globenewswire.com/news-release/2020/12/10/2142922/0/en/Leading-German-Private-
Health-Insurer-Enters-Contract-to-Provide-ReWalk-Exoskeletons-to-Individuals-with-Spinal-Cord-Injuries.html (accessed on 4
January 2021).
138. Korus, S. Industrial Robot Cost Declines Should Trigger Tipping Points in Demand. Available online: https://ark-invest.com/
articles/analyst-research/industrial-robot-cost-declines/ (accessed on 4 January 2021).
139. Wan, S.; Gu, Z.; Ni, Q. Cognitive computing and wireless communications on the edge for healthcare service robots. Comput.
Commun. 2020, 149, 99–106. [CrossRef]
140. ISO 13482:2014 Robots and robotic devices—Safety requirements for personal care robots. Available online: https://www.iso.
org/standard/53820.html (accessed on 6 January 2021).
141. Villaronga, E.F. ISO 13482: 2014 and its confusing categories. Building a bridge between law and robotics. In New Trends in
Medical and Service Robots; Springer: Berlin/Heidelberg, Germany, 2016; pp. 31–44.
142. Fosch-Villaronga, E. Robots, Healthcare, and the Law: Regulating Automation in Personal Care; Routledge: Abingdon-on-Thames, UK,
2019.
143. Grasso, C. Challenges and advantages of robotic nursing care: A social and ethical analysis. Available online: https://
corporatesocialresponsibilityblog.com/2018/06/26/robotic-nursing-care/ (accessed on 5 January 2021).
144. Wachsmuth, I. Robots like me: Challenges and ethical issues in aged care. Front. Psychol. 2018, 9, 432. [CrossRef]
145. Taylor, A.; Lee, H.R.; Kubota, A.; Riek, L.D. Coordinating clinical teams: Using robots to empower nurses to stop the line. Proc.
ACM Hum.-Comput. Interact. 2019, 3, 1–30. [CrossRef]
146. Irfan, B.; Ramachandran, A.; Spaulding, S.; Glas, D.F.; Leite, I.; Koay, K.L. Personalization in long-term human-robot interaction.
In Proceedings of the 2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14
March 2019; pp. 685–686.
147. Rajavenkatanarayanan, A.; Nambiappan, H.R.; Kyrarini, M.; Makedon, F. Towards a Real-Time Cognitive Load Assessment
System for Industrial Human-Robot Cooperation. In Proceedings of the 2020 29th IEEE International Conference on Robot and
Human Interactive Communication (RO-MAN), Naples, Italy, 31 August–4 September 2020; pp. 698–705.
148. Evans, W. Ruthless Quotas at Amazon Are Maiming Employees. Atlantic 2019. Available online: https://www.theatlantic.com/
technology/archive/2019/11/amazon-warehouse-reports-show-worker-injuries/602530/ (accessed on 6 January 2021).
149. Del Rey, J. How robots are transforming Amazon warehouse jobs—For better and worse. Available online: https://www.vox.
com/recode/2019/12/11/20982652/robots-amazon-warehouse-jobs-automation (accessed on 6 January 2021).