Sensors: A Human Support Robot For The Cleaning and Maintenance of Door Handles Using A Deep-Learning Framework
Sensors: A Human Support Robot For The Cleaning and Maintenance of Door Handles Using A Deep-Learning Framework
Article
A Human Support Robot for the Cleaning and
Maintenance of Door Handles Using a Deep-Learning
Framework
Balakrishnan Ramalingam 1,* , Jia Yin 1,*, Mohan Rajesh Elara 1, Yokhesh Krishnasamy Tamilselvam 2,
Madan Mohan Rayguru 1 , M. A. Viraj J. Muthugala 1 and Braulio Félix Gómez 1
1 Engineering Product Development Pillar, Singapore University of Technology and Design (SUTD),
Singapore 487372, Singapore; [email protected] (M.R.E.); [email protected] (M.M.R.);
[email protected] (M.A.V.J.M.); [email protected] (B.F.G.)
2 Department of Electrical Engineering, University of Western Ontario, London, ON N6A 3K7, Canada;
[email protected]
* Correspondence: [email protected] (B.R.); [email protected] (J.Y.)
Received: 17 May 2020; Accepted:14 June 2020; Published: 23 June 2020
Abstract: The role of mobile robots for cleaning and sanitation purposes is increasing worldwide.
Disinfection and hygiene are two integral parts of any safe indoor environment, and these factors
become more critical in COVID-19-like pandemic situations. Door handles are highly sensitive
contact points that are prone to be contamination. Automation of the door-handle cleaning task
is not only important for ensuring safety, but also to improve efficiency. This work proposes an
AI-enabled framework for automating cleaning tasks through a Human Support Robot (HSR).
The overall cleaning process involves mobile base motion, door-handle detection, and control of the
HSR manipulator for the completion of the cleaning tasks. The detection part exploits a deep-learning
technique to classify the image space, and provides a set of coordinates for the robot. The cooperative
control between the spraying and wiping is developed in the Robotic Operating System. The control
module uses the information obtained from the detection module to generate a task/operational space
for the robot, along with evaluating the desired position to actuate the manipulators. The complete
strategy is validated through numerical simulations, and experiments on a Toyota HSR platform.
Keywords: HSR; door handle cleaning; human service robot; cleaning; object detection; deep learning
1. Introduction
Cleaning and disinfecting are very important steps in preventing the acquisition and spread of
infectious diseases inside closed environments such as apartments, community centers, shopping
malls, hospitals, etc. In particular, high-contact-point areas such as doors, lifts, handrails, etc. are
major sources of contamination. In [1], the authors reveal that the door handle and its connected
areas are highly sensitive contact points that are prone to be contaminated, and a key medium for
spreading the germs. Therefore, the frequent cleaning or disinfection of doors handles is essential
for preventing the acquisition and spread of infection. However, due to a shortage of manpower,
the frequent cleaning of indoor areas has become a key challenge [2]. In addition to that, involving
manpower carries a high risk of infection while working for long time in such areas. Mobile robots can
be used as a viable solution to the problems associated with conventional cleaning and disinfecting
methods due to their proven ability to assist humans in diverse application areas such as hospitals,
elderly homes, and industries.
Many robotic solutions have been designed and developed, targeting routine tasks in healthcare
facilitie. The works [3,4] demonstrate that service robots are used to deliver food and medicine.
UV-light-installed mobile robot was introduced by Danish-based company UVD Robot [5] for efficiently
disinfecting hospital rooms, which could slow the disease spread through viruses. These robots
can disinfect anything much better than other techniques, using a mobile array of powerful
short-wavelength ultraviolet-C (UVC) lights that emit enough energy to shred the DNA or RNA of any
microorganisms exposed to them. In [6], author proposed an autonomous robot to performs bed baths
in the pursuit of patient hygiene. In addition to this core functionality, the robot is equipped with a
fall-detection system consisting of a video camera and a 3D LiDAR to identify patients who have fallen
to the ground. The robot can notify the medical staff for the assistance in an emergency condition of a
person detected by the fall detection module and remote condition monitoring module. GeckoH13 [7]
has been developed as an automated solution for the cleaning of walls in health sectors.This robot can
climb on walls with the aid of vacuum adhesion. The task is carried out by spraying different liquid
mixers apart from steam. GeckoH13 is effective in sterilizing uniform wall surfaces to a great extent.
In [8–14] mobile robot platform had demonstrated for indoor and outdoor cleaning application such
as floor cleaning [8,9], lawn mowing [10], wall cleaning [15], window cleaning [16], pool cleaning [12],
pavement [11] and external facade cleaning [14]. Similarly, the service robots are utilized for cleaning
the table in food courts [13] and collecting the used plates from the customer.
However, the cleaning of sensitive contact points such as door handles is highly challenging for
service robots. To carry out this task, the service robot needs an optimal vision system to identify
the target object and the surrounding environment, and generate a trajectory planning scheme
corresponding to arm-manipulation function to safely perform the proper cleaning tasks. In the
literature, various techniques have been developed for mobile robots to recognize a door and door
handle, such as laser scanner technique, ultrasonic sensors, and image-processing schemes. Here,
laser scanning and ultrasonic sensors schemes have lot of limitations. They work mostly on a known
model of a door and accuracy is not good [17]. Image-processing-based schemes are widely studied
for mobile robot object-recognition tasks. These are also widely studied for door and door-handle
detection tasks. A door-frame detection model was proposed using Hough Transform to detect edges,
and a fuzzy logic-based algorithm was used to find the relationship between them [18]. However,
the model was unable to differentiate doors from large objects typically indoors. In another work,
a computer vision-based door-detection algorithm was reported for blind people conducting indoor
activity [19]. Here, the author used the generic geometric of a door model and the edges and corner
information for detecting doors. In another study [20], Huy-Hieu Pham used Kinect sensor-generated
point cloud data and a 3D-image-processing scheme to detect a door and other indoor objects. In this
work, the author uses a Voxel Grid Filter to downsample the point cloud data, and Euclidean Cluster
Extraction is used to determine the type of the object. The author reported that the model detected the
door with 90% precision, and took 0.41 seconds. Yimin et al. [21] developed a door-detection algorithm
for an indoor mobile robot pioneer using Kinect sensor data. Here, the author used histogram-based
edge detection to identify the door and used bilateral filtering to remove the depth image noise.
Furthermore, the author used waypoints to indicate the route of the robot. Fernández-Caramés
developed a real-time door-detection system for domestic robot navigation where the author uses
haar-like features and integral images to detect the door. The experiment was tested with MoRLaCo
mobile robot and detected a door with 4% false positives. Furthermore, the author reports that the
false detection ration varies according to distance between the door and the robot [22].
Deep Learning is an emerging technique. It has been widely used for image classification and
object detection. Generally, DL has different definition. In zhang et.al [23] clearly described definition
of DL. The authors stated that DL is a class of machine learning algorithms that learns the structure
between inputs and outputs, besides learning the relationship between two or multiple variables
but also the knowledge that manages the relationship as well as the insights that makes sense of
the relation. Recently, DL based object detection algorithms are widely used for various automation
applications. To pick and place the objects [24], monitoring construction sites [25], recognized trash for
cleaning robots [9,26], etc. These were also used for door and door-handle detection.
Sensors 2020, 20, 3543 3 of 18
In [27], Benjamin and Zhang proposed the door-handle detection method for the Stanford AI
Robot (STAIR), where the AdaBoost technique and direct feature selection-based cascaded approach
was used for door detection. Alejandra et al. used a machine-learning algorithm and contour detection
algorithm with spatial filtering technique to detect the door. The experiment was developed for
the TurtleBot indoor robot and the detection algorithm obtained 78.0% detection accuracy with a
miss-classification rate of 21.93%. Wei et al. [28] proposed a door and door poses detection method for
the SIAT robot using a deep-learning framework. The CNN model was trained with 20,500 test images
and its detection error was 2.82%. Elen et al. used computer vision and machine-learning technique to
identify the door handle and its types where the author used a 2D sliding-window object-detection
algorithm and k-means clustering technique for detecting and localizing a door. The experiment
was tested with the Star robot, and the model obtained 94.5% recognition and 93.2% localization
accuracy, respectively. In another case study, a door and cabinet recognition system was proposed by
Maurin et al. for a mobile robot. The author used a 7 ×7 darknet deep-learning model to recognize
the door and k-means color clustering for segmentation of the handle point [29].
Another important aspect of the indoor cleaning environment is motion-planning and arm
manipulation. Once the vision module can detect and localize the cleaning subject (doors, walls etc.),
the information should be exploited by the robot platform such that it will move to the desired location
and perform the cleaning task. The authors of [30] proposed a motion-planning strategy for the
opening of doors, using the concept of action primitives. In a similar line of work, a motion-planning
and arm-manipulation technique was proposed for grasping books from any table, and returning them
to a predetermined position [31]. The motion-planning and manipulation for an HSR is discussed
in [32,33].
However, the above-mentioned ML and DL studies were focused on door and door-handle
detection for a mobile service robot in different applications that were not related to cleaning services.
Moreover, very few studies have discussed integrating the vision module data for motion-planning
and manipulation. The objective of this case study is to develop a fully autonomous robotic cleaner to
conduct cleaning and wiping of door handles to avoid direct contact with germs or viruses, especially
in case of virus outbreak. To implement this case study, a lightweight deep-learning framework is
developed for a human support robot to detect a door handle. The detection module’s output is
exploited by the motion-planning and arm-manipulation algorithms to complete the cleaning task.
The experiment was tested in Toyota HSR and its arm part is customized to accomplish the cleaning
task. Using a cleaning tool, the robot can spray the disinfecting liquids then clean the sprayed region
with a cleaning brush to prevent infections. Furthermore, the indoor path-planning scheme was
adopted for a robot to navigate indoors and clean multiple doors. The arm-manipulation function was
added to control both the arm and cleaning tool to successfully accomplish the cleaning task.
The paper is organized as follows. Section 2 reviews the proposed system and explains the
deep-learning methods used to recognize door handles. This section also describes the cleaning
strategy and the arm manipulation. Section 3 presents the validated results in a real scenario. Section 4
presents conclusions and describes future work for this project.
2. Proposed System
Figure 1 shows the Human Support Robot (HSR) with the customized cleaning module.
The system architecture for implementing the door-handle cleaning function in the HSR platform is
shown in Figure 2. The details of the HSR platform, functionality of the system architecture and the
integration of the functional module in the HSR platform are described as follows.
The camera has 1280 × 1024 resolution and run on 30 frames per second. In addition, the wide-angle
camera installed on the manipulator is used to help locate the position of the door handle. IMU and
Hokuyo URG-04LX 2D LiDAR are provided in the HSR base for localization and mapping. The HSR is
also provided with pneumatic bumpers to prevent any possible collision.
images. After detecting the images, the corresponding bounding boxes are used to collect the matching
depth images and to project the pixels in 3D space. A point cloud is formed using the projected depth
pixels. A ROS object-recognition package is used to remove the pixels belonging to the ground surface
or other objects. After removing the points in the detection plane, the point closest to the robot is
further moved. By the end, the center point of the current point cloud is assigned as the position
of object.
Deep-Learning Framework
A deep-learning-based object-detection framework is used in the HSR platform to recognize the
door handle. The framework proposed in this paper is built to essentially identify and classify types of
door handles. The base framework used in this paper is YOLO V3 built using Dark Flow in Python.
Sensors 2020, 20, 3543 6 of 18
The key advantage is that unlike other object-detection algorithms such as R-CNN, Faster R-CNN,
or SSD mobile net, YOLO splits the images into grids. These grids are then visited to spot any objects
that resemble the object of interest and a probability value is issued for each grid cell. If the probability
value exceeds a specified threshold, then the algorithm detects that as a region containing the desired
object. This grid cell is further separated to localize the desired object within that specific grid cell,
draw a bounding box, and classify the object. In many other alternative algorithms, the entire image is
searched to localize and spot the desired object. This makes the performance of the YOLO much better
than its counterparts regarding swiftness and accuracy.
The object-detection network is trained on the training data and its performance is checked using
the validation data. In this paper, we have used the loss function as a performance check for the model.
If the performance is good, i.e., if the stopping criteria is attained, the trained model would be saved.
If not, the training is resumed until the stopping criteria is met. The final step represents implementing
the trained model for the real-time scenario in the robot. The trained model is used to process real-time
video from the robot’s camera to detect and classify door handles.
Figure 3 shows the architecture for the proposed network. It consists of 10 convolutional layers
and 6 pooling layers. It is a general rule of thumb to have a pooling layer after a convolutional layer and
therefore the convolutional and pooling layers are placed alternately. The reason for this is to enhance
the feature extraction process and to downsample to the spatial dimensions of the image matrix.
where f and g are two variables that are involved in convolution. f is the input and g is
the filter function. Equation (2) explains the convolution operation for discrete time signals [37].
∞
x (n) ∗ h(n) = ∑ x (k)h(n − k) (2)
k =−∞
Sensors 2020, 20, 3543 7 of 18
where x and h are two discrete time signals involved in convolution. In discrete-time signals,
the convolutions are explained as the sum of element-wise multiplication of two signals when one is
reversed and shifted by a factor of −k. The factor of k can be customized in the convolution, and this
factor is called as stride length. The filter is basically to detect edges in the images which can be
extracted as features while updating the weights. The trained model containing the filter with updated
weights is used in the testing process, which would detect similar features in the testing images and
assist with the classification.
P
Xmin
Y
min
y = Xmax (3)
Ymax
C1
C2
where P is a binary value which determines if there is an object of interest in the image, Xmin is the
upper-left x bounding box coordinate, Ymin is the upper-left Y bounding box coordinate, Xmax is the
lower-right x bounding box coordinate, Ymax is the lower-right y bounding box coordinate, C1 is 1 if
the object belongs to class 1, else zero, C2 is if the object belongs to class 2, else zero.
The detected bounding box coordinate of the door handle is converted into the 3D word
coordinate, which is essential for the robot to manipulate the arm. The bounding box coordinate,
i.e., top-right, top-left, bottom-left, and center are the 3D coordinates in the word frame in meters from
the base frame of the robot. Here, the projective geometry function is adopted to carry out the image
plane to 3D point in the real world coordinate [38]. The translation matrix, such as T = [t1 t2 t3 ] is
presumed from the vision module after performing the calibration from the source of world coordinates
and the orientation matrix R = [roll, pitch, yawn]. The intrinsic parameter of the camera lies in the
focal length f x , f y , distortion coefficient parameter σ, principle point ( xc , yc ), and size of pixel (sx, sy);
any pixel on the image plane p = [ p x py ] with 3D coordinate W = [ X Y d] world plane and W can
be calculated by p = H ∗ W, where H = K ∗ [ R T ] is the homogeneous matrix. After predicting the
point in the 3D world coordinates frame and setting it aside to the distinct object frame, the ROS
transformation frame service package will manage the respective location between this point to the
HSR coordinate frame.
manipulator is kept constant while the base platform motion and base platform position is kept constant
while performing the cleaning task. Numerous path-planning and arm-manipulation approaches
are available in the literature. A technique similar to [30] is proposed for motion-planning and arm
manipulation. The notations and symbols are similar to [33]. For that purpose, we represent the
base platform pose as (xb , yb , ψb ). The arm height with respect to the base is represented by z arm .
The roll, pitch, and yaw of the end-effector/gripper is defined as (ψg , θ g , φg ). The orientation of the
link connecting the base and gripper is denoted as (θ1 , θ2 ). It should be noted that the arm cannot roll
independently along the horizontal direction, so ψg = ψb . The cleaning task is done by manipulating
the gripper pitch θ g and gripper yaw φg .
Xmax + Xmin
xmid = ,
2
Ymax + Ymin
Ymid = .
2
The goal of the robot platform is to reach a small neighborhood around the detected door handle
mid-position (Xmid , Ymid ). The control objective is to decide the desired velocities for the robot platform.
By keeping the arms at rest, the platform motion can be moved by a differential drive methodology, i.e.,
ẋb cos(ψb ) 0 " #
v
ṗ = ẏb = sin(ψb ) 0 b (4)
ωb
ψ̇b 0 1
The robot can reach near (Xmid , Ymid ) by setting the desired positions as (Xmid ± e, Ymid ± e),
where e is a positive number to be tuned according to arm length and interior geometry. The desired
velocities (vb , ωb ) are the output of a dynamic feedback linearizing controller. The input velocity is
decided through the integrator defined as:
ζ̇ = f , vb = ζ. (5)
ζ̇ = f , vb = ζ
d( xb − Xmid )
f = Ẍmid − k1 − k2 ( xb − Xmid ± e) (7)
dt
d(yb − Ymid )
ω = Ÿmid − k1 − k2 (yb − Ymid ± e)
dt
where k1 , k2 ∈ R+ . With the control law (7), the closed-loop system can be described as:
d( xb − Xmid )
ẍb − Ẍmid = −k1 − k2 ( xb − Xmid ± e)
dt (8)
d(yb − Ymid )
ÿb − Ÿmid = −k1 − k2 (yb − Ymid ± e)
dt
Sensors 2020, 20, 3543 9 of 18
The differential Equation (8) describes a stable error dynamic, and hence the robot platform
reaches an e neighborhood of the door handle, i.e.,:
xb → Xmid ± e, yb → Ymid ± e.
dψb 0
where J is the Jacobean of the robot arm. The error in joint space is fed back through a PID to actuate
the robot arm. After this stage, the arm camera should be able to detect the door handle, and the image
of the door handle should be on the center pixels of the image matrix. This step should be repeated
until a desirable distance between the door handle and arm gripper is reached.
(2) After Step 1, one should get
xb ≈ xd , yb ≈ yd , z arm ≈ zd
where xd , yd , zd are desired distances from the door. The cleaning should be performed only by
actuating the gripper, without moving the base. The following actions should be repeated a fixed
number of times depending on cleaning requirements.
(I) At this position, give an error signal (0, 0, zmid + lh /2) (lh is the approximate length of the door
handle), and solve the inverse kinematics (9) to transform this error into joint space. Feed the error
into a PID controller, and exploit the output to actuate the gripper.
(II) Give an error signal (0, 0, zmid − lh /2), and follow Step 2 (I).
2.10. Navigation
A two-dimensional map of the environment is generated by ROS hector-slam using the data
from 2D LiDAR and IMU sensors from HSR. To localize itself at every time-step in the environment,
an Adaptive Monte Carlo Localization (AMCL) algorithm is used. Both a static and dynamic obstacle
map is needed to generate a combined obstacle map (costmap) for the proper navigation of HSR
in the environment. A static obstacle map is the 2D map already generated using hector-slam.
LIDAR and RGBD sensors update the dynamic obstacle map. Once the target is fixed, the path
is generated using the move_base algorithm on the combined obstacle map. As the robot has a
global path to follow and a costmap, the local planar generates the command velocities, (vx,vy
Sensors 2020, 20, 3543 10 of 18
and vθ) based on the global plan and costmaps, to navigate the HSR to the target position along
the planned path.The following ROS packages are used to accomplish the navigation tasks, amcl,
base_local_planner, base_local_planner, clear_costmap_recovery, costmap_2d, dwa_local_planner,
fake_localization, global_planner, map_server, move_base, move_base_msgs, move_slow_and_clear,
navfn, nav_core, robot_pose_ekf, rotate_recovery, voxel_grid.
3. Experimental Results
This section describes the experimental results of door detection and the cleaning mechanism.
Figure 5 shows the block diagram of of experimental procedure. The experiment has three phases,
which are training the deep-learning framework, mapping the indoor environment, and configuring
and configuring and testing with HSR robot platform.
Sensors 2020, 20, 3543 11 of 18
Here, the dataset preparation process involves collecting the different door images at high
resolution. The dataset contains about 4500 images collected online through Bing image search,
the Miguel et al. [39] door dataset and also our institution’s different door handles. The data set
images are collected in the robot perspective with Intel real-sense depth camera at different angles and
different lighting conditions. Furthermore, the collected door datasets are sorted into three classes,
such as circle type, lever type, and bar type handles. To balance the images, 1200 images are used for
each class in the training, and 300 images are used for validation. Image resolution of 640 × 480 was
used in the experiments for both training and testing the CNN model. Furthermore, data expansion
techniques are used to enhance the CNN learning rate and controlling overfitting on training phase.
In the data enhancement process, rotation, scaling, and flipping are applied on the collected image.
K-fold (here K = 10) cross-validation process is involved for assessing the model. In this assessment
process, the dataset is divided into 10 sections; 9 of these sets are used for training the model, and
the remaining one is used for testing the model. This process is repeated 10 times to eliminate any
biasing conditions due to particular training or testing datasets. The results from the performance
metrics are repeated 10 times and the means of the results are provided. The resulting images from the
highest-accuracy models are provided here. The deep-learning models were developed in Tensor-flow
1.9 Ubuntu 18.04 version and trained using the following hardware configuration: Intel core i7-8700k,
64 GB RAM, and Nvidia GeForce GTX 1080 Ti Graphics Cards.
Standard statistical measures are used to assess the performance of the proposed scheme.
These include accuracy, precision, recall, and Fmeasure , which are computed using the equations below:
tp + tn
Accuracy( Acc) = (10)
tp + f p + tn + f n
Sensors 2020, 20, 3543 12 of 18
tp
Precision( Prec) = (11)
tp + f p
tp
Recall ( Rec) = (12)
tp + f n
2 × precision × recall
Fmeasure ( F1 ) = (13)
precision + recall
Here, tp, f p, tn, f n represent the true positives, false positives, true negatives, and false negatives,
respectively, as per the standard confusion matrix.
For real-time door-handle disinfection trail, two indoor environments were chosen as test bed;
those environments were first mapped manually by mapping algorithm, with HSR and door location
information stored as wave-point data. Our first test bed has four doors fixed with a lever-type door
handle and another test bed has three doors, all fixed with a circle-type door handle, the localization
result was shown in Figure 7. Figure 8 shows the real-time door-handle cleaning task (test bed1)
accomplished by HSR, where Figure 8a,b shows the HSR navigation and door-handle detection and
Figure 8c,d demonstrates the spraying of disinfection liquid and cleaning of the door handle.
Figure 7. Localization result for test bed1 and test bed2
365 ensures that the proposed system was not profoundly affected by environmental factors like varying
Sensors 2020, 20, 3543 13 of 18
365 ensures that the proposed system was not profoundly affected by environmental factors like varying
In this trial, the HSR RGBD vision system was used to capture the door runs at 10 fps, and image
resolution was set to 640 × 480. The robot was operated in autonomous mode and door-handle
detection and navigation was recorded from remote console. The experiment was tested with different
lighting conditions and its detection results are shown in Figure 9. The experiment was trailed with
different lighting conditions and it was observed that the detection model detects door handles with a
confidence level of 88 to 92%.
The statistical measures indicate that the trained CNN model has detected the door handle with
an average of 94.56% accuracy for the offline test and 91.2% average accuracy for the online test.
The study ensures that the proposed system was not heavily affected by environmental factors such as
varying lighting conditions, but slightly affected by the mirror effect. However, the mirror issue has
been resolved by fusing depth information with the detection results.
Sensors 2020, 20, 3543 14 of 18
Furthermore, statistical measures have been performed for estimating the robustness of the
detection model for both the online and offline experiments. Table 1 shows the statistical measure
results for online and offline experiments.
Table 1. Statistical measures for door-handle detection.
The table results indicate that the proposed CNN framework took an average of 11.07 s to detect
the door handle in 180 test images, and took 18 and 10 min to clean the test bed 1 and test bed 2
door handles, respectively.
3.3. Comparison with Other Object Detection Framework and Existing Schemes
Table 3 shows the comparative analysis of the proposed object detection framework with other
object detection frameworks includes SSD Mobile Net, SSD Inception. The detection and classification
accuracy of two networks has been trained using the same data-set for a constant amount of time.
Standard metrics were used for the performance estimation of these methods, as shown in Table 3.
SSD-MobilNet and proposed object detection framework had almost the same accuracy. However,
the execution time and detection accuracy of SSD-Inception is higher than the proposed framework.
Our proposed system has achieved better trade-off between accuracy and computational cost with
reduced execution time.
To the best of our knowledge, there is no similar work for comparing our proposed system
performance metrics. Hence we consider the similar door handle detection schemes for comparative
analysis. In this contest, we took a learning-based door handle scheme for comparison analysis,
and models’ detection accuracy was considered for evaluation metrics. The detection result of various
door handle detection schemes is shown in Table 4.
Sensors 2020, 20, 3543 15 of 18
From the above table it is evident that our model has shown better performance metrics compared
to other models. In our work, the darkflow framework was used to build the CNN model. To improve
the detection accuracy, we fixed the low learning rate of existing works and also used a reduced
number of hidden layers after the convolution and pooling layer.
4. Conclusions
This work proposed a novel technique for door-handle cleaning automation using the Toyota HSR
platform. The detection and classification algorithm used a CNN based deep learning methodology,
which generates a set of coordinates surrounding the door handles. These coordinates are exploited,
to generate the operational space of the robot using ROS, and to develop manipulator motion
planning. The veracity of the technique is tested offline and online through a series of simulations
and experiments. The accuracy of the detection is calculated to be more than 90% in both cases.
The real-time experiments are performed on two different testbeds successfully, which confirmed the
efficiency of the proposed framework. It can be concluded that the proposed system has the potential
for cleaning door handles in the indoor and public places, which in turn helps to overcome the shortage
of manpower in the cleaning industry. Future aspects of this work involve obstacle avoidance, human
interaction and safe navigation.
Author Contributions: Conceptualization, B.R. and M.R.E.; methodology, M.R.E., B.R.; software, B.F.G., M.M.R.,
J.Y., Y.K.T.; validation, Y.K.T., M.A.V.J.M., M.M.R., B.R.; formal analysis, M.R.E., M.M.R., M.A.V.J.M.; investigation,
M.R.E.; resources, M.R.E.; data, B.R., B.F., Y.K.T., J.Y.; writing—original draft preparation, B.R., M.M.R., M.A.V.J.M.,
Y.K.T.; writing—review and editing, M.R.E., J.Y., M.M.R., M.A.V.J.M.; visualization, B.F.G., M.R.E.; supervision,
M.R.E.; project administration, M.R.E.; funding acquisition, M.R.E. All authors have read and agreed to the
published version of the manuscript.
Sensors 2020, 20, 3543 16 of 18
Funding: This research is supported by the National Robotics Programme under its Robotics Enabling Capabilities
and Technologies (Funding Agency Project No. 192 25 00051), National Robotics Programme under its Robot
Domain Specific (Funding Agency Project No. 192 22 00058) and administered by the Agency for Science,
Technology and Research.
Acknowledgments: The authors would like to thank, the National Robotics Programme, the Agency for Science,
Technology and Research, Rithu Varman internship student from Ngee Ann Polytechnic for spray module design
assistance and SUTD for their support.
Conflicts of Interest: The authors declar no conflict of interest.
References
1. Stop Touching Your Face! Available online: https://www.nytimes.com/2020/03/02/well/live/coronavirus-
spread-transmission-face-touching-hands.html (accessed on 7 May 2020).
2. Cross, S.; Gon, G.; Morrison, E.; Afsana, K.; Ali, S.; Manjang, T.; Manneh, L.; Rahman, A.; Saxena, D.;
Kranti, V.; et al. An invisible workforce: The neglected role of cleaners in patient safety on maternity units.
Glob. Health Action 2019, 12, 1480085. [CrossRef] [PubMed]
3. Medical Supplies Delivery Robots. Available online: https://arstechnica.com/tech-policy/2020/04/
delivery-robots-move-medical-supplies-to-help-with-covid-19-response/ (accessed on 2 March 2020).
4. Food Delivery Robots. Available online: https://www.unido.org/stories/china-robot-delivery-vehicles-
deployed-help-covid-19-emergency (accessed on 2 March 2020).
5. UVD Robots. Available online: https://spectrum.ieee.org/automaton/robotics/medical-robots/autonomous-
robots-are-helping-kill-coronavirus-in-hospitals (accessed on 2 March 2020).
6. King, C.H.; Chen, T.L.; Jain, A.; Kemp, C.C. Towards an assistive robot that autonomously performs bed
baths for patient hygiene. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent
Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 319–324.
7. Cepolina, F.E.; Muscolo, G.G. Design of a robot for hygienization of walls in hospital environments.
In Proceedings of the ISR/Robotik 2014, 41st International Symposium on Robotics, Munich, Germany,
2–3 June 2014; pp. 1–7.
8. Prabakaran, V.; Elara, M.R.; Pathmakumar, T.; Nansai, S. Floor cleaning robot with reconfigurable mechanism.
Autom. Constr. 2018, 91, 155–165. [CrossRef]
9. Ramalingam, B.; Prabakaran, V.; Ilyas, M.; Mohan, R.E.; Arunmozhi, M. Vision-Based Dirt Detection and
Adaptive Tiling Scheme for Selective Area Coverage. J. Sens. 2018, 2018, 1–13. [CrossRef]
10. Wang, K.; Huang, C. Intelligent Robotic Lawn Mower Design. In Proceedings of the 2018 International
Conference on System Science and Engineering (ICSSE), New Taipei, Taiwan, 28–30 June 2018; pp. 1–5.
11. Hayat, A.A.; Parween, R.; Elara, M.R.; Parsuraman, K.; Kandasamy, P.S. Panthera: Design of a reconfigurable
pavement sweeping robot. In Proceedings of the 2019 International Conference on Robotics and Automation
(ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 7346–7352.
12. Fu-Cai, Y.; Shi-jian, H.; Hai-liang, S.; Li-Zhu, W. Design of cleaning robot for swimming pools. In Proceedings
of the MSIE 2011, Harbin, China, 8–11 January 2011; pp. 1175–1178.
13. Yin, J.; Apuroop, K.G.S.; Tamilselvam, Y.K.; Mohan, R.E.; Ramalingam, B.; Le, A.V. Table Cleaning Task by
Human Support Robot Using Deep Learning Technique. Sensors 2020, 20, 1698. [CrossRef] [PubMed]
14. Kouzehgar, M.; Tamilselvam, Y.K.; Heredia, M.V.; Elara, M.R. Self-reconfigurable façade-cleaning robot
equipped with deep-learning-based crack detection based on convolutional neural networks. Autom. Constr.
2019, 108, 102959. [CrossRef]
15. Muthugala, M.A.V.J.; Vega-Heredia, M.; Mohan, R.E.; Vishaal, S.R. Design and Control of a Wall Cleaning
Robot with Adhesion-Awareness. Symmetry 2020, 12, 122. [CrossRef]
16. Vega-Heredia, M.; Mohan, R.E.; Wen, T.Y.; Aisyah, J.S.; Vengadesh, A.; Ghanta, S.; Vinu, S. Design and
modelling of a modular window cleaning robot. Autom. Constr. 2019, 103, 268–278. [CrossRef]
17. Chitta, S.; Cohen, B.; Likhachev, M. Planning for autonomous door opening with a mobile manipulator.
In Proceedings of the 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK,
USA, 3–7 May 2012; pp. 1799–1806.
18. Mu, R.; Miguel, A. Door-Detection Using Computer Vision and Fuzzy Logic. WSEAS Trans. Syst. 2004,
91, 3047–3052,
Sensors 2020, 20, 3543 17 of 18
19. Tian, Y.; Yang, X.; Arditi, A. Computer Vision-Based Door Detection for Accessibility of Unfamiliar
Environments to Blind Persons. In Computers Helping People with Special Needs; Miesenberger, K., Klaus, J.,
Zagler, W., Karshmer, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 263–270.
20. Pham, H.H.; Le, T.; Vuillerme, N. Real-Time Obstacle Detection System in Indoor Environment for the
Visually Impaired Using Microsoft Kinect Sensor. J. Sens. 2016, 2016, 1–13. [CrossRef]
21. Zhou, Y.; Jiang, G.; Xu, G.; Wu, X.; Krundel, L. Kinect depth image based door detection for autonomous
indoor navigation. In Proceedings of the 23rd IEEE International Symposium on Robot and Human
Interactive Communication, Edinburgh, UK, 25–29 August 2014; Volume 2014, pp. 147–152. [CrossRef]
22. Fernández-Caramés, C.; Moreno, V.; Curto, B.; Rodríguez-Aragón, J.; Rodríguez, F. A Real-time Door
Detection System for Domestic Robotic Navigation. J. Intell. Robot. Syst. 2014, 76, 119–136. [CrossRef]
23. Zhang, W.J.; Yang, G.; Lin, Y.; Ji, C.; Gupta, M.M. On Definition of Deep Learning. In Proceedings of the
2018 World Automation Congress (WAC), Stevenson, WA, USA, 3–6 June 2018; pp. 1–5.
24. Caldera, S.; Rassau, A.; Chai, D. Review of Deep Learning Methods in Robotic Grasp Detection.
Multimodal Technol. Interact. 2018, 2, 57. [CrossRef]
25. Spencer, B.F.; Hoskere, V.; Narazaki, Y. Advances in Computer Vision-Based Civil Infrastructure Inspection
and Monitoring. Engineering 2019, 5, 199–222. [CrossRef]
26. Ramalingam, B.; Lakshmanan, A.K.; Ilyas, M.; Le, A.V.; Elara, M.R. Cascaded Machine-Learning Technique
for Debris Classification in Floor-Cleaning Robot Application. Appl. Sci. 2018, 8, 2649. [CrossRef]
27. Sapp, B.; Zhang, J.J. Door Handle Detection for the Stanford AI Robot (STAIR). 2005. Available online: http:
//cs229.stanford.edu/proj2005/SappZhang-DoorHandleDetectionForSTAIR.pdf (accessed on 21 June 2020).
28. Chen, W.; Qu, T.; Zhou, Y.; Weng, K.; Wang, G.; Fu, G. Door recognition and deep learning algorithm for
visual based robot navigation. In Proceedings of the 2014 IEEE International Conference on Robotics and
Biomimetics (ROBIO 2014), Bali, Indonesia, 5–10 December 2014; pp. 1793–1798.
29. Llopart, A.; Ravn, O.; Andersen, N.A. Door and cabinet recognition using Convolutional Neural Nets and
real-time method fusion for handle detection and grasping. In Proceedings of the 2017 3rd International
Conference on Control, Automation and Robotics (ICCAR), Nagoya, Japan, 22–24 April 2017; pp. 144–149.
30. Nagatani, K.; Yuta, S. Designing strategy and implementation of mobile manipulator control system
for opening door. In Proceedings of the IEEE International Conference on Robotics and Automation,
Minneapolis, MN, USA, 22–28 April 1996; Volume 3, pp. 2828–2834.
31. Takahama, T.; Nagatani, K.; Tanaka, Y. Motion planning for dual-arm mobile manipulator-realization of
“tidying a room motion”. In Proceedings of the IEEE International Conference on Robotics and Automation,
ICRA’04, New Orleans, LA, USA, 26 April–1 May 2004; Volume 5, pp. 4338–4343.
32. Yi, J.B.; Yi, S.J. Mobile Manipulation for the HSR Intelligent Home Service Robot. In Proceedings of the 2019
16th International Conference on Ubiquitous Robots (UR), Jeju, Korea, 24–27 June 2019; pp. 169–173.
33. Berenstein, R.; Fox, R.; McKinley, S.; Carpin, S.; Goldberg, K. Robustly adjusting indoor drip irrigation
emitters with the toyota hsr robot. In Proceedings of the 2018 IEEE International Conference on Robotics
and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 2236–2243.
34. Naranjo-Torres, J.; Mora, M.; Hernández-García, R.; Barrientos, R.J.; Fredes, C.; Valenzuela, A. A Review of
Convolutional Neural Network Applied to Fruit Image Processing. Appl. Sci. 2020, 10, 3443. [CrossRef]
35. Almabdy, S.; Elrefaei, L. Deep Convolutional Neural Network-Based Approaches for Face Recognition.
Appl. Sci. 2019, 9, 4397. [CrossRef]
36. Elhassouny, A.; Smarandache, F. Trends in deep convolutional neural Networks architectures: A review.
In Proceedings of the 2019 International Conference of Computer Science and Renewable Energies (ICCSRE),
Agadir, Morocco, 22–24 July 2019; pp. 1–8. [CrossRef]
37. Katkar, P.; Sridhar, T.N.; Sharath, G.M.; Sivanantham, S.; Sivasankaran, K. VLSI implementation of fast
convolution. In Proceedings of the 2015 Online International Conference on Green Engineering and
Technologies (IC-GET), Coimbatore, India, 27 November 2015; pp. 1–5. [CrossRef]
38. Le, A.V.; Hayat, A.A.; Elara, M.R.; Nhan, N.H.K.; Prathap, K. Reconfigurable Pavement Sweeping Robot and
Pedestrian Cohabitant Framework by Vision Techniques. IEEE Access 2019, 7, 159402–159414. [CrossRef]
39. Arduengo, M.; Torras, C.; Sentis, L. Robust and Adaptive Door Operation with a Mobile Manipulator Robot.
2019. Available online: https://arxiv.org/abs/1902.09051 (accessed on 21 June 2020).
40. Jauregi, E.; Martinez-Otzeta, J.M.; Sierra, B.; Lazkano, E. Door handle identification: A three-stage approach.
IFAC Proc. Vol. 2007, 40, 517–522. [CrossRef]
Sensors 2020, 20, 3543 18 of 18
41. Niu, L.; Qian, C.; Rizzo, J.; Hudson, T.; Li, Z.; Enright, S.; Sperling, E.; Conti, K.; Wong, E.; Fang, Y. A Wearable
Assistive Technology for the Visually Impaired with Door Knob Detection and Real-Time Feedback for
Hand-to-Handle Manipulation. In Proceedings of the 2017 IEEE International Conference on Computer
Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 1500–1508.
42. Klingbeil, E.; Saxena, A.; Ng, A.Y. Learning to open new doors. In Proceedings of the 2010 IEEE/RSJ
International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010;
pp. 2751–2757.
43. Zhang, W.; Van Luttervelt, C. Toward a resilient manufacturing system. CIRP Ann. 2011, 60, 469–472.
[CrossRef]
c 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).