Neural_network-based_algorithm
Neural_network-based_algorithm
com/scientificreports
The ability to recognize and interact with a variety of doorknob designs is an important component on
the path to true robot adaptability, allowing robotic systems to effectively interact with a variety of
environments and objects The problem addressed in this paper is to develop and implement a method
for recognizing the position of a door handle by a robot using data from an RGBD camera. To achieve
this goal, we propose a revolutionary approach designed for autonomous robots that allows them
to identify and manipulate door handles in different environments using data obtained from RGBD
cameras. This was achieved by creating and annotating a complete dataset consisting of 5000 images
of door handles from different angles, with the coordinates of the vertices of the bounding rectangles
labeled. The architectural basis of the proposed approach is based on MobileNetV2, combined with
a special decoder that optimally increases the resolution to 448 pixels. A new activation function
specially designed for this neural network is implemented to ensure increased accuracy and efficiency
of raw data processing. The most important achievement of this study is the model’s ability to work in
real-time, processing up to 16 images per second. This research paves the way for new advancements
in the fields of robotics and computer vision, making a substantial contribution to the practical
deployment of autonomous robots in a myriad of life’s spheres.
Keywords Robotics, RGBD camera, Neural network, Framework TensorFlow, The Mean Squared Error
Vol.:(0123456789)
www.nature.com/scientificreports/
Furthermore, the integration of depth information with RGB data, while promising, has not been fully
exploited in many existing approaches. Most methods predominantly focus on either RGB or depth data, failing
to harness the combined potential of RGBD cameras to enhance accuracy and robustness. This oversight results
in suboptimal performance, particularly in complex scenarios requiring precise spatial awareness.
This research aims to fill these gaps by developing a robust and efficient method for door handle recognition
using an RGBD camera, capable of real-time operation. By leveraging the full spectrum of data provided by
RGBD cameras, our approach enhances the robot’s ability to accurately detect and interact with door handles
under diverse lighting conditions and in various environments. Additionally, the proposed method integrates
advanced machine learning techniques to ensure rapid processing and decision-making, enabling real-time
application. Through comprehensive testing and validation, we demonstrate the method’s superiority in both
accuracy and operational efficiency, paving the way for more versatile and reliable autonomous robotic systems.
The primary objective of this work is the development and optimization of a method for the recognition of
door handle positions utilizing data from an RGBD camera. This encompasses the exploration and implementa-
tion of computer vision and machine learning methodologies, such as Convolutional Neural Network (CNN),
alongside the creation of specialized algorithms for the analysis of received data and the ascertainment of the
door handle’s isometry.
In the sphere of safety and assistance, robots endowed with the capability to autonomously operate doors
play a pivotal role, particularly in aiding individuals with disabilities or the elderly5. This function not only
enhances their mobility within residential settings but also promotes a greater degree of independence. In the
context of search and rescue operations6, the robotic ability to manipulate doors becomes critically valuable,
especially in scenarios of emergencies such as fires or earthquakes. Robots, capable of accessing perilous zones,
can potentially save lives by conducting search and rescue missions in situations where human intervention
carries substantial risk.
Within the ambit of intelligent home and office systems7, robotic interaction with door handles can be lever-
aged to manage access and bolster security, thereby augmenting the comfort and efficiency associated with the
use of such intelligent systems. In industrial environments, robots capable of transporting materials across various
sections, segregated by doors, contribute to a more flexible and efficient production p rocess8. The automation of
door operation paves the way for the development of more integrated production lines.
Moreover, in the field of research and exploration, the ability of robots to open doors allows for the investi-
gation of extreme environments, such as space missions or underwater e xplorations9–11, significantly reducing
risks to human life. This capability is instrumental in extending the frontiers of human knowledge, as it enables
robotic exploration in conditions that are otherwise inaccessible or hazardous for human beings.
The real-time operational capability of the model heralds significant prospects across various sectors where
the celerity of response is of paramount i mportance12,13. In the realms of robotics, autonomous vehicular systems,
and video surveillance frameworks, the immediacy in responding to incoming data can avert potential incidents
necessitating urgent intervention. This empowers autonomous systems to rapidly formulate decisions, adeptly
adapting to alterations in their ambient environment.
The motivation for this paper stems from the key role that autonomous mobile robots play in various fields,
including logistics, manufacturing, healthcare, and transportation, where their ability to perform tasks autono-
mously increases productivity and safety. In particular, recognizing and manipulating door handles is a critical
task in modern robotics that has significant implications for various sectors. For example, in the area of security
and assistance, autonomous robots capable of manipulating doors can significantly help people with disabilities
or the elderly by facilitating their mobility and independence. In addition, in industrial plants, the interaction of
robots with door handles contributes to more flexible and efficient production processes. In addition, the ability
of robots to open doors is crucial for research and exploration in extreme environments, such as space missions
or underwater exploration, where human access is limited.
The primary objective of this work is the development and optimization of a method for the recognition of
door handle positions utilizing data from an RGBD camera. This encompasses the exploration and implementa-
tion of computer vision and machine learning methodologies, such as Convolutional Neural Network (CNN),
alongside the creation of specialized algorithms for the analysis of received data and the ascertainment of the
door handle’s isometry.
The principal contributions of this article are summarized as follows:
1. An investigation and analysis of existing object recognition and image segmentation methodologies.
2. The proposition of a model architecture based on MobileNetV2, integrated with a bespoke decoder for
optimal resolution enhancement.
3. The development of an innovative approach enabling autonomous robots to identify and interact with door
handles in diverse environments, based on real-time RGBD data.
4. The execution of verification and validation procedures for the proposed algorithm on an array of door handles
under varying lighting conditions and backgrounds, coupled with an analysis of the accuracy and reliability
of the developed recognition system to ascertain its practical applicability in real-world robotic scenarios.
The structure of the remainder of the article is as follows: “State-of-the-arts” section provides an analysis of
different approaches to door opening by autonomous robots and an overview of existing analogaazs in control-
ling mobile robotic manipulators. In “Task” section articulates the research problem. The proposed algorithm
is detailed in “Proposed algorithm” section. In “Results of numerical experiments” section elaborates on the
test data and the results obtained. Finally, the conclusions and prospects for future research are presented in the
“Conclusions” section.
Vol:.(1234567890)
www.nature.com/scientificreports/
State‑of‑the‑arts
Existing research on door handle recognition and manipulation by autonomous robots can be broadly categorized
into three main approaches: methods utilizing RGB cameras for image recognition, techniques employing depth
sensors for spatial analysis, and hybrid approaches integrating both RGB and depth data. Within these categories,
various algorithms have been developed, including YOLOv5 for object detection, convolutional neural networks
for feature extraction, and reinforcement learning for control optimization.
In their scholarly article14, Wang et al. elucidate a methodology designed to enhance the reliability of control-
ling a mobile robotic manipulator in scenarios characterized by non-deterministic environmental behaviors.
The method incorporates an augmented Proximal Policy Optimization (PPO) algorithm. The focal point of this
research is the application of mobile robotic arms for door opening. The process commences with the identifica-
tion of the door handle’s position using an image recognition technique predicated on YOLOv515,16. Subsequently,
the interaction between the robotic arm and its environment is simulated utilizing the CoppeliaSim platform17.
Following this, a control strategy founded on a reward function is formulated to facilitate the training of a robotic
arm to open doors within actual environmental settings. Experimental outcomes demonstrate that this proposed
methodology expedites the convergence of the learning process and notably diminishes minor movements of
the robotic arm, thereby augmenting control stability.
In their research endeavor18, Arduengo et al. have developed a framework tailored for the robust and adaptive
operation of conventional doors by an autonomous mobile manipulator. This framework merges a convolutional
neural network with efficient point cloud analysis to enhance both robustness and speed performance. This inno-
vative approach allows for the real-time ascertainment of door handle grasping postures using RGBD imagery, a
crucial aspect of human-centered assistive applications. Additionally, a multifaceted Bayesian framework is pro-
posed, enabling the robot to analyze the door’s kinematic model based on motion observations and incorporate
prior experiences or learning derived from human demonstrations. The amalgamation of these algorithms with
the Task Space Region motion p lanner19 facilitates proficient interaction with doors, irrespective of their specific
kinematic models. The efficacy of this framework has been empirically validated through practical experiments
conducted with the Toyota Human Support R obot20.
In their scholarly work21, Stuede et al. introduce a strategy for door opening employing the KUICA KMR
iiwa mobile robot22. This robot demonstrates autonomous capabilities in recognizing doorknobs, opening doors,
and navigating through doorways without pre-existing knowledge of the door models or geometries. A CNN
architecture is utilized for doorknob detection, offering robustness across various doorknob designs and colors.
A 100% success rate in detection was achieved among a diverse set of 38 door handles, utilizing the highest
confidence level as the selection criterion. The door’s plane was identified through the segmentation of depth
data, facilitating the creation of bounding boxes for the handles. This methodology operates in real-time on an
external computer, achieving approximately 20 Hz, constrained by limited access to the KMR iiwa’s internal
control loops. Employing this approach, the robot successfully opened doors and navigated through them in 22
out of 25 trials across five distinct doors.
In their scholarly publication, Quintana et al.23 introduce a novel methodology for the detection of open
and closed doors within 3D laser scanning data, a pivotal task for the indoor navigation of autonomous mobile
robots. Their approach synthesizes both geometric (XYZ coordinates) and chromatic (RGB/HSV) data, derived
from a calibrated 3D laser scanner coupled with a color camera. The method is developed within the ambit of a
sophisticated 6D spatial structure, integrating geometric and color features. This integration, along with other
distinctive attributes of the methodology, imparts robustness against overlaps and the capacity to accommodate
minor door color variations, which may arise due to disparate lighting conditions encountered during scanning
at various locations. The efficacy of this method has been validated through assessments conducted on both
simulated and real-world scenarios, yielding promising outcomes.
Additionally, the approach proposed by Mochurad et al.24 for door handle detection warrants attention. Their
article delineates a novel algorithm that empowers autonomous mobile robots to operate a range of door types
independently, without human intervention. This algorithm employs machine learning techniques, notably the
YOLOv5 object detection model, the iterative RANSAC method for parameter estimation, and the DBSCAN clus-
tering algorithm25, with comparative analyses of alternative clustering techniques also undertaken. The algorithm’s
performance was evaluated in both simulated and real-world contexts, achieving a notable success rate of 95% in
100 door-opening attempts. The doorknob detection algorithm exhibited an average localization error of 1.98 mm
across 10,000 samples, underscoring its precision in pinpointing actual doorknobs. These results attest to the
algorithm’s accuracy and its efficacy for real-time, autonomous operation across diverse door types. Nonetheless,
the experiments identified certain limitations of this approach. The method relies on selecting the optimal cluster
and ascertaining its centroid as the door handle’s position, necessitating a specific angle between the camera and
the door handle. Deviations from this requisite angle may lead to centroid displacement, potentially culminating
in unsuccessful door operation attempts. Moreover, consistently achieving a favorable joint configuration of the
robot arm that ensures the desired angle between the camera and the door plane is not invariably feasible.
Unlike previous methods, the proposed MobileIsometryDet model mitigates the high processing time and
limited generalization issues by leveraging the efficient MobileNetV2 architecture combined with a bespoke
decoder. This innovative architecture enables the model to perform real-time operation and deliver robust
performance across diverse lighting conditions and door handle designs. By utilizing an optimized decoder,
the MobileIsometryDet model significantly enhances the resolution and accuracy of door handle detection,
ensuring that the system can function effectively in varied and dynamic environments. Additionally, the inte-
gration of a specialized activation function tailored for this architecture improves the overall efficiency of raw
data processing, further contributing to the model’s high accuracy and reliability. These advancements position
the MobileIsometryDet model as a superior solution for real-world applications where speed, accuracy, and
adaptability are critical.
Vol.:(0123456789)
www.nature.com/scientificreports/
Table 1 in the publication concisely enumerates the advantages and disadvantages of each of the existing
analogous methods under review, providing a comprehensive comparative analysis.
Therefore, various approaches have been analyzed above, including the use of convolutional neural networks,
optimization algorithms, image recognition techniques, interactive simulations, and the integration of these
technologies to perform specific tasks such as opening doors.
Task
Input data for model training
To facilitate the training of the model, a meticulously annotated dataset was curated, comprising 5000 images.
Each image is accompanied by the corresponding coordinates of the bounding box (BB) vertices of the door
handle, situated within the camera’s coordinate framework. The dataset’s labeling is visually represented in Fig. 1,
where two doors from the dataset are displayed, accompanied by their respective labels, indicated by a green
BB.
In a formal capacity, the dataset is represented as an ordered pair of tensors, denoted as Xdata , ydata where
Xdata ∈ Rn×h×w×3 corresponds to the tensor of images and ydata ∈ Rn×24 corresponds to the tensor of BB coor-
dinates. Here, n denotes the total number of data points within the dataset, amounting to 5000, h represents the
height of the images, and w denotes the width of the images. It is pertinent to note that the dataset utilized for
this research endeavor is not available for public access.
Figure 1. Visualization of two doors from the dataset and their labels in the form of a green BB.
Vol:.(1234567890)
www.nature.com/scientificreports/
These data inputs are critically analyzed and processed by the algorithm to ascertain with precision the spatial
position and orientation of the door handles. This process involves meticulous computations and the application
of advanced image-processing techniques to ensure accurate results. The integration of these components is piv-
otal in the algorithm’s ability to reliably determine the exact spatial parameters necessary for effective interaction
with door handles in varying environmental settings.
There are no stringent specifications regarding the orientation of the camera relative to the door handle in
the dataset. It is imperative that the data utilized during the model’s operation closely resemble the training
data in terms of visual characteristics and perspective. Optimally, the identical camera, equipped with the same
intrinsic parameters, should be employed throughout both the training and application phases. This ensures that
the camera’s viewing angles on the door handle during model deployment closely approximate those captured
in the dataset.
Proposed algorithm
MobileNetV226, a significant advancement in the sphere of deep learning and computer vision, was developed
by the Google AI team. This model, evolving from its predecessor MobileNet, is architecturally optimized for
heightened efficiency, a crucial attribute for deployment in mobile devices and systems with constrained com-
putational resources. Distinguished for its high accuracy, MobileNetV2 achieves this with considerably fewer
parameters in comparison to larger, more intricate models. A hallmark of MobileNetV2 is the incorporation
of inverted residuals and linear blocks, innovations that underpin the model’s efficiency. This structural design
enables the reduction of computational demand and the number of parameters, without compromising recogni-
tion quality. Such characteristics render MobileNetV2 particularly suitable for real-time applications where swift
data processing is imperative. Employed extensively in various computer vision tasks like image classification,
object detection, and segmentation, MobileNetV2’s versatility and efficiency have led to its widespread applica-
tion in mobile applications, robotics, automated video surveillance systems, and other areas requiring fast and
reliable image processing within the bounds of limited computational resources. Its adeptness in handling diverse
data types solidifies its role as a vital tool in the contemporary realm of artificial intelligence. The MobileNetV2
framework represents an evolutionary advancement of the prototypical MobileNet structure, crafted with an
emphasis on augmented operational efficiency for deployment across mobile apparatuses and devices character-
ized by computational resource constraints. Salient Features of MobileNetV2:
1. Inverted Residuals: This pivotal innovation underpins MobileNetV2’s architecture, facilitating the propaga-
tion of features via "inverted residuals" with minimal informational attrition.
2. 2. Linear Bottlenecks: Positioned at the terminal phase of each block, these bottlenecks are instrumental in
preserving the representational integrity of features.
The dimensionality of the MobileNetV2 model, or the weight file, is contingent upon the selected configura-
tion and the applied depth multiplier. MobileNetV2 proffers an array of models varying in “depth multipliers”
and input resolutions, which confer the flexibility to calibrate the equilibrium between precision and model
dimensions.
Figure 2. Visualization of the corresponding depth pictures to the data from the dataset, which are previously
presented in Fig. 1.
Vol.:(0123456789)
www.nature.com/scientificreports/
For instance, the archetypal iteration of MobileNetV2, utilizing a depth multiplier of 1.0 and an input dimen-
sion of 224 × 224 pixels, typically encompasses a model size approximating 14 megabytes (MB). Conversely,
opting for a reduced depth multiplier, such as 0.5, can result in a commensurately reduced model size.
Vol:.(1234567890)
www.nature.com/scientificreports/
the determination of this specific point in space relative to the camera. The remaining values are essential for
ascertaining the orientation of the door handle, accommodating the variability in angles from which the camera
may view the handle.
Thus, the decoder plays a crucial role in the architecture of the proposed neural network (see Fig. 3). Its main
function is to recover the high-resolution original data from the low-dimensional representation generated by
the encoder. In this architecture, the encoder compresses input images with a width and height of 448 pixels to a
smaller tensor with a width and height of 28 pixels. This compression is achieved through a series of convolution
layers that extract features from the input images.
After the encoder has created this compressed representation, the decoder samples this representation to the
original 448-pixel resolution. This process involves four layers of sampling, each of which increases the spatial
dimensions of the tensor, gradually expanding it to its original size. This reconstruction is necessary to preserve
the spatial information and details lost during encoding.
In summary, the encoder extracts relevant features from the input images, compressing them into a lower-
dimensional representation. The decoder then reconstructs this representation back to its original resolution,
while also assigning specific values to each pixel to facilitate door handle localization, spatial positioning, and
orientation determination. This interplay between the encoder and decoder enables the neural network to effec-
tively analyze and interpret the input images in the context of door handle detection and localization.
Postprocessing
The process of transmuting a depth image into a constellation of points necessitates the assignment of each pixel’s
depth value to a corresponding three-dimensional point within spatial coordinates. This intricate procedure
mandates the acquisition of specific intrinsic camera parameters alongside the requisite depth data. Intrinsic
parameters are typically constituted by the focal lengths ( fx and fy ), the coordinates of the principal point (cx
and cy ), and occasionally, parameters accounting for lens distortion.
A depth image is essentially an array of distance metrics, where each pixel details the measured distance
from the camera to the point in the scene it denotes. In the endeavor to metamorphose a depth image into a
point cloud, it is imperative to first procure the depth information. This involves the interpretation of each pixel
within the depth image as a quantitative indicator of the range between the camera’s lens and the scene’s physi-
cal elements.
∗ To∗ calculate the 3D coordinates for each depth value at the pixel x, y , you can calculate the 3D point
x , y , z using the Eq. (1).
∗
x∗ = ((x
�� − cx )�· d)/f
� x,
y ∗ = y − cy · d /fy , (1)
z ∗ = d.
Loss function
At the output of the neural network, we have the tensor ypred ∈ R448×448×5 . For each training set, we have
ytrue ∈ R448×448×5. Then the Mean Squared Error (MSE) function looks like (3).
448 448 5
1 2
MSE ytrue , ypred = ytrue,ijk − ypred,ijk . (3)
448 × 448 × 5
i=1 j=1 k=1
Training setup
The training regimen was executed utilizing the TensorFlow framework. Training of the network was conducted
on a singular GPU, utilizing the Adam optimization algorithm with a batch size of 32 instances. The learning rate
was initially established at 0.01 and was methodically reduced to 0.001 across a span of 30 epochs. To facilitate
the deployment of the trained models on mobile devices, a conversion process into the TensorFlow Lite (TFLite)
format was implemented. This conversion is essential for ensuring that the models are compatible with the
hardware and computational constraints inherent to mobile technology platforms.
Augmentation
During the training phase, it is imperative to employ augmentation techniques that do not induce alterations
to the intrinsic parameters of the camera. The intrinsic parameters encompass characteristics intrinsic to the
camera that influence the projection of three-dimensional scenes onto two-dimensional images. These param-
eters entail focal length, coordinates of the image center, and distortion coefficients, among others. Maintaining
the integrity of these parameters throughout the augmentation process signifies that any modifications must
Vol.:(0123456789)
www.nature.com/scientificreports/
not simulate variations in camera focal length, position, or orientation. Adherence to this guideline is vital to
ensure that the neural network is trained to recognize and accurately interpret the spatial positioning of objects,
utilizing images that accurately represent actual scenes, as opposed to those in which fundamental perspective
or geometric properties have been synthetically modified. Such a methodological approach is instrumental in
circumventing erroneous learning scenarios, wherein the network may become unduly sensitive to non-realistic
transformations in the visual appearance of objects or their spatial arrangement.
In this research, two specific types of augmentation were implemented that respect the camera’s internal
parameters: blur and horizontal flip, as depicted in Fig. 4. These augmentation techniques have been identified
as optimally congruent with the task at hand. The horizontal flip effectively mirrors the scenario of a door as
viewed from the opposite side, while the application of blur can mimic the effect of slight camera defocus, both
of which are practical phenomena that do not compromise the camera’s intrinsic attributes.
Algorithm 1. DetectHandleWithNN
Vol:.(1234567890)
www.nature.com/scientificreports/
Algorithm 1 delineates the comprehensive procedural pipeline which commences with the utilization of a
neural network and extends to the post-processing phase that employs the intrinsic parameters of the camera.
This systematic sequence culminates in the acquisition of the isometry of the doorknob, characterized by both
a position vector and its corresponding quaternion.
Figure 5. Visualization of the first channel, which is responsible for the probability distribution of the center of
the door handle.
Vol.:(0123456789)
www.nature.com/scientificreports/
Figure 6. The results of the algorithm on the door handle from several angles.
Table 2. Quantitative indicators for comparing the effectiveness of the proposed model with existing analogs:
processing speed and memory size.
Vol:.(1234567890)
www.nature.com/scientificreports/
YOLO-6D. Furthermore, the MobileIsometryDet model processes only six frames per second less than a model
designed for a more rudimentary task of classification MobileNet. The proposed MobileIsometryDet model has
a fairly high FPS and a small size. It is capable of processing 16 frames per second and takes up 16 megabytes.
Consequently, this efficiency positions the MobileIsometryDet model as a viable option for deployment on mobile
devices, an attribute that is particularly pertinent for the majority of autonomous mobile robots.
Conclusions
This study marks a noteworthy advancement in the realm of algorithm development for the identification of door
handle positions utilizing RGBD camera data. The research entailed a sequence of essential phases, commencing
with the compilation and annotation of a substantial dataset comprising 5000 images, each annotated with the
bounding box (BB) vertices’ coordinates of door handles. This preliminary stage was pivotal in guaranteeing the
effective training of the neural network.
The foundational element of this development was the MobileIsometryDet architecture, an innovative inte-
gration of an encoder based on MobileNetV2 and a decoder facilitating upsampling to a resolution of 448 pixels.
A uniquely devised activation function was employed within this architecture, optimizing the efficiency of
processing the network’s output. Throughout the model training phase, the MSE function was utilized, enabling
the precise quantification of discrepancies between predicted outcomes and actual results. This aspect is crucial
in enhancing the accuracy of the model.
The study placed significant emphasis on data augmentation, particularly methods that do not alter the
camera’s intrinsic parameters. This approach ensured a superior quality of network training, equipping the
model to adapt to diverse lighting scenarios and perspectives. The implementation of blur and horizontal flip
as primary augmentation techniques contributed to increased variability in the training dataset, maintaining
the data’s relevance.
One of the paramount achievements of this work is the model’s capability to operate in real-time, processing
up to 16 images per second, indicative of the model’s high efficiency and rapid processing ability. This feature is
critically important for the practical application of the model. The algorithm demonstrated exceptional adapt-
ability and precision in identifying isometric parameters of objects, even when tested on door handles not
included in the training dataset.
The ability to recognize and interact with doorknobs signifies a significant advancement toward integrat-
ing robots seamlessly into everyday life, thereby improving human productivity and overall quality of life. This
capability extends the potential applications of autonomous robots, making them more effective in diverse
scenarios such as assisting individuals with disabilities, enhancing industrial automation, and contributing to
household convenience.
In summation, this research exemplifies a successful approach to the intricate task of doorknob recognition
through the application of deep learning techniques. Future investigations could explore the expansion of the
dataset to encompass a wider array of doorknob types and lighting conditions, thereby enhancing the model’s
applicability across varied real-world scenarios.
Moreover, the integration of the developed model with other robotics systems, such as autonomous naviga-
tion or smart home technologies, presents a promising avenue for creating more sophisticated and functional
solutions.
Continued research is imperative to mitigate the impact of external variables, such as fluctuating lighting
conditions or visual obstacles, on the model’s accuracy and stability, further refining the system’s reliability and
operational efficiency under varying usage conditions. Despite this, experiments conducted demonstrate the
approach’s notable reliability in detecting door handles, even in scenarios where they haven’t been seen before,
the material of the door handles changes, or the lighting conditions vary significantly. Leveraging data from an
annotated set of 5000 images with corresponding coordinates of the door handle bounding box, the algorithm
Vol.:(0123456789)
www.nature.com/scientificreports/
operates with RGBD images and internal camera parameters to accurately determine the handles’ parameters.
Future research plans include conducting additional experiments and further diversifying the test data.
Potential collision hazards arising from accuracy and uncertainty issues include unintended contact with
objects or obstacles due to inaccurate object detection and unpredictable robot behavior. To address these chal-
lenges, countermeasures such as sensor redundancy, dynamic collision avoidance algorithms, safety margins,
adaptive grasping strategies, and continuous monitoring can be implemented to enhance the robot’s ability to
perceive its environment accurately and react effectively to potential collision hazards.
Data availability
The datasets generated and analyzed in this study are not publicly available due to restrictions their containing
information that could compromise the privacy of research participants, but can be obtained from the respective
author upon reasonable request.
References
1. Andreasson, H., Grisetti, G., Stoyanov, T. & Pretto, A. Sensors for mobile robots (2023) [cited 2024 Apr 23] 1–22. http://arxiv.org/
abs/2206.03223.
2. Mochurad, L., Hladun, Y. & Tkachenko, R. An obstacle-finding approach for autonomous mobile robots using 2D LiDAR data.
BDCC 7, 43 (2023).
3. Law, W. T., Fan, K. W., Li, K. S. & Mo, T. Applied design and methodology of delivery robots based on human–robot interaction
in smart cities. EAI Endorsed Trans. Smart Cities https://doi.org/10.4108/eetsc.2649 (2023).
4. Gujarathi, A., Kulkarni, A., Patil, U., Phalak, Y., Deotalu, R., Jain, A. et al. Design and development of autonomous delivery robot
[Internet] (2021) [cited 2024 Apr 23]. http://arxiv.org/abs/2103.09229.
5. Sawik, B. et al. Robots for elderly care: Review, multi-criteria optimization model and qualitative case study. Healthcare 11, 1286
(2023).
6. Su, H.-R. & Chen, K.-Y. Design and implementation of a mobile robot with autonomous door opening ability. Int. J. Fuzzy Syst.
21, 333–342 (2019).
7. Vanus, J., Hercik, R. & Bilik, P. Using interoperability between mobile robot and KNX technology for occupancy monitoring in
smart home care. Sensors 23, 8953 (2023).
8. Çiğdem, Ş, Meidute-Kavaliauskiene, I. & Yıldız, B. Industry 4.0 and industrial robots: A study from the perspective of manufactur-
ing company employees. Logistics 7, 17 (2023).
9. Indha, W. A., Zamzam, N. S., Saptari, A., Alsayaydeh, J. A. & Hassim, N. B. Development of security system using motion sensor
powered by RF energy harvesting. In 2020 IEEE Student Conference on Research and Development (SCOReD) [Internet] 254–258
(IEEE; 2020) [cited 2024 Feb 6]. https://ieeexplore.ieee.org/document/9250984/.
10. Alsayaydeh, J., Abedalrahim, J., Ahmad Idil, A.R., Baharudin, Z. A. et al. Development of programmable home security using GSM
system for early prevention. Asian Res. Publ. Netw. (2021). http://eprints.utem.edu.my/id/eprint/25751/2/JEAS_0121_8470.PDF.
11. Kovtun, V., Izonin, I. & Gregus, M. Model of information system communication in aggressive cyberspace: Reliability, functional
safety, economics. IEEE Access 10, 31494–31502 (2022).
12. Izonin, I., Tkachenko, R., Pidkostelnyi, R., Pavliuk, O., Khavalko, V. & Batyuk, A. Experimental evaluation of the effectiveness of
ANN-based numerical data augmentation methods for diagnostics tasks. In Proceedings of the 4th International Conference on
Informatics & Data-Driven Medicine 223–232 (2021).
13. Mochurad, L., Matviiv, O.-V., Lema, H. & Vilhutska, R. CUDA-based algorithm for Lidar position determination in mobile robot-
ics. MoMLeT&DS 2023, 193–203 (2023).
14. Wang, Y., Wang, L. & Zhao, Y. Research on door opening operation of mobile robotic arm based on reinforcement learning. Appl.
Sci. 12, 5204 (2022).
15. Mohanapriya, S., Mohana Saranya, S., Kumaravel, T. & Sumithra, P. Image detection and segmentation using YOLO v5 for surveil-
lance. ACE 8, 142–147 (2023).
16. Terven, J. & Cordova-Esparza, D. A comprehensive review of YOLO architectures in computer vision: From YOLOv1 to YOLOv8
and YOLO-NAS. MAKE 5, 1680–1716 (2023).
17. Robot simulator CoppeliaSim: Create, compose, simulate, any robot - Coppelia Robotics. [Internet]. [cited 2023 Nov 9]. https://
www.coppeliarobotics.com/.
18. Arduengo, M., Torras, C. & Sentis, L. Robust and adaptive door operation with a mobile robot. Intell. Serv. Robot. 14, 409–425
(2021).
19. Berenson, D., Srinivasa, S. & Kuffner, J. Task space regions: A framework for pose-constrained manipulation planning. Int. J. Robot.
Res. 30, 1435–1460 (2011).
20. Huaman Quispe, A. C., Martinson, E. & Oguchi, K. Learning user preferences for robot–human handovers. In 2017 IEEE/RSJ
International Conference on Intelligent Robots and Systems (IROS) [Internet] 834–839 (IEEE, 2017) [cited 2024 Feb 7]. https://ieeex
plore.ieee.org/document/8202245/.
21. Stuede, M., Nuelle, K., Tappe, S. & Ortmaier, T. Door opening and traversal with an industrial cartesian impedance controlled
mobile robot. In 2019 International Conference on Robotics and Automation (ICRA) [Internet] 966–972 (IEEE, 2019) [cited 2024
Feb 6]. https://ieeexplore.ieee.org/document/8793866/.
22. KMR iiwa | KUKA AG [Internet]. [cited 2023 Nov 14]. https://www.kuka.com/en-us/products/mobility/mobile-robot-systems/
kmr-iiwa.
23. Quintana, B., Prieto, S. A., Adan, A. & Bosche, F. Door detection in 3D colored laser scans for autonomous indoor navigation. In
2016 International Conference on Indoor Positioning and Indoor Navigation (IPIN) [Internet] 1–8 (IEEE, 2016) [cited 2024 Feb 7].
http://ieeexplore.ieee.org/document/7743677/.
24. Mochurad, L., Hladun, Y., Zasoba, Y. & Gregus, M. An approach for opening doors with a mobile robot using machine learning
methods. BDCC 7, 69 (2023).
25. Mochurad, L., Sydor, A. & Ratinskiy, O. A fast parallelized DBSCAN algorithm based on OpenMp for detection of criminals on
streaming services. Front. Big Data 6, 1292923 (2023).
26. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A. & Chen, L.-C. MobileNetV2: Inverted residuals and linear bottlenecks. In 2018
IEEE/CVF Conference on Computer Vision and Pattern Recognition [Internet] 4510–4520 (IEEE, 2018) [cited 2024 Feb 7]. https://
ieeexplore.ieee.org/document/8578572/.
27. Wang, Q., Ihme, M., Chen, Y.-F. & Anderson, J. A TensorFlow simulation framework for scientific computing of fluid flows on
tensor processing units. Comput. Phys. Commun. 274, 108292 (2022).
Vol:.(1234567890)
www.nature.com/scientificreports/
28. 20.04 LTS | Ubuntu [Internet]. [cited 2023 Nov 14]. https://ubuntu.com/blog/tag/20-04-lts.
29. Ahn, M. S., Chae, H., Noh. D., Nam, H. & Hong, D. Analysis and noise modeling of the intel RealSense D435 for mobile robots.
In 2019 16th International Conference on Ubiquitous Robots (UR) [Internet] 707–711 (IEEE, 2019) [cited 2024 Feb 8]. https://ieeex
plore.ieee.org/document/8768489/.
30. Trybała, P., Szrek, J., Remondino, F., Wodecki, J. & Zimroz, R. calibration of a multi-sensor wheeled robot for the 3D mapping of
underground mining tunnels. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, 135–142 (2022).
31. Tekin, B., Sinha, S. N. & Fua, P. Real-time seamless single shot 6D object pose prediction (2017) [cited 2024 Feb 8]. https://arxiv.
org/abs/1711.08848.
32. Hu, Y., Hugonot, J., Fua, P. & Salzmann, M. Segmentation-driven 6D object pose estimation (2018) [cited 2024 Feb 8]. https://
arxiv.org/abs/1812.02541.
Acknowledgements
The authors thank the reviewers for the relevant comments that helped to present the paper better. The National
Research Foundation of Ukraine funds this study from the state budget of Ukraine within the Project No.
2023.03/0029.
Author contributions
Conceptualization, L.M.; methodology, L.M.; software, L.M. and Y.H.; validation, L.M. and Y.H.; formal analysis,
L.M.; investigation, L.M.; resources, Y.H.; data curation, Y.H.; writing—original draft preparation, L.M.; writ-
ing—review and editing L.M.; visualization, Y.H.; supervision, L.M.; All authors have read and agreed to the
published version of the manuscript.
Competing interests
The authors declare no competing interests.
Additional information
Correspondence and requests for materials should be addressed to L.M.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or
format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the
Creative Commons licence, and indicate if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the
material. If material is not included in the article’s Creative Commons licence and your intended use is not
permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Vol.:(0123456789)
© The Author(s) 2024. This work is published under
http://creativecommons.org/licenses/by/4.0/(the “License”). Notwithstanding
the ProQuest Terms and Conditions, you may use this content in accordance
with the terms of the License.