Immersive Assistance System For Intuitive Robot Programming Using Mixed-Reality and Digital Twin
Immersive Assistance System For Intuitive Robot Programming Using Mixed-Reality and Digital Twin
Twin
Niklas Hübel Siegfried Müller Eric Guiffo Kaigom
Dept. of Comp. Science & Engineering Dept. of Comp. Science & Engineering Dept. of Comp. Science & Engineering
Frankfurt University of Applied Sciences Frankfurt University of Applied Sciences Frankfurt University of Applied Sciences
Frankfurt a.M., Germany Frankfurt a.M., Germany Frankfurt a.M., Germany
[email protected] [email protected] [email protected]
acumen while remaining natural, responsive, and skillful Twin (DT) of the robot acts as planning and control interface
enough to accommodate subtle changes in practice. between the virtual and physical operational spaces (see
fig. 1). Specifically, the DT can be used to specify and
II. C ONTRIBUTIONS analyze desired motions while the physical robot is carrying
To address previously mentioned motion planning chal- out productive tasks. Analyzed and potentially optimized
lenges, this paper (e.g., for collision avoidance or energy savings) joint states
• describes the conception and implementation of a of the DT are sent to the controller of the physical robot
framework that allows for a fast and reconfigurable through an application- and robot-independent extensible
programming of robots without writing any line of communication layer that supports standard information
code, as shown in fig. 1. The responsive approach models (e.g., URDF) and communication protocols like TCP
leverages Mixed Reality (MR) for programming and (see fig. 4 and table I). Downtimes are therefore avoided
adjusting the desired robot motion in the augmented through the loose coupling.
3D operational space of the robot. Since the approach III. R ELATED WORK
is enriched with virtual cues for planning purposes, it is Approaches have been proposed with the aim of lever-
easily and fully accessible even to novices. This helps aging AR to simplify robot programming efforts. Markers
mitigate the lack of skilled collaborators previously [5], virtual artifacts [6] and visual metaphors [7] have
mentioned in section I. been employed to augment the physical environment in
• demonstrates the simplified incorporation of human
which the robot operates. Experts and non-experts can then
dexterity and ingenuity in motion planning without use artifacts and metaphors to program the robot. Despite
requiring any technical background in robotics. Citizens their intuitiveness, artifacts and metaphors still represent a
are thereby empowered with individualized capabilities cognitive load to non-experts and require a certain level
to move a manipulator from a desired Cartesian frame of interpretability and duration to familiarize with them. In
to another while harnessing knowledge from multiple many practical circumstances, it is necessary to design task
domains. This fosters creativity, innovation, and com- specific markers, artifacts, and metaphors, which unfortu-
fort in planning robotics-driven applications in industry, nately increases complexity and roll out time in the case of
society, and in-between. changing demands or a high product mix.
In fact, the planning is intuitive and versatile since natural Fig. 3. Loosely coupled planning, communication, and execution layers.
hand motions (see fig. 2) are made for the flexible and
seamless placement of targeted Cartesian frames in the Recently, [8], [9] have proposed AR-based approaches to
virtual planning space while the robot remains in production easily program (mobile) manipulators. However, the motion
in the physical operational space, as illustrated in fig. 3. planning framework (e.g., way-points) is not explicitly pre-
Immersive technologies are leveraged to this end with a sented or demonstrated. By contrast, we show in this work
twofold objective. First, to decouple the virtual motion how the combination of MR and DT technologies helps in-
planning space from the physical operational space of the tuitively program and move a robot from an initial Cartesian
robot (see fig. 3) using Augmented Reality (AR). Second, frame to another without writing any line of code. Although
to immerse human operators in the virtual space that can way-points for motion planning have been presented in [10],
interact on demand with the physical one in a MR way, as their placement is realized by clicking in the scene. By
depicted in fig. 1. contrast, we propose an approach in which novices place
The loose coupling between virtual planning and physical way-points through natural hand motions. Therefore, human
operational spaces heavily contrasts with standard robot pro- dexterity is combined with cognitive capabilities and domain
gramming in which the robot must be physically maneuvered knowledge to economically yield a meaningful placement of
to the desired frames for teach-in. In this work, a Digital way-points.
Authorized licensed use limited to: R V College of Engineering. Downloaded on March 19,2024 at 07:25:34 UTC from IEEE Xplore. Restrictions apply.
2023 IEEE 3rd International Conference on Digital Twins and Parallel Intelligence (DTPI)
IV. T ECHNOLOGIES AND SYSTEM ARCHITECTURE with the connectivity architecture (see fig. 5) of the assis-
tance system. On the left hand side of fig. 4, a server with
The system is composed of three layers, as shown in fig. 3. the host operating system, i.e., a Windows 11 22H2, runs a
The first bottom layer hosts the Kinova Gen3 lightweight Windows Subsystem for Linux. The robot is controlled via
robot (see fig. 1) controlled via ROS (Robot Operating ROS, as previously mentioned. ROS runs in a subsystem us-
System). The third top layer includes the AR glass with built- ing Ubuntu 18.04. The HoloLens is located on the client side
in software packages, as well as the 3D VR/AR development in fig. 4. The operating system of the HoloLens is Windows
environment Unity for the visualization of the DT of the Holographic 22H2. The developed AR/MR application runs
physical robot via the HoloLens 2. Data exchange between on the HoloLens 2 (see figs. 4 and 5). It is created and tested
the robot control unit and the AR glasses is carried out under the host operating system with Unity 2021.2.18f1. The
through the intermediate second layer. This occurs via a server-client communication is carried out via a TCP-based
wireless TCP network connection. For this purpose, a TCP connection. Tab. I highlights involved software packages.
endpoint developed by Unity Technologies is available on
the ROS side (bottom layer) [11]. On the Unity side (top A. Network communication
layer), a package provided by Unity Technologies called In the developed framework, employed hard- and software
TCP connector handles the data traffic [12]. are configured with appropriate IP addresses for a seamless
Fig. 4 shows the communication architecture associated and reliable communication between server and client based
Authorized licensed use limited to: R V College of Engineering. Downloaded on March 19,2024 at 07:25:34 UTC from IEEE Xplore. Restrictions apply.
2023 IEEE 3rd International Conference on Digital Twins and Parallel Intelligence (DTPI)
TABLE I
U SED SOFTWARE AND MULTIMEDIA COMPONENTS AND FORMATS .
Description Version
Authorized licensed use limited to: R V College of Engineering. Downloaded on March 19,2024 at 07:25:34 UTC from IEEE Xplore. Restrictions apply.
2023 IEEE 3rd International Conference on Digital Twins and Parallel Intelligence (DTPI)
Authorized licensed use limited to: R V College of Engineering. Downloaded on March 19,2024 at 07:25:34 UTC from IEEE Xplore. Restrictions apply.
2023 IEEE 3rd International Conference on Digital Twins and Parallel Intelligence (DTPI)
Authorized licensed use limited to: R V College of Engineering. Downloaded on March 19,2024 at 07:25:34 UTC from IEEE Xplore. Restrictions apply.