A Brief Introduction To Geometric Deep Learning - by Jason McEwen - Towards Data Science
A Brief Introduction To Geometric Deep Learning - by Jason McEwen - Towards Data Science
Deep Learning
AI for complex data
318 2
Open in app Sign up Sign in
Search Write
The term geometric deep learning was first coined by Michael Bronstein, a
pioneer of the field (see his posts for interesting insights on a lot of the latest
deep learning research, as well as extensive overviews of the field). In this
post, rather then getting deep into the technical weeds, we present a very
brief introduction to geometric deep learning. We largely follow the
excellent recent book by Bronstein and colleagues [1] but provide our own
unique take, and focus on high level concepts rather than technical details.
Geometric Priors
2. Stability
3. Multiscale representations
Illustration of stability of mapping to representation space. Small distortions are responsible for intra-class
variations, whereas large distortions are responsible for inter-class variations. Stability of the mapping is
required to ensure measures of similarity between data instances, i.e. the size of the distortion between them,
is preserved in the representation space in order to facilitate effective learning. [Diagram created by author for
[2].]
A third common geometric prior is to encode a multiscale, hierarchical
representation of data. In a data instance, many of the datum are not
independent but are correlated in complex ways. Consider an image for
example. Each image pixel is not independent but rather nearby pixels are
often related and very similar. Different notions of “nearby” are also possible
depending on content structure. Effective representational spaces can
therefore be constructed by capturing the multiscale, hierarchical nature of
much data.
A multiscale, hierarchical representation of an image. A low-resolution version of the original image is shown
in the top-left corner and then remaining image content at different resolutions is captured in the other panels
of the diagram. Similar representations can be exploited to provide effective representational spaces for
learning. [Source wikipedia.]
Categories of geometric deep learning. [Image sourced from article [1], with permission, with annotated
overview and examples added.]
The group category covers homogenous spaces with global symmetries. The
canonical example of this category is the sphere (covered in greater detail in
our previous article [3]). Spherical data arise in myrad applications, not only
when data is acquired directly on the sphere (such as over the Earth or by
360° cameras that capture panoramic photos and videos), but also when
considering spherical symmetries (such as in molecular chemistry or
magnetic resonance imaging). While the sphere is the most common group
setting, other groups and their corresponding symmetries can also be
considered.
The final geodesics and gauges category involves deep learning on more
complex shapes, such as more general maniolds and 3D meshes. Such
approaches can be of great use in computer vision and graphics, for
example, where one can perform deep learning with 3D models and their
deformations.
Building Blocks
While there are a number of different categories of geometric deep learning,
as described above, and different types of geometric priors than can be
exploited, all approaches to geometric deep learning essentially adopt
different incarnations of the following fundamental underlying building
blocks.
All approaches to geometric deep learning leverage a core set of fundamental underlying building blocks.
[Photo by Sen on Unsplash.]
Future perspectives
Deep learning is now commonplace for standard types of data, such as
structured, sequential and image data. However, to exend the application of
deep learning to other more complex — geometric — datasets, the geometry
of such data must be encoded in deep learning models, giving rise to the
field of geometric deep learning.
Geometric deep learning is a topical and rapidly evolving field, where much
progress has been made. However, many unsolved questions remain, not
only in models themselves but also around scalability and practical
application. We will address these issuses in upcoming articles, showing
how solving such issues is critical to unlocking the remarkable potential of
deep learning for a host of new applications.
References
[1] Bronstein, Bruna, Cohen, Velickovic, Geometric Deep Learning: Grids,
Groups, Graphs, Geodesics, and Gauges (2021), arXix:2104.13478
Jason McEwen in Towards Data Science Tim Sumner in Towards Data Science
7 min read · Mar 14, 2024 10 min read · Mar 31, 2024
106 1 2.6K 35
Cristian Leo in Towards Data Science Jason McEwen in Towards Data Science
· 28 min read · Mar 29, 2024 9 min read · Apr 26, 2022
2.8K 17 35 1
See all from Jason McEwen See all from Towards Data Science
Alex Honchar in Towards Data Science Abdulkader Helwan in Python in Plain English
7 min read · Mar 15, 2024 · 3 min read · Feb 23, 2024
2K 11 382
Lists
Predictive Modeling w/ Natural Language Processing
Python 1380 stories · 874 saves
20 stories · 1111 saves
584 6 83 4
Rosaria Silipo in Low Code for Data Science Benedict Neo in bitgrit Data Science Publication
6 min read · Mar 11, 2024 11 min read · Mar 11, 2024