Topological deep learning explained

Topological Deep Learning (TDL).[1] is a research field that extends deep learning to handle complex, non-Euclidean data structures. Traditional deep learning models, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), excel in processing data on regular grids and sequences. However, scientific and real-world data often exhibit more intricate data domains encountered in scientific computations, including point clouds, meshes, time series, scalar fields graphs, or general topological spaces like simplicial complexes and CW complexes.[2] TDL addresses this by incorporating topological concepts to process data with higher-order relationships, such as interactions among multiple entities and complex hierarchies. This approach leverages structures like simplicial complexes and hypergraphs to capture global dependencies and qualitative spatial properties, offering a more nuanced representation of data. TDL also encompasses methods from computational and algebraic topology that permit studying properties of neural networks and their training process, such as their predictive performance or generalization properties.,[3] [4] [5] [6] [7] [8] .

History and motivation

Traditional techniques from deep learning often operate under the assumption that a dataset is residing in a highly-structured space (like images, where convolutional neural networks exhibit outstanding performance over alternative methods) or a Euclidean space. The prevalence of new types of data, in particular graphs, meshes, and molecules, resulted in the development of new techniques, culminating in the field of geometric deep learning, which originally proposed a signal-processing perspective for treating such data types.[9] While originally confined to graphs, where connectivity is defined based on nodes and edges, follow-up work extended concepts to a larger variety of data types, including simplicial complexes and CW complexes, with recent work proposing a unified perspective of message-passing on general combinatorial complexes.

An independent perspective on different types of data originated from topological data analysis, which proposed a new framework for describing structural information of data, i.e., their "shape," that is inherently aware of multiple scales in data, ranging from local information to global information.[10] While at first restricted to smaller datasets, subsequent work developed new descriptors that efficiently summarized topological information of datasets to make them available for traditional machine-learning techniques, such as support vector machines or random forests. Such descriptors ranged from new techniques for feature engineering over new ways of providing suitable coordinates for topological descriptors,[11] [12] [13] or the creation of more efficient dissimilarity measures.[14] [15] [16] [17]

Contemporary research in this field is largely concerned with either integrating information about the underlying data topology into existing deep-learning models or obtaining novel ways of training on topological domains.

Learning on topological spaces

Focusing broadly on topology in the sense of point set topology, an active branch of TDL is concerned with learning on topological spaces, or, put differently, on certain topological domains.

An introduction to topological domains

One of the core concepts in topological deep learning is the domain upon which this data is defined and supported. In case of Euclidian data, such as images, this domain is a grid, upon which the pixel value of the image is supported. In a more general setting this domain might be a topological domain. Next, we introduce the most common topological domains that are encountered in a deep learning setting. These domains include, but not limited to, graphs, simplicial complexes, cell complexes, combinatorial complexes and hypergraphs.

Given a finite set S of abstract entities, a neighborhood function

l{N}

on S is an assignment that attach to every point

x

in S a subset of S or a relation. Such a function can be induced by equipping S with an auxiliary structure. Edges provide one way of defining relations among the entities of S. More specifically, edges in a graph allow one to define the notion of neighborhood using, for instance, the one hop neighborhood notion. Edges however, limited in their modeling capacity as they can only be used to model binary relations among entities of S since every edge is connected typically to two entities. In many applications, it is desirable to permit relations that incorporate more than two entities. The idea of using relations that involve more than two entities is central to topological domains. Such higher-order relations allow for a broader range of neighborhood functions to be defined on S to capture multi-way interactions among entities of S.

Next we review the main properties, advantages, and disadvantages of some commonly studied topological domains in the context of deep learning, including (abstract) simplicial complexes, regular cell complexes, hypergraphs, and combinatorial complexes.

Comparisons among topological domains

Each of the enumerated topological domains has its own characteristics, advantages, and limitations:

Hierarchical structure and set-type relations

The properties of simplicial complexes, cell complexes, and hypergraphs give rise to two main features of relations on higher-order domains, namely hierarchies of relations and set-type relations.

Rank function

A rank function on a higher-order domain X is an order-preserving function rk: XZ, where rk(x) attaches a non-negative integer value to each relation x in X, preserving set inclusion in X. Cell and simplicial complexes are common examples of higher-order domains equipped with rank functions and therefore with hierarchies of relations.

Set-type relations

Relations in a higher-order domain are called set-type relations if the existence of a relation is not implied by another relation in the domain. Hypergraphs constitute examples of higher-order domains equipped with set-type relations. Given the modeling limitations of simplicial complexes, cell complexes, and hypergraphs, we develop the combinatorial complex, a higher-order domain that features both hierarchies of relations and set-type relations.

The learning tasks in TDL can be broadly classified into three categories:

In practice, to perform the aforementioned tasks, deep learning models designed for specific topological spaces must be constructed and implemented. These models, known as topological neural networks, are tailored to operate effectively within these spaces.

Topological neural networks

Central to TDL are topological neural networks (TNNs), specialized architectures designed to operate on data structured in topological domains. Unlike traditional neural networks tailored for grid-like structures, TNNs are adept at handling more intricate data representations, such as graphs, simplicial complexes, and cell complexes. By harnessing the inherent topology of the data, TNNs can capture both local and global relationships, enabling nuanced analysis and interpretation.

Message passing topological neural networks

In a general topological domain, higher-order message passing involves exchanging messages among entities and cells using a set of neighborhood functions.

Definition: Higher-Order Message Passing on a General Topological Domain

Let

l{X}

be a topological domain. We define a set of neighborhood functions

l{N}=\{l{N}1,\ldots,l{N}n\}

on

l{X}

. Consider a cell

x

and let

y\inl{N}k(x)

for some

l{N}k\inl{N}

. A message

mx,y

between cells

x

and

y

is a computation dependent on these two cells or the data supported on them. Denote

l{N}(x)

as the multi-set

\{\{l{N}1(x),\ldots,l{N}n(x)\}\}

, and let
(l)
h
x
represent some data supported on cell

x

at layer

l

. Higher-order message passing on

l{X}

, induced by

l{N}

, is defined by the following four update rules:

mx,y=\alphal{Nk}(h

(l)
x
(l)
,h
y

)

k
m
x

=oplusyk(x)}mx,y

, where

oplus

is the intra-neighborhood aggregation function.

mx=otimesk\inl{N}}

k
m
x
, where

otimes

is the inter-neighborhood aggregation function.
(l+1)
h
x

=\beta

(l)
(h
x

,mx)

, where

\alphal{Nk},\beta

are differentiable functions.

Some remarks on Definition above are as follows.

First, Equation 1 describes how messages are computed between cells

x

and

y

. The message

mx,y

is influenced by both the data
(l)
h
x
and
(l)
h
y
associated with cells

x

and

y

, respectively. Additionally, it incorporates characteristics specific to the cells themselves, such as orientation in the case of cell complexes. This allows for a richer representation of spatial relationships compared to traditional graph-based message passing frameworks.

Second, Equation 2 defines how messages from neighboring cells are aggregated within each neighborhood. The function

oplus

aggregates these messages, allowing information to be exchanged effectively between adjacent cells within the same neighborhood.

Third, Equation 3 outlines the process of combining messages from different neighborhoods. The function

otimes

aggregates messages across various neighborhoods, facilitating communication between cells that may not be directly connected but share common neighborhood relationships.

Fourth, Equation 4 specifies how the aggregated messages influence the state of a cell in the next layer. Here, the function

\beta

updates the state of cell

x

based on its current state
(l)
h
x
and the aggregated message

mx

obtained from neighboring cells.

Non-message passing topological neural networks

While the majority of TNNs follow the message passing paradigm from graph learning, several models have been suggested that do not follow this approach. For instance, Maggs et al.[18] leverage geometric information from embedded simplicial complexes, i.e., simplicial complexes with high-dimensional features attached to their vertices.This offers interpretability and geometric consistency without relying on message passing. Furthermore, in a contrastive loss-based method was suggested to learn the simplicial representation.

Learning on topological descriptors

Motivated by the modular nature of deep neural networks, initial work in TDL drew inspiration from topological data analysis, and aimed to make the resulting descriptors amenable to integration into deep-learning models. This led to work defining new layers for deep neural networks. Pioneering work by Hofer et al.,[19] for instance, introduced a layer that permitted topological descriptors like persistence diagrams or persistence barcodes to be integrated into a deep neural network. This was achieved by means of end-to-end-trainable projection functions, permitting topological features to be used to solve shape classification tasks, for instance. Follow-up work expanded more on the theoretical properties of such descriptors and integrated them into the field of representation learning.[20] Other such topological layers include layers based on extended persistent homology descriptors,[21] persistence landscapes,[22] or coordinate functions.[23] In parallel, persistent homology also found applications in graph-learning tasks. Noteworthy examples include new algorithms for learning task-specific filtration functions for graph classification or node classification tasks.[24] [25]

Applications

TDL is rapidly finding new applications across different domains, including data compression, enhancing the expressivity and predictive performance of graph neural networks,[26] [27] [28] action recognition, and trajectory prediction.

References

  1. 2304.10031 . cs.LG . M. . Papillon . S. . Sanborn . Architectures of topological deep learning: A survey on topological neural networks . Hajij . M. . Miolane . N. . 2023.
  2. Uray . Martin . Giunti . Barbara . Kerber . Michael . Huber . Stefan . 2024-10-01 . Topological Data Analysis in smart manufacturing: State of the art and future directions . Journal of Manufacturing Systems . 76 . 75–91 . 2310.09319 . 10.1016/j.jmsy.2024.07.006 . 0278-6125.
  3. Bianchini . Monica . Scarselli . Franco . 2014 . On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures . IEEE Transactions on Neural Networks and Learning Systems . 25 . 8 . 1553–1565 . 10.1109/TNNLS.2013.2293637 . 2162-237X.
  4. Naitzat . Gregory . Zhitnikov . Andrey . Lim . Lek-Heng . 2020 . Topology of Deep Neural Networks . Journal of Machine Learning Research . 21 . 1 . 184:7503–184:7542 . 1532-4435.
  5. Birdal . Tolga . Lou . Aaron . Guibas . Leonidas J . Simsekli . Umut . 2021 . Intrinsic Dimension, Persistent Homology and Generalization in Neural Networks . Advances in Neural Information Processing Systems . Curran Associates, Inc. . 34 . 6776–6789.
  6. Ballester . Rubén . Clemente . Xavier Arnal . Casacuberta . Carles . Madadi . Meysam . Corneanu . Ciprian A. . Escalera . Sergio . 2024 . Predicting the generalization gap in neural networks using topological data analysis . Neurocomputing . en . 596 . 127787 . 2203.12330 . 10.1016/j.neucom.2024.127787.
  7. Rieck . Bastian . Togninalli . Matteo . Bock . Christian . Moor . Michael . Horn . Max . Gumbsch . Thomas . Borgwardt . Karsten . 2018-09-27 . Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology . International Conference on Learning Representations . en.
  8. Dupuis . Benjamin . Deligiannidis . George . Simsekli . Umut . 2023-07-03 . Generalization Bounds using Data-Dependent Fractal Dimensions . Proceedings of the 40th International Conference on Machine Learning . en . PMLR . 8922–8968.
  9. Bronstein . Michael M. . Bruna . Joan . LeCun . Yann . Szlam . Arthur . Vandergheynst . Pierre . 2017 . Geometric Deep Learning: Going beyond Euclidean data . IEEE Signal Processing Magazine . 34 . 4 . 18–42 . 10.1109/MSP.2017.2693418 . 1053-5888. 1611.08097 .
  10. Carlsson . Gunnar . 2009-01-29 . Topology and data . Bulletin of the American Mathematical Society . en . 46 . 2 . 255–308 . 10.1090/S0273-0979-09-01249-X . 0273-0979. free .
  11. Adcock . Aaron . Carlsson . Erik . Carlsson . Gunnar . 2016 . The ring of algebraic functions on persistence bar codes . Homology, Homotopy and Applications . en . 18 . 1 . 381–402 . 10.4310/HHA.2016.v18.n1.a21. 1304.0530 .
  12. Adams . Henry . Emerson . Tegan . Kirby . Michael . Neville . Rachel . Peterson . Chris . Shipman . Patrick . Chepushtanova . Sofya . Hanson . Eric . Motta . Francis . Ziegelmeier . Lori . 2017 . Persistence Images: A Stable Vector Representation of Persistent Homology . Journal of Machine Learning Research . 18 . 8 . 1–35 . 1533-7928.
  13. Bubenik . Peter . 2015 . Statistical Topological Data Analysis using Persistence Landscapes . Journal of Machine Learning Research . 16 . 3 . 77–102 . 1533-7928.
  14. Kwitt . Roland . Huber . Stefan . Niethammer . Marc . Lin . Weili . Bauer . Ulrich . 2015 . Statistical Topological Data Analysis - A Kernel Perspective . Advances in Neural Information Processing Systems . Curran Associates, Inc. . 28.
  15. Carrière . Mathieu . Cuturi . Marco . Oudot . Steve . 2017-07-17 . Sliced Wasserstein Kernel for Persistence Diagrams . Proceedings of the 34th International Conference on Machine Learning . en . PMLR . 664–673.
  16. Kusano . Genki . Fukumizu . Kenji . Hiraoka . Yasuaki . 2018 . Kernel Method for Persistence Diagrams via Kernel Embedding and Weight Factor . Journal of Machine Learning Research . 18 . 189 . 1–41 . 1533-7928.
  17. Le . Tam . Yamada . Makoto . 2018 . Persistence Fisher Kernel: A Riemannian Manifold Kernel for Persistence Diagrams . Advances in Neural Information Processing Systems . Curran Associates, Inc. . 31.
  18. Maggs . Kelly . Hacker . Celia . Rieck . Bastian . 2023-10-13 . Simplicial Representation Learning with Neural k-Forms . International Conference on Learning Representations . en.
  19. Hofer . Christoph . Kwitt . Roland . Niethammer . Marc . Uhl . Andreas . 2017 . Deep Learning with Topological Signatures . Advances in Neural Information Processing Systems . Curran Associates, Inc. . 30.
  20. Hofer . Christoph D. . Kwitt . Roland . Niethammer . Marc . 2019 . Learning Representations of Persistence Barcodes . Journal of Machine Learning Research . 20 . 126 . 1–45 . 1533-7928.
  21. Carriere . Mathieu . Chazal . Frederic . Ike . Yuichi . Lacombe . Theo . Royer . Martin . Umeda . Yuhei . 2020-06-03 . PersLay: A Neural Network Layer for Persistence Diagrams and New Graph Topological Signatures . Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics . en . PMLR . 2786–2796.
  22. Kim . Kwangho . Kim . Jisu . Zaheer . Manzil . Kim . Joon . Chazal . Frederic . Wasserman . Larry . 2020 . PLLay: Efficient Topological Layer based on Persistent Landscapes . Advances in Neural Information Processing Systems . Curran Associates, Inc. . 33 . 15965–15977.
  23. Gabrielsson . Rickard Brüel . Nelson . Bradley J. . Dwaraknath . Anjan . Skraba . Primoz . 2020-06-03 . A Topology Layer for Machine Learning . Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics . en . PMLR . 1553–1563.
  24. Hofer . Christoph . Graf . Florian . Rieck . Bastian . Niethammer . Marc . Kwitt . Roland . 2020-11-21 . Graph Filtration Learning . Proceedings of the 37th International Conference on Machine Learning . en . PMLR . 4314–4323.
  25. Immonen . Johanna . Souza . Amauri . Garg . Vikas . 2023-12-15 . Going beyond persistent homology using persistent homology . Advances in Neural Information Processing Systems . en . 36 . 63150–63173.
  26. Bodnar . Cristian . Frasca . Fabrizio . Wang . Yuguang . Otter . Nina . Montufar . Guido F. . Lió . Pietro . Bronstein . Michael . 2021-07-01 . Weisfeiler and Lehman Go Topological: Message Passing Simplicial Networks . Proceedings of the 38th International Conference on Machine Learning . en . PMLR . 1026–1037.
  27. Bodnar . Cristian . Frasca . Fabrizio . Otter . Nina . Wang . Yuguang . Liò . Pietro . Montufar . Guido F . Bronstein . Michael . 2021 . Weisfeiler and Lehman Go Cellular: CW Networks . Advances in Neural Information Processing Systems . Curran Associates, Inc. . 34 . 2625–2640.
  28. Horn . Max . Brouwer . Edward De . Moor . Michael . Moreau . Yves . Rieck . Bastian . Borgwardt . Karsten . 2021-10-06 . Topological Graph Neural Networks . International Conference on Learning Representations . en.