Plenary - Physics-inspired learning on graphs
Michael Bronstein, University of Oxford, UK
-
CIS
IEEE Members: Free
Non-members: FreeLength: 01:01:50
Michael Bronstein, University of Oxford, UK;
ABSTRACT: The message-passing paradigm has been the “battle horse” of deep learning
on graphs for several years, making graph neural networks a big success
in a wide range of applications, from particle physics to protein
design. From a theoretical viewpoint, it established the link to the
Weisfeiler-Lehman hierarchy, allowing to analyse the expressive power of
GNNs. I argue that the very “node-and-edge”-centric mindset of current
graph deep learning schemes may hinder future progress in the field. As
an alternative, I propose physics-inspired “continuous” learning models
that open up a new trove of tools from the fields of differential
geometry, algebraic topology, and differential equations so far largely
unexplored in graph ML.