Two aspects of learning algorithms: generalization under shifts and loss functions

Stefanie Jegelka
Massachusetts Institute of Technology

Graph Neural Networks (GNNs) have become a popular tool for learning algorithmic tasks, in particular related to combinatorial optimization. In this talk, we will focus on the “algorithmic reasoning” task of learning a full discrete algorithm. First, we will focus on the stability of GNNs to data perturbations: what may be an appropriate metric for measuring shift? Under what conditions will a GNN generalize to larger graphs? Second, we will consider loss functions for learning with discrete objects, beyond GNNs. In particular, neural networks work best with continuous, high-dimensional spaces. Can we extend discrete loss functions accordingly?

This talk is based on joint work with Ching-Yao Chuang, Keyulu Xu, Joshua Robinson, Nikos Karalias, Jingling Li, Mozhi Zhang, Simon S. Du, Ken-ichi Kawarabayashi and Andreas Loukas.

Presentation (PDF File)

Back to Artificial Intelligence and Discrete Optimization