Local2Global: a distributed approach for scaling representation learning on graphs
| Authors |
|
|---|---|
| Publication date | 05-2023 |
| Journal | Machine Learning |
| Volume | Issue number | 112 | 5 |
| Pages (from-to) | 1663-1692 |
| Number of pages | 30 |
| Organisations |
|
| Abstract |
We propose a decentralised “local2global” approach to graph representation learning, that one can a-priori use to scale any embedding technique. Our local2global approach proceeds by first dividing the input graph into overlapping subgraphs (or “patches”)
and training local representations for each patch independently. In a
second step, we combine the local representations into a globally
consistent representation by estimating the set of rigid motions that
best align the local representations using information from the patch
overlaps, via group synchronization. A key distinguishing feature of local2global
relative to existing work is that patches are trained independently
without the need for the often costly parameter synchronization during
distributed training. This allows local2global to scale to
large-scale industrial applications, where the input graph may not even
fit into memory and may be stored in a distributed manner. We apply local2global
on data sets of different sizes and show that our approach achieves a
good trade-off between scale and accuracy on edge reconstruction and
semi-supervised classification. We also consider the downstream task of
anomaly detection and show how one can use local2global to highlight anomalies in cybersecurity networks.
|
| Document type | Article |
| Language | English |
| Published at | https://doi.org/10.1007/s10994-022-06285-7 |
| Other links | https://github.com/LJeub/Local2Global https://www.scopus.com/pages/publications/85148623035 |
| Downloads |
s10994-022-06285-7
(Final published version)
|
| Permalink to this page | |
