## DeepWalking Backwards: From Embeddings Back to Graphs

### Sudhanshu Chanpuriya · Cameron Musco · Konstantinos Sotiropoulos · Charalampos Tsourakakis

Keywords: [ Networks and Relational Learning ] [ Algorithms ] [ Regularization ] [ Uncertainty Estimation ] [ Algorithms -> Missing Data; Theory ]

[ Abstract ]
[ Paper ] [ ]
Tue 20 Jul 9 a.m. PDT — 11 a.m. PDT

Spotlight presentation: Deep Learning Algorithms 3
Tue 20 Jul 7 a.m. PDT — 8 a.m. PDT

Abstract: Low-dimensional node embeddings play a key role in analyzing graph datasets. However, little work studies exactly what information is encoded by popular embedding methods, and how this information correlates with performance in downstream learning tasks. We tackle this question by studying whether embeddings can be inverted to (approximately) recover the graph used to generate them. Focusing on a variant of the popular DeepWalk method \cite{PerozziAl-RfouSkiena:2014, QiuDongMa:2018}, we present algorithms for accurate embedding inversion -- i.e., from the low-dimensional embedding of a graph $G$, we can find a graph $\tilde G$ with a very similar embedding. We perform numerous experiments on real-world networks, observing that significant information about $G$, such as specific edges and bulk properties like triangle density, is often lost in $\tilde G$. However, community structure is often preserved or even enhanced. Our findings are a step towards a more rigorous understanding of exactly what information embeddings encode about the input graph, and why this information is useful for learning tasks.