Sébastien Lerique, Jacobo Levy-Abitbol, Márton Karsai & Éric Fleury
network—feature dependencies
In Twitter (courtesy of Jacobo)
2-way: how much do network structure and feature structure correlate
3-way: how well does feature set 1 + network structure predict feature set 2
(and how much does network structure help in predicting)
Combined: how much can we compress network + features (down to what dimension)
network—feature dependencies
network—feature independence
Use deep learning to explore these
Graph convolutional neural networks + Auto-encoders
Reformulate our questions
Where are we at & what's next
x
y
green
red
\(H^{(l+1)} = \sigma(H^{(l)}W^{(l)})\)
\(H^{(0)} = X\)
\(H^{(L)} = Z\)
Inspired by colah's blog
\(H^{(l+1)} = \sigma(H^{(l)}W^{(l)})\)
\(H^{(0)} = X\)
\(H^{(L)} = Z\)
\(H^{(l+1)} = \sigma(\color{DarkRed}{\tilde{D}^{-\frac{1}{2}}\tilde{A}\tilde{D}^{-\frac{1}{2}}}H^{(l)}W^{(l)})\)
\(H^{(0)} = X\)
\(H^{(L)} = Z\)
\(\color{DarkGreen}{\tilde{A} = A + I}\)
\(\color{DarkGreen}{\tilde{D}_{ii} = \sum_j \tilde{A}_{ij}}\)
Kipf & Welling (2017)
Four well-marked communities of size 10, small noise
Five well-marked communities of size 20, moderate noise
Overlapping communities of size 12, small noise
Two feature communities in a near-clique, small noise
From blog.keras.io
high dimension
high dimension
low dimension
MNIST Examples by Josef Steppan (CC-BY-SA 4.0)
60,000 training images
28x28 pixels
784 dims
784 dims
2D
From blog.keras.io
node features
embedding
GCN
node features
adjacency matrix
Socio-economic status
Language style
Topics
Socio-economic status
Language style
Topics
Compressed & combined representation of nodes + network
a.k.a., questions we can (will be able to) ask
2-way: how much do network structure and feature structure correlate
Combined: how much can we compress network + features
(down to what dimension)
3-way: how well does feature set 1 + network structure predict feature set 2
(and how much does network structure help in predicting)
Force a 2D embedding with one axis for each type of information
AE input = feature set 1
AE output = feature set 2
with / without convolutions
Plot the embedding dimension vs. reconstruction accuracy
✔ Graph convolutions exploration
✔ Variational AE basic implementation
✔ Feature + Adjacency reconstruction
✔ Adjacency reconstruction normalisation
✔ Sensitivity analysis
✔ Mini-batch training
✔ Smarter mini-batch sampling
👷 Exploring small real data sets BlogCatalog, Flickr, YouTube
👷 2nd sensitivity analysis for mini-batching on real datasets
👷 Refining target questions
👷 Choosing language features
👷 Scaling up to Twitter
node2vec, Grover & Leskovec (2016)
walk back \(\propto \frac{1}{p}\)
walk out \(\propto \frac{1}{q}\)
walk in triangle \(\propto 1\)
Non-biased
p=1, q=1
Triangles
p=100, q=100
Walk out
p=1, q=.01
Walk back
p=.01, q=1
Jacobo, Márton, Éric