SlideShare ist ein Scribd-Unternehmen logo
1 von 38
Downloaden Sie, um offline zu lesen
Localized methods for
diffusions in large graphs
David F. Gleich!
Purdue University!
Joint work with 
Kyle Kloster @"
Purdue &
Michael
Mahoney @
Berkeley
supported by "
NSF CAREER
CCF-1149756
Code "www.cs.purdue.edu/homes/dgleich/codes/nexpokit !
"www.cs.purdue.edu/homes/dgleich/codes/l1pagerank!
David Gleich · Purdue
1
MMDS 2014
Image from rockysprings, deviantart, CC share-alike
Everything in the world can be
explained by a matrix, and we see
how deep the rabbit hole goes
The talk ends, you
believe -- whatever
you want to.
3
Graph diffusions
David Gleich · Purdue
4
f =
1X
k=0
↵k Pk
s
ate
t in
on
work, or mesh, from a typical problem in scientific computing
high
low
A – adjacency matrix!
D – degree matrix!
P – column stochastic operator
s – the “seed” (a sparse vector)
f – the diffusion result
𝛼k – the path weights
P = AD 1
Px =
X
j!i
1
dj
xj
Graph diffusions help:
1.  Attribute prediction
2.  Community detection
3.  “Ranking”
4.  Find small conductance sets
MMDS 2014
Graph diffusions
David Gleich · Purdue
5
ate
t in
on
work, or mesh, from a typical problem in scientific computing
high
low
h = e t
1X
k=0
tk
k!
Pk
s
h = e t
exp{tP}s
PageRank
Heat kernel
x = (1 )
1X
k=0
k
Pk
s
(I P)x = (1 )s
P = AD 1
Px =
X
j!i
1
dj
xj
MMDS 2014
Graph diffusions
David Gleich · Purdue
6
h = e t
1X
k=0
tk
k!
Pk
s
h = e t
exp{tP}s
PageRank
Heat kernel
0 20 40 60 80 100
10
−5
10
0
t=1 t=5 t=15 α=0.85
α=0.99
Weight
Length
x = (1 )
1X
k=0
k
Pk
s
(I P)x = (1 )s
MMDS 2014
Uniformly localized "
solutions in livejournal
1 2 3 4 5
x 10
6
0
0.5
1
1.5
nnz = 4815948
magnitude
plot(x)
10
0
10
1
10
2
10
3
10
4
10
5
10
6
10
−14
10
−12
10
−10
10
−8
10
−6
10
−4
10
−2
10
0
1−normerror
largest non−zeros retained
10
0
10
1
10
2
10
3
10
4
10
5
10
6
10
−14
10
−12
10
−10
10
−8
10
−6
10
−4
10
−2
10
0
1−normerror
largest non−zeros retained
x = exp(P)ec
David Gleich · Purdue
7
nnz(x) = 4, 815, 948
Gleich & Kloster,
arXiv:1310.3423
MMDS 2014
Our mission!
Find the solution with work "
roughly proportional to the "
localization, not the matrix.
David Gleich · Purdue
8
MMDS 2014
Two types of localization
David Gleich · Purdue
9
kx x⇤
k1  " kD 1
(x x⇤
)k1  "
x ⇡ x⇤
Uniform (Strong)! Entry-wise (Weak)!
Localized vectors are not sparse, but they
can be approximated by sparse vectors. 
Good global approximation
using only a local region.
“Hard” to prove.
“Need” a graph property.
Good approximation for
cuts and communities.
“Easy” to prove.
“Fast” algorithms
MMDS 2014
We have four results
1.  A new interpretation for the PageRank
diffusion in relationship with a mincut
problem.
2.  A new understanding of the scalable,
localized PageRank “push” method
3.  A new algorithm for the heat kernel
diffusion in a degree weighted norm.
4.  Algorithms for diffusions as functions of
matrices (K. Kloster’s poster on Thurs.) 
David Gleich · Purdue
10
Undirected
graphs only

Entry-wise
localization
Directed,
uniform
localization
MMDS 2014
Our algorithms for uniform localization"
www.cs.purdue.edu/homes/dgleich/codes/nexpokit
10
0
10
1
10
2
10
3
10
4
10
5
10
6
10
−8
10
−6
10
−4
10
−2
10
0
non−zeros
1−normerror
gexpm
gexpmq
expmimv
10
0
10
1
10
2
10
3
10
4
10
5
10
6
10
−8
10
−6
10
−4
10
−2
10
0
non−zeros
1−normerror
David Gleich · Purdue
11
MMDS 2014
work = O
⇣
log(1
" )(1
" )3/2
d2
(log d)2
⌘
nnz = O
⇣
log(1
" )(1
" )3/2
d(log d)
⌘
PageRank, mincuts, and the
push method via
Algorithmic Anti-Differentiation
David Gleich · Purdue
12
Gleich & Mahoney,
ICML 2014
 MMDS 2014
The PageRank problem & "
the Laplacian on undirected graphs
Combinatorial Laplacian L = D - A!
David Gleich · Purdue
13

The PageRank random surfer
1.  With probability beta, follow a random-walk step
2.  With probability (1-beta), jump randomly ~ dist. s.
Goal find the stationary dist. x!
x = (1 )
1X
k=0
k
Pk
s1. (I AD 1
)x = (1 )s;
2. [↵D + L]z = ↵s where = 1/(1 + ↵) and x = Dz.
MMDS 2014
minimize kBxkC,1 =
P
ij2E Ci,j |xi xj |
subject to xs = 1, xt = 0, x 0.
The s-t min-cut problem
Unweighted incidence matrix
Diagonal capacity matrix
14
David Gleich · Purdue
t
s
In the unweighted case, "
solve via max-flow.

In the weighted case,
solve via network simplex
or industrial LP.
MMDS 2014
The localized cut graph



Related to a construction
used in “FlowImprove” "
Andersen & Lang (2007); and
Orecchia & Zhu (2014)
AS =
2
4
0 ↵dT
S 0
↵dS A ↵d¯S
0 ↵dT
¯S 0
3
5
Connect s to vertices
in S with weight ↵ · degree
Connect t to vertices
in ¯S with weight ↵ · degree
David Gleich · Purdue
15
MMDS 2014
The localized cut graph
Connect s to vertices
in S with weight ↵ · degree
Connect t to vertices
in ¯S with weight ↵ · degree
BS =
2
4
e IS 0
0 B 0
0 I¯S e
3
5
minimize kBSxkC(↵),1
subject to xs = 1, xt = 0
x 0.
Solve the s-t min-cut
David Gleich · Purdue
16
MMDS 2014
The localized cut graph
Connect s to vertices
in S with weight ↵ · degree
Connect t to vertices
in ¯S with weight ↵ · degree
BS =
2
4
e IS 0
0 B 0
0 I¯S e
3
5
Solve the “electrical flow” 

s-t min-cut
minimize kBSxkC(↵),2
subject to xs = 1, xt = 0
David Gleich · Purdue
17
MMDS 2014
s-t min-cut à PageRank 
 Proof
Square and expand
the objective into
a Laplacian, then
apply constraints.
David Gleich · Purdue
18
MMDS 2014
The PageRank vector z that solves
(↵D + L)z = ↵s
with s = dS/vol(S) is a renormalized
solution of the electrical cut computation:
minimize kBSxkC(↵),2
subject to xs = 1, xt = 0.
Specifically, if x is the solution, then
x =
2
4
1
vol(S)z
0
3
5
PageRank à s-t min-cut
That equivalence works if s is degree-weighted.
What if s is the uniform vector? 
A(s) =
2
4
0 ↵sT
0
↵s A ↵(d s)
0 ↵(d s)T
0
3
5 .
David Gleich · Purdue
19
MMDS 2014
Insight 1!
PageRank implicitly approximates the
solution of these s-t mincut problems
David Gleich · Purdue
20
MMDS 2014
The Push Algorithm for PageRank
Proposed (in closest form) in Andersen, Chung, Lang "
(also by McSherry, Jeh & Widom) for personalized PageRank
Strongly related to Gauss-Seidel on Ax=b (see my talk at Simons)
Derived to show improved runtime for balanced solvers
1. x(1)
= 0, r(1)
= (1 )ei , k = 1
2. while any rj > ⌧dj (dj is the degree of node j)
3. x(k+1)
= x(k)
+ (rj ⌧dj ⇢)ej
4. r(k+1)
i =
8
><
>:
⌧dj ⇢ i = j
r(k)
i + (rj ⌧dj ⇢)/dj i ⇠ j
r(k)
i otherwise
5. k k + 1
The
Push
Method!
⌧, ⇢
David Gleich · Purdue
21
a
b
c
MMDS 2014
Why do we care
about push?

1.  Used for empirical stud-
ies of “communities”
2.  Local Cheeger inequality.
3.  Used for “fast Page-
Rank approximation”
4.  It produces weakly
localized approximations
to PageRank!

Newman’s netscience!
379 vertices, 1828 nnz
“zero” on most of the nodes
s has a single "
one here
22
kD 1
(x x⇤
)k1  "
1
(1 )"
edges
The push method revisited
Let x be the output from the push method
with 0 < < 1, v = dS/vol(S),
⇢ = 1, and ⌧ > 0.
Set ↵ = 1
,  = ⌧vol(S)/ , and let zG solve:
minimize 1
2 kBSzk
2
C(↵),2 + kDzk1
subject to zs = 1, zt = 0, z 0
,
where z =
h 1
zG
0
i
.
Then x = DzG/vol(S).
Proof Write out KKT conditions
Show that the push method
solves them. Slackness was “tricky”
Regularization
for sparsity
David Gleich · Purdue
23
Need for
normalization
MMDS 2014
Insight 2!
The PageRank push method
implicitly solves a 1-norm regularized
2-norm cut approximation. 
David Gleich · Purdue
24
MMDS 2014
Insight 2’
We get 3-digits of accuracy on P and 
16-digits of accuracy on P’.
David Gleich · Purdue
25
MMDS 2014
David Gleich · Purdue
26
Anti-di↵erentiating Approximat
16 nonzeros 15 nonzeros
Figure 2. Examples of the di↵erent cut vectors on a portion of the netscience
with its vertices enlarged. In the other subfigures, we show the solution vectors
(4), and (6), solved with min-cut, PageRank, and ACL) for this set S . Each v
values are large and dark. White vertices with outlines are numerically non-zer
outlined, in contrast to the third figure). The true min-cut set is large in all ve
with many fewer non-zeros than the vanilla PageRank problem.
References
Andersen, Reid and Lang, Kevin. An algorithm for improving
graph partitions. In Proceedings of the 19th annual ACM-SIAM
Symposium on Discrete Algorithms, pp. 651–660, 2008.
Leskov
Mic
clus
Inte
Mahon
Anti-di↵erentiating Approximation Algorithms
eros 15 nonzeros 284 nonzeros 24 nonzeros
of the di↵erent cut vectors on a portion of the netscience graph. In the left subfigure, we show the set S highlighted
arged. In the other subfigures, we show the solution vectors from the various cut problems (from left to right, Probs. (2),
Push’s sparsity
helps it identify
the “right” graph
feature with fewer
non-zeros
The set S
 The mincut solution
The push solution
The PageRank solution
MMDS 2014
The push method revisited
Let x be the output from the push method
with 0 < < 1, v = dS/vol(S),
⇢ = 1, and ⌧ > 0.
Set ↵ = 1
,  = ⌧vol(S)/ , and let zG solve:
minimize 1
2 kBSzk
2
C(↵),2 + kDzk1
subject to zs = 1, zt = 0, z 0
,
where z =
h 1
zG
0
i
.
Then x = DzG/vol(S).
Regularization
for sparsity in
solution and
residual
David Gleich · Purdue
27
The push method is scalable because it gives
us sparse solutions AND sparse residuals r. 
MMDS 2014
This is a case of 
Algorithmic Anti-differentiation!
28
MMDS 2014
David Gleich · Purdue
Understand why H works!
Show heuristic H solves P’
Guess and check!
until you find something H
solves
Derive characterization of
heuristic H
The real world
Given “find-communities”
Hack around "

Write paper presenting
“three steps of the power
method on P finds
communities”
Algorithmic Anti-differentiation!
Given heuristic H, is there a problem P’
such that H is an algorithm for P’ ? 
MMDS 2014
David Gleich · Purdue
29
e.g. Mahoney & Orecchia,
Dhillon et al. (Graclus);
Saunders
Without these insights, we’d
draw the wrong conclusion.
David Gleich · Purdue
30
Gleich & Mahoney,
Submitted
Our s-t mincut framework extends to many
diffusions used in semi-supervised learning.
MMDS 2014
Without these insights, we’d
draw the wrong conclusion.
David Gleich · Purdue
31
Gleich & Mahoney,
Submitted
Our s-t mincut framework extends to many
diffusions used in semi-supervised learning.
2 4 6 8 10
0
0.2
0.4
0.6
0.8
errorrate
average training samples per class
K2
RK2
K3
RK3
Off the shelf SSL procedure
MMDS 2014
Without these insights, we’d
draw the wrong conclusion.
David Gleich · Purdue
32
Gleich & Mahoney,
Submitted
Our s-t mincut framework extends to many
diffusions used in semi-supervised learning.
2 4 6 8 10
0
0.2
0.4
0.6
0.8
errorrate
average training samples per class
K2
RK2
K3
RK3
2 4 6 8 10
0
0.2
0.4
0.6
0.8
errorrate
average training samples per class
K2
RK2
K3
RK3
Off the shelf SSL procedure
Rank-rounded SSL
MMDS 2014
Recap so far
1.  Used the relationship between
PageRank and mincut to get a new
understanding of the implicit properties
of the push method
2.  Showed that this insight helps improve
semi-supervised learning.
(next) A new algorithm for the heat kernel
diffusion in a degree weighted norm.

David Gleich · Purdue
33
MMDS 2014
Graph diffusions
David Gleich · Purdue
34
h = e t
1X
k=0
tk
k!
Pk
s
h = e t
exp{tP}s
PageRank
Heat kernel
0 20 40 60 80 100
10
−5
10
0
t=1 t=5 t=15 α=0.85
α=0.99
Weight
Length
x = (1 )
1X
k=0
k
Pk
s
(I P)x = (1 )s
Many “empirically useful” properties of PageRank
also hold for the Heat kernel diffusion, e.g. "
Chung (2007) showed a local Cheeger inequality.

No “local” algorithm until a randomized method by
Simpson & Chung (2013).
MMDS 2014
We can turn the heat kernel
into a linear system
Direct expansion!


"
!
!
!

David Gleich · Purdue
35
x = exp(P)ec ⇡
PN
k=0
1
k! Pk
ec = xN
2
6
6
6
6
6
6
4
III
P/1 III
P/2
...
... III
P/N III
3
7
7
7
7
7
7
5
2
6
6
6
6
6
6
4
v0
v1
...
...
vN
3
7
7
7
7
7
7
5
=
2
6
6
6
6
6
6
4
ec
0
...
...
0
3
7
7
7
7
7
7
5
xN =
NX
i=0
vi
(III ⌦ IIIN SN ⌦ P)v = e1 ⌦ ec
Lemma we approximate xN well if we approximate v well
Kloster & Gleich,
WAW2013
MMDS 2014
There is a fast deterministic
adaptation of the push method
David Gleich · Purdue
36
Kloster & Gleich,
KDD2014


ons
hat
hen,
erm
s is
d to
the
s:
(7)
(8)
tity
em
(9)
to
k ⇡
we
# G is graph as dictionary -of -sets ,
# seed is an array of seeds ,
# t, eps , N, psis are precomputed
x = {} # Store x, r as dictionaries
r = {} # initialize residual
Q = collections.deque () # initialize queue
for s in seed:
r[(s ,0)] = 1./ len(seed)
Q.append ((s ,0))
while len(Q) > 0:
(v,j) = Q.popleft () # v has r[(v,j)] ...
rvj = r[(v,j)]
# perform the hk -relax step
if v not in x: x[v] = 0.
x[v] += rvj
r[(v,j)] = 0.
mass = (t*rvj/( float(j)+1.))/ len(G[v])
for u in G[v]: # for neighbors of v
next = (u,j+1) # in the next block
if j+1 == N: # last step , add to soln
x[u] += rvj/len(G(v))
continue
if next not in r: r[next] = 0.
thresh = math.exp(t)*eps*len(G[u])
thresh = thresh /(N*psis[j+1])/2.
if r[next] < thresh and 
r[next] + mass >= thresh:
Q.append(next) # add u to queue
r[next] = r[next] + mass
Figure 2: Pseudo-code for our algorithm as work-
ing python code. The graph is stored as a dic-
Let h = e t
exp{tP}s.
Let x = hk-push(") output
Then kD 1
(x h)k1  "
after looking at 2Net
" edges.
We believe that the bound below suffices
N  2t log(1/")
MMDS 2014
PageRank vs. Heat Kernel
David Gleich · Purdue
37
5 6 7 8 9
0
0.5
1
1.5
2
Runtime: hk vs. ppr
log10(|V|+|E|)
Runtime(s)
hkgrow 50%
25%
75%
pprgrow 50%
25%
75%
5 6 7 8 9
10
−2
10
−1
10
0
Conductances: hk vs. ppr
log10(|V|+|E|)
log10(Conductances)
hkgrow 50%
25%
75%
pprgrow 50%
25%
75%
5 6 7 8 9
0
0.5
1
1.5
2
Runtime: hk vs. ppr
log10(|V|+|E|)
Runtime(s)
hkgrow 50%
25%
75%
pprgrow 50%
25%
75%
10
−2
10
−1
10
0
Conductances: hk vs. ppr
log10(Conductances)
hkgrow 50%
25%
75%
pprgrow 50%
25%
75%
On large graphs, our heat kernel
takes slightly longer than a
localized PageRank, but
produces sets with smaller
(better) conductance scores. 

Our python code on clueweb12
(72B edges) via libbvg:
•  99 seconds to load
•  1 second to compute
MMDS 2014
References and ongoing work
Gleich and Kloster – Relaxation methods for the matrix exponential,
Submitted"
Kloster and Gleich – Heat kernel based community detection KDD2014
Gleich and Mahoney – Algorithmic Anti-differentiation, ICML 2014 "
Gleich and Mahoney – Regularized diffusions, Submitted
www.cs.purdue.edu/homes/dgleich/codes/nexpokit!
www.cs.purdue.edu/homes/dgleich/codes/l1pagerank
•  Improved localization bounds for functions of matrices
•  Asynchronous and parallel “push”-style methods
David Gleich · Purdue
38
Supported by NSF CAREER 1149756-CCF 
 www.cs.purdue.edu/homes/dgleich

Weitere ähnliche Inhalte

Was ist angesagt?

Non-exhaustive, Overlapping K-means
Non-exhaustive, Overlapping K-meansNon-exhaustive, Overlapping K-means
Non-exhaustive, Overlapping K-meansDavid Gleich
 
Big data matrix factorizations and Overlapping community detection in graphs
Big data matrix factorizations and Overlapping community detection in graphsBig data matrix factorizations and Overlapping community detection in graphs
Big data matrix factorizations and Overlapping community detection in graphsDavid Gleich
 
Higher-order organization of complex networks
Higher-order organization of complex networksHigher-order organization of complex networks
Higher-order organization of complex networksDavid Gleich
 
Spectral clustering with motifs and higher-order structures
Spectral clustering with motifs and higher-order structuresSpectral clustering with motifs and higher-order structures
Spectral clustering with motifs and higher-order structuresDavid Gleich
 
Spacey random walks and higher-order data analysis
Spacey random walks and higher-order data analysisSpacey random walks and higher-order data analysis
Spacey random walks and higher-order data analysisDavid Gleich
 
Iterative methods with special structures
Iterative methods with special structuresIterative methods with special structures
Iterative methods with special structuresDavid Gleich
 
Correlation clustering and community detection in graphs and networks
Correlation clustering and community detection in graphs and networksCorrelation clustering and community detection in graphs and networks
Correlation clustering and community detection in graphs and networksDavid Gleich
 
Gaps between the theory and practice of large-scale matrix-based network comp...
Gaps between the theory and practice of large-scale matrix-based network comp...Gaps between the theory and practice of large-scale matrix-based network comp...
Gaps between the theory and practice of large-scale matrix-based network comp...David Gleich
 
Lesson 26: Integration by Substitution (handout)
Lesson 26: Integration by Substitution (handout)Lesson 26: Integration by Substitution (handout)
Lesson 26: Integration by Substitution (handout)Matthew Leingang
 
A new generalized lindley distribution
A new generalized lindley distributionA new generalized lindley distribution
A new generalized lindley distributionAlexander Decker
 
Relaxation methods for the matrix exponential on large networks
Relaxation methods for the matrix exponential on large networksRelaxation methods for the matrix exponential on large networks
Relaxation methods for the matrix exponential on large networksDavid Gleich
 
Uncertainty Modeling in Deep Learning
Uncertainty Modeling in Deep LearningUncertainty Modeling in Deep Learning
Uncertainty Modeling in Deep LearningSungjoon Choi
 
High-Performance Approach to String Similarity using Most Frequent K Characters
High-Performance Approach to String Similarity using Most Frequent K CharactersHigh-Performance Approach to String Similarity using Most Frequent K Characters
High-Performance Approach to String Similarity using Most Frequent K CharactersHolistic Benchmarking of Big Linked Data
 
A lattice-based consensus clustering
A lattice-based consensus clusteringA lattice-based consensus clustering
A lattice-based consensus clusteringDmitrii Ignatov
 
Different kind of distance and Statistical Distance
Different kind of distance and Statistical DistanceDifferent kind of distance and Statistical Distance
Different kind of distance and Statistical DistanceKhulna University
 
Pattern-based classification of demographic sequences
Pattern-based classification of demographic sequencesPattern-based classification of demographic sequences
Pattern-based classification of demographic sequencesDmitrii Ignatov
 
An application of gd
An application of gdAn application of gd
An application of gdgraphhoc
 
Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...
Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...
Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...Atsushi Nitanda
 

Was ist angesagt? (20)

Non-exhaustive, Overlapping K-means
Non-exhaustive, Overlapping K-meansNon-exhaustive, Overlapping K-means
Non-exhaustive, Overlapping K-means
 
Big data matrix factorizations and Overlapping community detection in graphs
Big data matrix factorizations and Overlapping community detection in graphsBig data matrix factorizations and Overlapping community detection in graphs
Big data matrix factorizations and Overlapping community detection in graphs
 
Higher-order organization of complex networks
Higher-order organization of complex networksHigher-order organization of complex networks
Higher-order organization of complex networks
 
Spectral clustering with motifs and higher-order structures
Spectral clustering with motifs and higher-order structuresSpectral clustering with motifs and higher-order structures
Spectral clustering with motifs and higher-order structures
 
Spacey random walks and higher-order data analysis
Spacey random walks and higher-order data analysisSpacey random walks and higher-order data analysis
Spacey random walks and higher-order data analysis
 
Iterative methods with special structures
Iterative methods with special structuresIterative methods with special structures
Iterative methods with special structures
 
Correlation clustering and community detection in graphs and networks
Correlation clustering and community detection in graphs and networksCorrelation clustering and community detection in graphs and networks
Correlation clustering and community detection in graphs and networks
 
Gaps between the theory and practice of large-scale matrix-based network comp...
Gaps between the theory and practice of large-scale matrix-based network comp...Gaps between the theory and practice of large-scale matrix-based network comp...
Gaps between the theory and practice of large-scale matrix-based network comp...
 
Lesson 26: Integration by Substitution (handout)
Lesson 26: Integration by Substitution (handout)Lesson 26: Integration by Substitution (handout)
Lesson 26: Integration by Substitution (handout)
 
A new generalized lindley distribution
A new generalized lindley distributionA new generalized lindley distribution
A new generalized lindley distribution
 
Relaxation methods for the matrix exponential on large networks
Relaxation methods for the matrix exponential on large networksRelaxation methods for the matrix exponential on large networks
Relaxation methods for the matrix exponential on large networks
 
Uncertainty Modeling in Deep Learning
Uncertainty Modeling in Deep LearningUncertainty Modeling in Deep Learning
Uncertainty Modeling in Deep Learning
 
High-Performance Approach to String Similarity using Most Frequent K Characters
High-Performance Approach to String Similarity using Most Frequent K CharactersHigh-Performance Approach to String Similarity using Most Frequent K Characters
High-Performance Approach to String Similarity using Most Frequent K Characters
 
Uncertainty in Deep Learning
Uncertainty in Deep LearningUncertainty in Deep Learning
Uncertainty in Deep Learning
 
A lattice-based consensus clustering
A lattice-based consensus clusteringA lattice-based consensus clustering
A lattice-based consensus clustering
 
Different kind of distance and Statistical Distance
Different kind of distance and Statistical DistanceDifferent kind of distance and Statistical Distance
Different kind of distance and Statistical Distance
 
Pattern-based classification of demographic sequences
Pattern-based classification of demographic sequencesPattern-based classification of demographic sequences
Pattern-based classification of demographic sequences
 
MUMS Opening Workshop - Extrapolation: The Art of Connecting Model-Based Pred...
MUMS Opening Workshop - Extrapolation: The Art of Connecting Model-Based Pred...MUMS Opening Workshop - Extrapolation: The Art of Connecting Model-Based Pred...
MUMS Opening Workshop - Extrapolation: The Art of Connecting Model-Based Pred...
 
An application of gd
An application of gdAn application of gd
An application of gd
 
Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...
Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...
Stochastic Gradient Descent with Exponential Convergence Rates of Expected Cl...
 

Andere mochten auch

13 임태훈
13 임태훈13 임태훈
13 임태훈KOS-ROBOT
 
[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream Job
[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream Job[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream Job
[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream JobChallenge:Future
 
The Future of Smart Disclosure (pdf)
The Future of Smart Disclosure (pdf)The Future of Smart Disclosure (pdf)
The Future of Smart Disclosure (pdf)Tim O'Reilly
 
DreamCode Season2 Intro
DreamCode Season2 IntroDreamCode Season2 Intro
DreamCode Season2 IntroUi-Gyun Jeong
 
The Future of Smart Disclosure
The Future of Smart DisclosureThe Future of Smart Disclosure
The Future of Smart DisclosureTim O'Reilly
 
Direct tall-and-skinny QR factorizations in MapReduce architectures
Direct tall-and-skinny QR factorizations in MapReduce architecturesDirect tall-and-skinny QR factorizations in MapReduce architectures
Direct tall-and-skinny QR factorizations in MapReduce architecturesDavid Gleich
 
MapReduce Tall-and-skinny QR and applications
MapReduce Tall-and-skinny QR and applicationsMapReduce Tall-and-skinny QR and applications
MapReduce Tall-and-skinny QR and applicationsDavid Gleich
 
Tall and Skinny QRs in MapReduce
Tall and Skinny QRs in MapReduceTall and Skinny QRs in MapReduce
Tall and Skinny QRs in MapReduceDavid Gleich
 
Iterative methods for network alignment
Iterative methods for network alignmentIterative methods for network alignment
Iterative methods for network alignmentDavid Gleich
 
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...David Gleich
 
The power and Arnoldi methods in an algebra of circulants
The power and Arnoldi methods in an algebra of circulantsThe power and Arnoldi methods in an algebra of circulants
The power and Arnoldi methods in an algebra of circulantsDavid Gleich
 
A multithreaded method for network alignment
A multithreaded method for network alignmentA multithreaded method for network alignment
A multithreaded method for network alignmentDavid Gleich
 
A history of PageRank from the numerical computing perspective
A history of PageRank from the numerical computing perspectiveA history of PageRank from the numerical computing perspective
A history of PageRank from the numerical computing perspectiveDavid Gleich
 
How does Google Google: A journey into the wondrous mathematics behind your f...
How does Google Google: A journey into the wondrous mathematics behind your f...How does Google Google: A journey into the wondrous mathematics behind your f...
How does Google Google: A journey into the wondrous mathematics behind your f...David Gleich
 
Tall-and-skinny QR factorizations in MapReduce architectures
Tall-and-skinny QR factorizations in MapReduce architecturesTall-and-skinny QR factorizations in MapReduce architectures
Tall-and-skinny QR factorizations in MapReduce architecturesDavid Gleich
 
A dynamical system for PageRank with time-dependent teleportation
A dynamical system for PageRank with time-dependent teleportationA dynamical system for PageRank with time-dependent teleportation
A dynamical system for PageRank with time-dependent teleportationDavid Gleich
 
Fast relaxation methods for the matrix exponential
Fast relaxation methods for the matrix exponential Fast relaxation methods for the matrix exponential
Fast relaxation methods for the matrix exponential David Gleich
 
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...Vertex neighborhoods, low conductance cuts, and good seeds for local communit...
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...David Gleich
 
MapReduce for scientific simulation analysis
MapReduce for scientific simulation analysisMapReduce for scientific simulation analysis
MapReduce for scientific simulation analysisDavid Gleich
 
Recommendation and graph algorithms in Hadoop and SQL
Recommendation and graph algorithms in Hadoop and SQLRecommendation and graph algorithms in Hadoop and SQL
Recommendation and graph algorithms in Hadoop and SQLDavid Gleich
 

Andere mochten auch (20)

13 임태훈
13 임태훈13 임태훈
13 임태훈
 
[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream Job
[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream Job[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream Job
[Challenge:Future] Inter-School Bridge System (I.S.B.S): My Dream Job
 
The Future of Smart Disclosure (pdf)
The Future of Smart Disclosure (pdf)The Future of Smart Disclosure (pdf)
The Future of Smart Disclosure (pdf)
 
DreamCode Season2 Intro
DreamCode Season2 IntroDreamCode Season2 Intro
DreamCode Season2 Intro
 
The Future of Smart Disclosure
The Future of Smart DisclosureThe Future of Smart Disclosure
The Future of Smart Disclosure
 
Direct tall-and-skinny QR factorizations in MapReduce architectures
Direct tall-and-skinny QR factorizations in MapReduce architecturesDirect tall-and-skinny QR factorizations in MapReduce architectures
Direct tall-and-skinny QR factorizations in MapReduce architectures
 
MapReduce Tall-and-skinny QR and applications
MapReduce Tall-and-skinny QR and applicationsMapReduce Tall-and-skinny QR and applications
MapReduce Tall-and-skinny QR and applications
 
Tall and Skinny QRs in MapReduce
Tall and Skinny QRs in MapReduceTall and Skinny QRs in MapReduce
Tall and Skinny QRs in MapReduce
 
Iterative methods for network alignment
Iterative methods for network alignmentIterative methods for network alignment
Iterative methods for network alignment
 
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...
 
The power and Arnoldi methods in an algebra of circulants
The power and Arnoldi methods in an algebra of circulantsThe power and Arnoldi methods in an algebra of circulants
The power and Arnoldi methods in an algebra of circulants
 
A multithreaded method for network alignment
A multithreaded method for network alignmentA multithreaded method for network alignment
A multithreaded method for network alignment
 
A history of PageRank from the numerical computing perspective
A history of PageRank from the numerical computing perspectiveA history of PageRank from the numerical computing perspective
A history of PageRank from the numerical computing perspective
 
How does Google Google: A journey into the wondrous mathematics behind your f...
How does Google Google: A journey into the wondrous mathematics behind your f...How does Google Google: A journey into the wondrous mathematics behind your f...
How does Google Google: A journey into the wondrous mathematics behind your f...
 
Tall-and-skinny QR factorizations in MapReduce architectures
Tall-and-skinny QR factorizations in MapReduce architecturesTall-and-skinny QR factorizations in MapReduce architectures
Tall-and-skinny QR factorizations in MapReduce architectures
 
A dynamical system for PageRank with time-dependent teleportation
A dynamical system for PageRank with time-dependent teleportationA dynamical system for PageRank with time-dependent teleportation
A dynamical system for PageRank with time-dependent teleportation
 
Fast relaxation methods for the matrix exponential
Fast relaxation methods for the matrix exponential Fast relaxation methods for the matrix exponential
Fast relaxation methods for the matrix exponential
 
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...Vertex neighborhoods, low conductance cuts, and good seeds for local communit...
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...
 
MapReduce for scientific simulation analysis
MapReduce for scientific simulation analysisMapReduce for scientific simulation analysis
MapReduce for scientific simulation analysis
 
Recommendation and graph algorithms in Hadoop and SQL
Recommendation and graph algorithms in Hadoop and SQLRecommendation and graph algorithms in Hadoop and SQL
Recommendation and graph algorithms in Hadoop and SQL
 

Ähnlich wie Localized methods for diffusions in large graphs

Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsChristian Robert
 
slides_nuclear_norm_regularization_david_mateos
slides_nuclear_norm_regularization_david_mateosslides_nuclear_norm_regularization_david_mateos
slides_nuclear_norm_regularization_david_mateosDavid Mateos
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixturesChristian Robert
 
Lego like spheres and tori, enumeration and drawings
Lego like spheres and tori, enumeration and drawingsLego like spheres and tori, enumeration and drawings
Lego like spheres and tori, enumeration and drawingsMathieu Dutour Sikiric
 
Cluster abundances and clustering Can theory step up to precision cosmology?
Cluster abundances and clustering Can theory step up to precision cosmology?Cluster abundances and clustering Can theory step up to precision cosmology?
Cluster abundances and clustering Can theory step up to precision cosmology?CosmoAIMS Bassett
 
Lecture on linerar discriminatory analysis
Lecture on linerar discriminatory analysisLecture on linerar discriminatory analysis
Lecture on linerar discriminatory analysisdevcb13d
 
A non-stiff numerical method for 3D interfacial flow of inviscid fluids.
A non-stiff numerical method for 3D interfacial flow of inviscid fluids.A non-stiff numerical method for 3D interfacial flow of inviscid fluids.
A non-stiff numerical method for 3D interfacial flow of inviscid fluids.Alex (Oleksiy) Varfolomiyev
 
Graph Edit Distance: Basics & Trends
Graph Edit Distance: Basics & TrendsGraph Edit Distance: Basics & Trends
Graph Edit Distance: Basics & TrendsLuc Brun
 
Developing fast low-rank tensor methods for solving PDEs with uncertain coef...
Developing fast  low-rank tensor methods for solving PDEs with uncertain coef...Developing fast  low-rank tensor methods for solving PDEs with uncertain coef...
Developing fast low-rank tensor methods for solving PDEs with uncertain coef...Alexander Litvinenko
 
Challenges for coupling approaches for classical linear elasticity and bond-b...
Challenges for coupling approaches for classical linear elasticity and bond-b...Challenges for coupling approaches for classical linear elasticity and bond-b...
Challenges for coupling approaches for classical linear elasticity and bond-b...Patrick Diehl
 
Sketching and locality sensitive hashing for alignment
Sketching and locality sensitive hashing for alignmentSketching and locality sensitive hashing for alignment
Sketching and locality sensitive hashing for alignmentssuser2be88c
 
Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...Vladimir Bakhrushin
 
Theories and Engineering Technics of 2D-to-3D Back-Projection Problem
Theories and Engineering Technics of 2D-to-3D Back-Projection ProblemTheories and Engineering Technics of 2D-to-3D Back-Projection Problem
Theories and Engineering Technics of 2D-to-3D Back-Projection ProblemSeongcheol Baek
 

Ähnlich wie Localized methods for diffusions in large graphs (20)

Interactive High-Dimensional Visualization of Social Graphs
Interactive High-Dimensional Visualization of Social GraphsInteractive High-Dimensional Visualization of Social Graphs
Interactive High-Dimensional Visualization of Social Graphs
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithms
 
slides_nuclear_norm_regularization_david_mateos
slides_nuclear_norm_regularization_david_mateosslides_nuclear_norm_regularization_david_mateos
slides_nuclear_norm_regularization_david_mateos
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixtures
 
CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)
 
CLIM Program: Remote Sensing Workshop, Optimization for Distributed Data Syst...
CLIM Program: Remote Sensing Workshop, Optimization for Distributed Data Syst...CLIM Program: Remote Sensing Workshop, Optimization for Distributed Data Syst...
CLIM Program: Remote Sensing Workshop, Optimization for Distributed Data Syst...
 
Lego like spheres and tori, enumeration and drawings
Lego like spheres and tori, enumeration and drawingsLego like spheres and tori, enumeration and drawings
Lego like spheres and tori, enumeration and drawings
 
Cluster abundances and clustering Can theory step up to precision cosmology?
Cluster abundances and clustering Can theory step up to precision cosmology?Cluster abundances and clustering Can theory step up to precision cosmology?
Cluster abundances and clustering Can theory step up to precision cosmology?
 
Lecture on linerar discriminatory analysis
Lecture on linerar discriminatory analysisLecture on linerar discriminatory analysis
Lecture on linerar discriminatory analysis
 
A non-stiff numerical method for 3D interfacial flow of inviscid fluids.
A non-stiff numerical method for 3D interfacial flow of inviscid fluids.A non-stiff numerical method for 3D interfacial flow of inviscid fluids.
A non-stiff numerical method for 3D interfacial flow of inviscid fluids.
 
Graph Edit Distance: Basics & Trends
Graph Edit Distance: Basics & TrendsGraph Edit Distance: Basics & Trends
Graph Edit Distance: Basics & Trends
 
Bayes gauss
Bayes gaussBayes gauss
Bayes gauss
 
Developing fast low-rank tensor methods for solving PDEs with uncertain coef...
Developing fast  low-rank tensor methods for solving PDEs with uncertain coef...Developing fast  low-rank tensor methods for solving PDEs with uncertain coef...
Developing fast low-rank tensor methods for solving PDEs with uncertain coef...
 
Presentation.pdf
Presentation.pdfPresentation.pdf
Presentation.pdf
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Challenges for coupling approaches for classical linear elasticity and bond-b...
Challenges for coupling approaches for classical linear elasticity and bond-b...Challenges for coupling approaches for classical linear elasticity and bond-b...
Challenges for coupling approaches for classical linear elasticity and bond-b...
 
Sketching and locality sensitive hashing for alignment
Sketching and locality sensitive hashing for alignmentSketching and locality sensitive hashing for alignment
Sketching and locality sensitive hashing for alignment
 
CLIM Fall 2017 Course: Statistics for Climate Research, Spatial Data: Models ...
CLIM Fall 2017 Course: Statistics for Climate Research, Spatial Data: Models ...CLIM Fall 2017 Course: Statistics for Climate Research, Spatial Data: Models ...
CLIM Fall 2017 Course: Statistics for Climate Research, Spatial Data: Models ...
 
Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...
 
Theories and Engineering Technics of 2D-to-3D Back-Projection Problem
Theories and Engineering Technics of 2D-to-3D Back-Projection ProblemTheories and Engineering Technics of 2D-to-3D Back-Projection Problem
Theories and Engineering Technics of 2D-to-3D Back-Projection Problem
 

Kürzlich hochgeladen

How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
Potential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and InsightsPotential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and InsightsRavi Sanghani
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Hiroshi SHIBATA
 
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...itnewsafrica
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesBernd Ruecker
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integrationmarketing932765
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersNicole Novielli
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...Farhan Tariq
 
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security ObservabilityGlenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security Observabilityitnewsafrica
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesThousandEyes
 

Kürzlich hochgeladen (20)

How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
Potential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and InsightsPotential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and Insights
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024
 
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architectures
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software Developers
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...
 
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security ObservabilityGlenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
 

Localized methods for diffusions in large graphs

  • 1. Localized methods for diffusions in large graphs David F. Gleich! Purdue University! Joint work with Kyle Kloster @" Purdue & Michael Mahoney @ Berkeley supported by " NSF CAREER CCF-1149756 Code "www.cs.purdue.edu/homes/dgleich/codes/nexpokit ! "www.cs.purdue.edu/homes/dgleich/codes/l1pagerank! David Gleich · Purdue 1 MMDS 2014
  • 2. Image from rockysprings, deviantart, CC share-alike Everything in the world can be explained by a matrix, and we see how deep the rabbit hole goes The talk ends, you believe -- whatever you want to.
  • 3. 3
  • 4. Graph diffusions David Gleich · Purdue 4 f = 1X k=0 ↵k Pk s ate t in on work, or mesh, from a typical problem in scientific computing high low A – adjacency matrix! D – degree matrix! P – column stochastic operator s – the “seed” (a sparse vector) f – the diffusion result 𝛼k – the path weights P = AD 1 Px = X j!i 1 dj xj Graph diffusions help: 1.  Attribute prediction 2.  Community detection 3.  “Ranking” 4.  Find small conductance sets MMDS 2014
  • 5. Graph diffusions David Gleich · Purdue 5 ate t in on work, or mesh, from a typical problem in scientific computing high low h = e t 1X k=0 tk k! Pk s h = e t exp{tP}s PageRank Heat kernel x = (1 ) 1X k=0 k Pk s (I P)x = (1 )s P = AD 1 Px = X j!i 1 dj xj MMDS 2014
  • 6. Graph diffusions David Gleich · Purdue 6 h = e t 1X k=0 tk k! Pk s h = e t exp{tP}s PageRank Heat kernel 0 20 40 60 80 100 10 −5 10 0 t=1 t=5 t=15 α=0.85 α=0.99 Weight Length x = (1 ) 1X k=0 k Pk s (I P)x = (1 )s MMDS 2014
  • 7. Uniformly localized " solutions in livejournal 1 2 3 4 5 x 10 6 0 0.5 1 1.5 nnz = 4815948 magnitude plot(x) 10 0 10 1 10 2 10 3 10 4 10 5 10 6 10 −14 10 −12 10 −10 10 −8 10 −6 10 −4 10 −2 10 0 1−normerror largest non−zeros retained 10 0 10 1 10 2 10 3 10 4 10 5 10 6 10 −14 10 −12 10 −10 10 −8 10 −6 10 −4 10 −2 10 0 1−normerror largest non−zeros retained x = exp(P)ec David Gleich · Purdue 7 nnz(x) = 4, 815, 948 Gleich & Kloster, arXiv:1310.3423 MMDS 2014
  • 8. Our mission! Find the solution with work " roughly proportional to the " localization, not the matrix. David Gleich · Purdue 8 MMDS 2014
  • 9. Two types of localization David Gleich · Purdue 9 kx x⇤ k1  " kD 1 (x x⇤ )k1  " x ⇡ x⇤ Uniform (Strong)! Entry-wise (Weak)! Localized vectors are not sparse, but they can be approximated by sparse vectors. Good global approximation using only a local region. “Hard” to prove. “Need” a graph property. Good approximation for cuts and communities. “Easy” to prove. “Fast” algorithms MMDS 2014
  • 10. We have four results 1.  A new interpretation for the PageRank diffusion in relationship with a mincut problem. 2.  A new understanding of the scalable, localized PageRank “push” method 3.  A new algorithm for the heat kernel diffusion in a degree weighted norm. 4.  Algorithms for diffusions as functions of matrices (K. Kloster’s poster on Thurs.) David Gleich · Purdue 10 Undirected graphs only Entry-wise localization Directed, uniform localization MMDS 2014
  • 11. Our algorithms for uniform localization" www.cs.purdue.edu/homes/dgleich/codes/nexpokit 10 0 10 1 10 2 10 3 10 4 10 5 10 6 10 −8 10 −6 10 −4 10 −2 10 0 non−zeros 1−normerror gexpm gexpmq expmimv 10 0 10 1 10 2 10 3 10 4 10 5 10 6 10 −8 10 −6 10 −4 10 −2 10 0 non−zeros 1−normerror David Gleich · Purdue 11 MMDS 2014 work = O ⇣ log(1 " )(1 " )3/2 d2 (log d)2 ⌘ nnz = O ⇣ log(1 " )(1 " )3/2 d(log d) ⌘
  • 12. PageRank, mincuts, and the push method via Algorithmic Anti-Differentiation David Gleich · Purdue 12 Gleich & Mahoney, ICML 2014 MMDS 2014
  • 13. The PageRank problem & " the Laplacian on undirected graphs Combinatorial Laplacian L = D - A! David Gleich · Purdue 13 The PageRank random surfer 1.  With probability beta, follow a random-walk step 2.  With probability (1-beta), jump randomly ~ dist. s. Goal find the stationary dist. x! x = (1 ) 1X k=0 k Pk s1. (I AD 1 )x = (1 )s; 2. [↵D + L]z = ↵s where = 1/(1 + ↵) and x = Dz. MMDS 2014
  • 14. minimize kBxkC,1 = P ij2E Ci,j |xi xj | subject to xs = 1, xt = 0, x 0. The s-t min-cut problem Unweighted incidence matrix Diagonal capacity matrix 14 David Gleich · Purdue t s In the unweighted case, " solve via max-flow. In the weighted case, solve via network simplex or industrial LP. MMDS 2014
  • 15. The localized cut graph Related to a construction used in “FlowImprove” " Andersen & Lang (2007); and Orecchia & Zhu (2014) AS = 2 4 0 ↵dT S 0 ↵dS A ↵d¯S 0 ↵dT ¯S 0 3 5 Connect s to vertices in S with weight ↵ · degree Connect t to vertices in ¯S with weight ↵ · degree David Gleich · Purdue 15 MMDS 2014
  • 16. The localized cut graph Connect s to vertices in S with weight ↵ · degree Connect t to vertices in ¯S with weight ↵ · degree BS = 2 4 e IS 0 0 B 0 0 I¯S e 3 5 minimize kBSxkC(↵),1 subject to xs = 1, xt = 0 x 0. Solve the s-t min-cut David Gleich · Purdue 16 MMDS 2014
  • 17. The localized cut graph Connect s to vertices in S with weight ↵ · degree Connect t to vertices in ¯S with weight ↵ · degree BS = 2 4 e IS 0 0 B 0 0 I¯S e 3 5 Solve the “electrical flow” 
 s-t min-cut minimize kBSxkC(↵),2 subject to xs = 1, xt = 0 David Gleich · Purdue 17 MMDS 2014
  • 18. s-t min-cut à PageRank Proof Square and expand the objective into a Laplacian, then apply constraints. David Gleich · Purdue 18 MMDS 2014 The PageRank vector z that solves (↵D + L)z = ↵s with s = dS/vol(S) is a renormalized solution of the electrical cut computation: minimize kBSxkC(↵),2 subject to xs = 1, xt = 0. Specifically, if x is the solution, then x = 2 4 1 vol(S)z 0 3 5
  • 19. PageRank à s-t min-cut That equivalence works if s is degree-weighted. What if s is the uniform vector? A(s) = 2 4 0 ↵sT 0 ↵s A ↵(d s) 0 ↵(d s)T 0 3 5 . David Gleich · Purdue 19 MMDS 2014
  • 20. Insight 1! PageRank implicitly approximates the solution of these s-t mincut problems David Gleich · Purdue 20 MMDS 2014
  • 21. The Push Algorithm for PageRank Proposed (in closest form) in Andersen, Chung, Lang " (also by McSherry, Jeh & Widom) for personalized PageRank Strongly related to Gauss-Seidel on Ax=b (see my talk at Simons) Derived to show improved runtime for balanced solvers 1. x(1) = 0, r(1) = (1 )ei , k = 1 2. while any rj > ⌧dj (dj is the degree of node j) 3. x(k+1) = x(k) + (rj ⌧dj ⇢)ej 4. r(k+1) i = 8 >< >: ⌧dj ⇢ i = j r(k) i + (rj ⌧dj ⇢)/dj i ⇠ j r(k) i otherwise 5. k k + 1 The Push Method! ⌧, ⇢ David Gleich · Purdue 21 a b c MMDS 2014
  • 22. Why do we care about push? 1.  Used for empirical stud- ies of “communities” 2.  Local Cheeger inequality. 3.  Used for “fast Page- Rank approximation” 4.  It produces weakly localized approximations to PageRank! Newman’s netscience! 379 vertices, 1828 nnz “zero” on most of the nodes s has a single " one here 22 kD 1 (x x⇤ )k1  " 1 (1 )" edges
  • 23. The push method revisited Let x be the output from the push method with 0 < < 1, v = dS/vol(S), ⇢ = 1, and ⌧ > 0. Set ↵ = 1 ,  = ⌧vol(S)/ , and let zG solve: minimize 1 2 kBSzk 2 C(↵),2 + kDzk1 subject to zs = 1, zt = 0, z 0 , where z = h 1 zG 0 i . Then x = DzG/vol(S). Proof Write out KKT conditions Show that the push method solves them. Slackness was “tricky” Regularization for sparsity David Gleich · Purdue 23 Need for normalization MMDS 2014
  • 24. Insight 2! The PageRank push method implicitly solves a 1-norm regularized 2-norm cut approximation. David Gleich · Purdue 24 MMDS 2014
  • 25. Insight 2’ We get 3-digits of accuracy on P and 16-digits of accuracy on P’. David Gleich · Purdue 25 MMDS 2014
  • 26. David Gleich · Purdue 26 Anti-di↵erentiating Approximat 16 nonzeros 15 nonzeros Figure 2. Examples of the di↵erent cut vectors on a portion of the netscience with its vertices enlarged. In the other subfigures, we show the solution vectors (4), and (6), solved with min-cut, PageRank, and ACL) for this set S . Each v values are large and dark. White vertices with outlines are numerically non-zer outlined, in contrast to the third figure). The true min-cut set is large in all ve with many fewer non-zeros than the vanilla PageRank problem. References Andersen, Reid and Lang, Kevin. An algorithm for improving graph partitions. In Proceedings of the 19th annual ACM-SIAM Symposium on Discrete Algorithms, pp. 651–660, 2008. Leskov Mic clus Inte Mahon Anti-di↵erentiating Approximation Algorithms eros 15 nonzeros 284 nonzeros 24 nonzeros of the di↵erent cut vectors on a portion of the netscience graph. In the left subfigure, we show the set S highlighted arged. In the other subfigures, we show the solution vectors from the various cut problems (from left to right, Probs. (2), Push’s sparsity helps it identify the “right” graph feature with fewer non-zeros The set S The mincut solution The push solution The PageRank solution MMDS 2014
  • 27. The push method revisited Let x be the output from the push method with 0 < < 1, v = dS/vol(S), ⇢ = 1, and ⌧ > 0. Set ↵ = 1 ,  = ⌧vol(S)/ , and let zG solve: minimize 1 2 kBSzk 2 C(↵),2 + kDzk1 subject to zs = 1, zt = 0, z 0 , where z = h 1 zG 0 i . Then x = DzG/vol(S). Regularization for sparsity in solution and residual David Gleich · Purdue 27 The push method is scalable because it gives us sparse solutions AND sparse residuals r. MMDS 2014
  • 28. This is a case of Algorithmic Anti-differentiation! 28 MMDS 2014 David Gleich · Purdue
  • 29. Understand why H works! Show heuristic H solves P’ Guess and check! until you find something H solves Derive characterization of heuristic H The real world Given “find-communities” Hack around " Write paper presenting “three steps of the power method on P finds communities” Algorithmic Anti-differentiation! Given heuristic H, is there a problem P’ such that H is an algorithm for P’ ? MMDS 2014 David Gleich · Purdue 29 e.g. Mahoney & Orecchia, Dhillon et al. (Graclus); Saunders
  • 30. Without these insights, we’d draw the wrong conclusion. David Gleich · Purdue 30 Gleich & Mahoney, Submitted Our s-t mincut framework extends to many diffusions used in semi-supervised learning. MMDS 2014
  • 31. Without these insights, we’d draw the wrong conclusion. David Gleich · Purdue 31 Gleich & Mahoney, Submitted Our s-t mincut framework extends to many diffusions used in semi-supervised learning. 2 4 6 8 10 0 0.2 0.4 0.6 0.8 errorrate average training samples per class K2 RK2 K3 RK3 Off the shelf SSL procedure MMDS 2014
  • 32. Without these insights, we’d draw the wrong conclusion. David Gleich · Purdue 32 Gleich & Mahoney, Submitted Our s-t mincut framework extends to many diffusions used in semi-supervised learning. 2 4 6 8 10 0 0.2 0.4 0.6 0.8 errorrate average training samples per class K2 RK2 K3 RK3 2 4 6 8 10 0 0.2 0.4 0.6 0.8 errorrate average training samples per class K2 RK2 K3 RK3 Off the shelf SSL procedure Rank-rounded SSL MMDS 2014
  • 33. Recap so far 1.  Used the relationship between PageRank and mincut to get a new understanding of the implicit properties of the push method 2.  Showed that this insight helps improve semi-supervised learning. (next) A new algorithm for the heat kernel diffusion in a degree weighted norm. David Gleich · Purdue 33 MMDS 2014
  • 34. Graph diffusions David Gleich · Purdue 34 h = e t 1X k=0 tk k! Pk s h = e t exp{tP}s PageRank Heat kernel 0 20 40 60 80 100 10 −5 10 0 t=1 t=5 t=15 α=0.85 α=0.99 Weight Length x = (1 ) 1X k=0 k Pk s (I P)x = (1 )s Many “empirically useful” properties of PageRank also hold for the Heat kernel diffusion, e.g. " Chung (2007) showed a local Cheeger inequality. No “local” algorithm until a randomized method by Simpson & Chung (2013). MMDS 2014
  • 35. We can turn the heat kernel into a linear system Direct expansion! " ! ! ! David Gleich · Purdue 35 x = exp(P)ec ⇡ PN k=0 1 k! Pk ec = xN 2 6 6 6 6 6 6 4 III P/1 III P/2 ... ... III P/N III 3 7 7 7 7 7 7 5 2 6 6 6 6 6 6 4 v0 v1 ... ... vN 3 7 7 7 7 7 7 5 = 2 6 6 6 6 6 6 4 ec 0 ... ... 0 3 7 7 7 7 7 7 5 xN = NX i=0 vi (III ⌦ IIIN SN ⌦ P)v = e1 ⌦ ec Lemma we approximate xN well if we approximate v well Kloster & Gleich, WAW2013 MMDS 2014
  • 36. There is a fast deterministic adaptation of the push method David Gleich · Purdue 36 Kloster & Gleich, KDD2014 ons hat hen, erm s is d to the s: (7) (8) tity em (9) to k ⇡ we # G is graph as dictionary -of -sets , # seed is an array of seeds , # t, eps , N, psis are precomputed x = {} # Store x, r as dictionaries r = {} # initialize residual Q = collections.deque () # initialize queue for s in seed: r[(s ,0)] = 1./ len(seed) Q.append ((s ,0)) while len(Q) > 0: (v,j) = Q.popleft () # v has r[(v,j)] ... rvj = r[(v,j)] # perform the hk -relax step if v not in x: x[v] = 0. x[v] += rvj r[(v,j)] = 0. mass = (t*rvj/( float(j)+1.))/ len(G[v]) for u in G[v]: # for neighbors of v next = (u,j+1) # in the next block if j+1 == N: # last step , add to soln x[u] += rvj/len(G(v)) continue if next not in r: r[next] = 0. thresh = math.exp(t)*eps*len(G[u]) thresh = thresh /(N*psis[j+1])/2. if r[next] < thresh and r[next] + mass >= thresh: Q.append(next) # add u to queue r[next] = r[next] + mass Figure 2: Pseudo-code for our algorithm as work- ing python code. The graph is stored as a dic- Let h = e t exp{tP}s. Let x = hk-push(") output Then kD 1 (x h)k1  " after looking at 2Net " edges. We believe that the bound below suffices N  2t log(1/") MMDS 2014
  • 37. PageRank vs. Heat Kernel David Gleich · Purdue 37 5 6 7 8 9 0 0.5 1 1.5 2 Runtime: hk vs. ppr log10(|V|+|E|) Runtime(s) hkgrow 50% 25% 75% pprgrow 50% 25% 75% 5 6 7 8 9 10 −2 10 −1 10 0 Conductances: hk vs. ppr log10(|V|+|E|) log10(Conductances) hkgrow 50% 25% 75% pprgrow 50% 25% 75% 5 6 7 8 9 0 0.5 1 1.5 2 Runtime: hk vs. ppr log10(|V|+|E|) Runtime(s) hkgrow 50% 25% 75% pprgrow 50% 25% 75% 10 −2 10 −1 10 0 Conductances: hk vs. ppr log10(Conductances) hkgrow 50% 25% 75% pprgrow 50% 25% 75% On large graphs, our heat kernel takes slightly longer than a localized PageRank, but produces sets with smaller (better) conductance scores. Our python code on clueweb12 (72B edges) via libbvg: •  99 seconds to load •  1 second to compute MMDS 2014
  • 38. References and ongoing work Gleich and Kloster – Relaxation methods for the matrix exponential, Submitted" Kloster and Gleich – Heat kernel based community detection KDD2014 Gleich and Mahoney – Algorithmic Anti-differentiation, ICML 2014 " Gleich and Mahoney – Regularized diffusions, Submitted www.cs.purdue.edu/homes/dgleich/codes/nexpokit! www.cs.purdue.edu/homes/dgleich/codes/l1pagerank •  Improved localization bounds for functions of matrices •  Asynchronous and parallel “push”-style methods David Gleich · Purdue 38 Supported by NSF CAREER 1149756-CCF www.cs.purdue.edu/homes/dgleich