-
Notifications
You must be signed in to change notification settings - Fork 17
Expand file tree
/
Copy pathlecture3.tex
More file actions
311 lines (254 loc) · 13.8 KB
/
lecture3.tex
File metadata and controls
311 lines (254 loc) · 13.8 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
\documentclass[11pt]{article}
%%%%%%%%% options for the file macros.tex
\def\showauthornotes{1}
\def\showkeys{0}
\def\showdraftbox{0}
% \allowdisplaybreaks[1]
\input{macros}
\allowdisplaybreaks
\usepackage{tikz}
\usepackage[
backend=biber,
% giveninits=true,
% natbib=true,
style=alphabetic,
url=false,
% doi=true,
hyperref,
backref=true,
backrefstyle=none,
maxbibnames=10,
sortcites
]{biblatex}
\addbibresource{papers.bib}
%%%%%%%%% Authornotes
\newcommand{\Snote}{\Authornote{S}}
\newenvironment{tight_enumerate}{
\begin{enumerate}
\setlength{\itemsep}{2pt}
\setlength{\parskip}{1pt}
}{\end{enumerate}}
\newenvironment{tight_itemize}{
\begin{itemize}
\setlength{\itemsep}{2pt}
\setlength{\parskip}{1pt}
}{\end{itemize}}
\addbibresource{refs.bib}
%%%%%%%%%%%%%%%%%%%%%%%%%
\begin{document}
\newcommand{\coursenum}{{CSC 2421H}}
\newcommand{\coursename}{{Graphs, Matrices, and Optimization}}
\newcommand{\courseprof}{Sushant Sachdeva}
\lecturetitle{3}{Cheeger's Inequality}{Muhammed Tahsin Rahman}{24 Sep 2018}
\section{Recall}
The conductance of a graph $G = (V,E)$ is defined as
\begin{equation}
\Phi(G) = \ \min_{\substack{ S \subset V \\ S \neq \varnothing}} \ \dfrac{\abs{E(S,\overline{S})}}{\min{\{\vol(S),\vol(\overline{S})\}}} \,,
\end{equation}
where $\overline{S} = S^c$, $\abs{E(S,\overline{S})} = \sum_{(u,v)} w_{u,v}\V{\mathbbm{1}}[(u,v) \in E(S,\overline{S}) ]$, and $\vol(S) = \sum_{v \in S} \text{deg}(v)$. Cheeger's inequality states that the conductance is bounded by
\begin{equation} \label{eq.cheegers}
\frac{\nu_2}{2} \leq \Phi(G) \leq \sqrt{2\nu_2}.
\end{equation}
$\nu_2$ is the second smallest eigenvalue of the normalized graph Laplacian, defined as
\begin{equation}
\nu_2 = \min_{\V{y}^{\top}\V{D}\V{\mathbbm{1}}=0} \dfrac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}} = \lambda_2(\V{N}),
\end{equation}
where $\V{y} \in\mathbb{R}^V$, $\V{D}$ is the degree matrix, $\V{L}$ is the graph Laplacian, and $\V{N}=\V{D}^{-1/2}\V{L}\V{D}^{-1/2}$ is the normalized graph Laplacian.
In the previous lecture we had proved the left side of eq. \ref{eq.cheegers}, and in this lecture we will prove the right side. To this end, we first state three lemmas; then, we use these lemmas to derive the bound; finally, we prove the lemmas. We will also begin the next topic on Random Walks.
%Section Stating Lemmas
\section{Required Lemmas}
%Lemma1
\begin{lemma} \label{lem1}
Given a vector $\V{y} \in \mathbb{R}^V$ \underline{s.t.} $\V{y}^{\top}\V{D}\mathbbm{1}=0$, we can find a vector $\V{z} \in \mathbb{R}^V$ \underline{s.t.}
\begin{align*}
\V{z} &\geq \V{0} \\
\dfrac{\V{z}^{\top}\V{L}\V{z}}{\V{z}^{\top}\V{D}\V{z}} &\leq \dfrac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}}, \,\,\, and \\
\vol(\mathrm{supp}(\V{z})) &\leq \frac{1}{2} \vol(V),
\end{align*}
where $\mathrm{supp}(\V{z}) = \{ v : \V{z}(v) > 0 \}$.
\end{lemma}
%Lemma2
\begin{lemma} \label{lem2}
Given a vector $\V{z} \in \mathbb{R}^V$ \underline{s.t.} $\V{z}\geq\V{0}$, we can sample a scalar $t$, where $S_t \subseteq V$, \underline{s.t.} $S_t \subseteq \mathrm{supp}(\V{z})$. Then,
\begin{equation*}
\dfrac{\mathbb{E}_t[\abs{E(S_t,\overline{S_t})}]}{\mathbb{E}_t[\vol(S_t)]} \leq \sqrt{ 2\dfrac{\V{z}^{\top}\V{L}\V{z}}{\V{z}^{\top}\V{D}\V{z}} }.
\end{equation*}
\end{lemma}
%Lemma3
\begin{lemma} \label{lem3}
Given a distribution over $t$ and associated random variables $X_t, Y_t$ \underline{s.t.} $Y_t > 0$, $\exists \ t_0$ \underline{s.t.}
\begin{equation*}
\dfrac{X_{t_0}}{Y_{t_0}} \leq \dfrac{\mathbb{E}_t[X_t]}{\mathbb{E}_t[Y_t]}.
\end{equation*}
\end{lemma}
%Using Lemmas to prove Cheeger's
\section{Cheeger's Upper Bound}
Let $\V{y} \in \mathbb{R}^V$ \underline{s.t.} $\V{y}^{\top}\V{D}\mathbbm{1}=0$ and apply Lemma \ref{lem1}. We obtain $\V{z} \geq \V{0}$, with $\vol(\mathrm{\V{z}}) \leq \frac{1}{2}\vol(V)$ and $\frac{\V{z}^{\top}\V{L}\V{z}}{\V{z}^{\top}\V{D}\V{z}} \leq \frac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}}$. Now, use this $\V{z}$ in Lemma \ref{lem2}, sampling $t$ with associated set $S_t \subseteq \mathrm{supp(\V{z})}$, to obtain $\vol(S_t) \leq \frac{1}{2}\vol(V)$ and $\frac{\mathbb{E}_t[\abs{E(S_t,\overline{S_t})}]}{\mathbb{E}_t[\vol(S_t)]} \leq \sqrt{ 2\frac{\V{z}^{\top}\V{L}\V{z}}{\V{z}^{\top}\V{D}\V{z}} } \leq \sqrt{ 2\frac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}} }$. Finally, apply Lemma \ref{lem3}, in that $\exists \ t_0$ that achieves $\Phi_G(S_{t_0}) = \frac{\abs{E(S_{t_0},\overline{S_{t_0}})}}{\vol(S_{t_0})} \leq \sqrt{ 2\frac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}} }$. To complete the proof plug in the $\V{y}$ that achieves $\nu_2$.
%Proving the Lemmas
\section{Proof of Lemmas}
%Proof of Lemma 1
\subsection{Lemma \ref{lem1}}
With $\V{y}^{\top}\V{D}\mathbbm{1}=0$, let
\begin{align*}
\V{\tilde{y}} &= \V{y} - c\mathbbm{1}, \\
\mathcal{V}_1 &= \vol(\{v:\V{\tilde{y}}(v)>0\}), \\
\mathcal{V}_2 &= \vol(\{v:\V{\tilde{y}}(v)<0\}).
\end{align*}
Note that $\mathcal{V}_1$ and $\mathcal{V}_2$ are strictly disjoint. As we change $c$ over the real line, the vertices move from $\mathcal{V}_1$ to $\mathcal{V}_2$, so there will be some $c$ \underline{s.t.}
\begin{equation*}
\mathcal{V}_1, \mathcal{V}_2 \leq \frac{1}{2}\vol(V).
\end{equation*}
Consider what happens to the ratio $\frac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}}$.
\begin{align*}
\V{\tilde{y}}^{\top}\V{L}\V{\tilde{y}} = \sum_{(u,v)} (&\V{\tilde{y}}(u) - \V{\tilde{y}}(v))^2 = \V{y}^{\top}\V{L}\V{y}, \\
\V{\tilde{y}}^{\top}\V{D}\V{\tilde{y}} = \V{y}^{\top}\V{D}\V{y} - 2&c\underbrace{\V{y}^{\top}\V{D}\mathbbm{1}}_{\footnotesize = \, 0} + \underbrace{c^2\mathbbm{1}^{\top}\V{D}\mathbbm{1}}_{\footnotesize \geq \, 0} \geq \V{y}^{\top}\V{D}\V{y}, \,\,\, \mathrm{and} \\
\dfrac{\V{\tilde{y}}^{\top}\V{L}\V{\tilde{y}}}{\V{\tilde{y}}^{\top}\V{D}\V{\tilde{y}}} &\leq \dfrac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}}.
\end{align*}
Now write $\V{\tilde{y}}=\V{z^+} - \V{z^-}$, where $\V{z^+}, \V{z^-} \geq \V{0}$; then,
\begin{align*}
\vol(\mathrm{supp}(\V{z^+})), &\ \vol(\mathrm{supp}(\V{z^-})) \leq \frac{1}{2}\vol(V), \\
\V{\tilde{y}}^{\top}\V{D}\V{\tilde{y}} &= \sum_v \text{deg}(v)\V{\tilde{y}}(v)^2 \\
&= \V{z^+}^{\top}\V{D}\V{z^+}+\V{z^-}^{\top}\V{D}\V{z^-}.
\end{align*}
Note that
\begin{align*}
\V{\tilde{y}}^{\top}\V{L}\V{\tilde{y}} &= \sum_{(u,v)} (\V{\tilde{y}}(u) - \V{\tilde{y}}(v))^2 \\
&= \sum_{(u,v)} ( (\V{z^+}(u) - \V{z^+}(v)) - (\V{z^-}(u) - \V{z^-}(v)) )^2 \\
&\geq \sum_{(u,v)} (\V{z^+}(u) - \V{z^+}(v))^2 + (\V{z^-}(u) - \V{z^-}(v))^2, \\
\end{align*}
since the cross terms of the expansion will always be $\geq 0$. Now, with a proof similar to Lemma \ref{lem3} shown in section \ref{sec.lem3},
\begin{equation*}
\dfrac{\V{z^+}^{\top}\V{L}\V{z^+}+\V{z^-}^{\top}\V{L}\V{z^-}}{\V{z^+}^{\top}\V{D}\V{z^+}+\V{z^-}^{\top}\V{D}\V{z^-}} \leq \dfrac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}}.
\end{equation*}
One of $\V{z^+}, \V{z^-}$ achieves a ratio $\leq \frac{\V{y}^{\top}\V{L}\V{y}}{\V{y}^{\top}\V{D}\V{y}}$.
%Proof of Lemma 2
\subsection{Lemma \ref{lem2}}
%Number line figure
\begin{wrapfigure}{r}{0.25\textwidth}
\centering
\includegraphics[width=0.22\textwidth]{images/lemma2}
\caption{Sampling $t$ and the corresponding $S_t$}
\label{fig.lemma2}
\end{wrapfigure}
With $\V{z} \geq \V{0}$, scale $\V{z}$ \underline{s.t.} $\max(\V{z})=1$. This scaling is acceptable because the claim depends on ratios with $\V{z}$ and not the absolute values. Now, sample a threshold $t \in [0,1]$ using the probability density function $f(t) = 2t$. Note that $\pr[t \in [a,b]] = \int_{a}^{b}f(t) dt = b^2 - a^2$. This distribution is picked as a proof trick and does not affect the outcome of Cheeger's inequality.
Define $S_t = \{v:\V{z}(v)>t\}$. So, picking $S_t$ depends on $t$ discretely. If $t$ is between $a$ \& $b$, then $\pr[v \in S_t] = b^2 - a^2$. This selection of $S_t$ is shown in Figure \ref{fig.lemma2}. Next, the denominator of Lemma \ref{lem2}
\begin{align*}
\mathbb{E}_t[\vol(S_t)] &= \mathbb{E}_t\bigg[\sum_v\mathrm{deg}(v)\mathbbm{1}[v\in S_t]\bigg], \\
&= \sum_v \mathrm{deg}(v) \underbrace{\mathbb{E}_t[\mathbbm{1}[v\in S_t]]}_{\parbox{6.5em}{\raggedright\scriptsize$\pr[v \in S_t] =\pr[t \in [0,\V{z}(v)]] =\V{z}(v)^2$}}, \\
&= \sum_v \mathrm{deg}(v) \V{z}(v)^2 \\
&= \V{z}^{\top}\V{D}\V{z}.
\end{align*}
For the numerator,
\begin{align*}
\mathbb{E}_t[\abs{E(S,\overline{S})}] &= \mathbb{E}_t\bigg[\sum_{(u,v)}\mathbbm{1}[(u,v)\in E(S_t,\overline{S_t})]\bigg] \\
&= \sum_{(u,v)}\pr[(u,v) \in E(S_t,\overline{S_t})] \\
&= \sum_{(u,v)}\abs{\V{z}(u)^2-\V{z}(v)^2}.
\end{align*}
Applying Cauchy-Shwartz inequality $\bigg( \sum_ia_ib_i \leq \sqrt{\sum_ia_i^2} \sqrt{\sum_ib_i^2} \bigg)$ gives
\begin{align*}
\sum_{(u,v)}\abs{\V{z}(u)^2-\V{z}(v)^2} &\leq \sqrt{(\underbrace{\sum_{(u,v)}(\V{z}(u)-\V{z}(v))^2}_{\footnotesize \text{Laplacian quadratic form}})(\sum_{(u,v)}(\V{z}(u)+\V{z}(v))^2)}, \\
\sum_{(u,v)}(\V{z}(u)+\V{z}(v))^2 &\leq 2\sum_{(u,v)}(\V{z}(u)^2+\V{z}(v)^2) \\
&= 2\sum_v\text{deg}(v)\V{z}(v)^2 \\
&= 2\V{z}^{\top}\V{D}\V{z}.
\end{align*}
Therefore,
\begin{equation*}
\dfrac{\mathbb{E}_t[\abs{E(S,\overline{S})}]}{\mathbb{E}_t[\vol(S_t)]} \leq \dfrac{\sqrt{2(\V{z}^{\top}\V{L}\V{z})(\V{z}^{\top}\V{D}\V{z})}}{\V{z}^{\top}\V{D}\V{z}} = \sqrt{\ 2\ \dfrac{\V{z}^{\top}\V{L}\V{z}}{\V{z}^{\top}\V{D}\V{z}}}
\end{equation*}
%Proof of Lemma 3
\subsection{Lemma \ref{lem3}} \label{sec.lem3}
Assuming the collection of $X_t, Y_t$ is finite, let $r$ denote the
minimum ratio, and without loss of generality, assume that it is
achieved for $t=0.$
\begin{align*}
r & \defeq \min_t \dfrac{X_t}{Y_t} \qquad \text{\underline{s.t.}} \, Y_t>0, \\
&= \dfrac{X_0}{Y_0}.
\end{align*}
% \noindent We want to prove that
% \[\frac{X_0}{Y_0} = r \le \frac{\av_t[X_t]}{\av_t[Y_t]}.\]
% \noindent We assume to the contrary that $r > $ that $r < $
The fact that $r$ is a minimum means that for all $t,$ we
have $r \le \frac{X_t}{Y_t},$ and hence $ r Y_t \le X_t$ (since $Y_t >
0$). Taking
expectation over $t,$ we get that
\[r \av_t [Y_t] \le \av_t[X_t],\]
or equivalently,
\[r \le \frac{\av_t[X_t]}{\av_t[Y_t]}.\]
(Note that $Y_t > 0$ for all $t,$ and hence $\av_t[Y_t] > 0.$)
Since $r = \frac{X_0}{Y_0},$ this proves the claim.
% \begin{align*}
% \dfrac{X_t}{Y_t} &\geq r, \\
% X_t &\geq rY_t, \\
% \mathbb{E}_t[X_t] &\geq r\mathbb{E}_t[Y_t]
% \end{align*}
%Examples
\section{Examples}
Given a vector $\V{y}$, how do we get a `good' cut? Consider two examples that we have seen previously, a single edge between two vertices and a ring graph.
%example1
\subsection{Two Vertices Connected by an Edge}
We know that $\Phi(G)=1$. Using Cheeger's inequality, we first compute the second smallest eigenvalue of the normalized graph Laplacian. Drawing from the first lecture,
\begin{align*}
\nu_2 &= \lambda_2(\V{D}^{-1/2}\V{L}\V{D}^{-1/2})\\
&= \lambda_2(\V{L}) \\
&= 2 \\
&= 2\Phi(G),
\end{align*}
showing that the lower bound of Cheeger's inequality is tight
%example2
\subsection{Ring Graph}
\begin{lemma}
For $R_n$, define
\begin{align*}
x_k(v) &= \sin(\frac{2\pi kv}{n}), \quad 0 \leq k \leq \frac{n}{2}, \quad \mathrm{and} \\
y_k(v) &= \cos(\frac{2\pi kv}{n}), \quad 0 \leq k \leq \frac{n}{2}.
\end{align*}
$\forall k$, $x_k, y_k$ are eigenvectors of $L_{R_n}$, with eigenvalues $2(1-\cos(\frac{2\pi k}{n}))$.
\end{lemma}
Proof of the above lemma is left as an exercise.
For the simple case of $n$ even,
\begin{equation*}
\Phi(R_n) = \dfrac{2}{2.\frac{n}{2}} = \dfrac{2}{n}
\end{equation*}
Knowing that $\V{D}=2\V{I}$,
\begin{align*}
\nu_2 &= \lambda_2(\V{D}^{-1/2}\V{L}\V{D}^{-1/2}) \\
&= \frac{1}{2}\lambda_2(\V{L}) \\
&= \frac{1}{2} (2-2\cos (\dfrac{2\pi}{n})) \\
&= 2 \sin^2(\dfrac{\pi}{n}) \\
&= 2 (\dfrac{\pi}{n})^2(1+o(1)), \quad \text{for large $n$}.
\end{align*}
Note that for a tight bound, we would need to take the square root of the above expression, as proven earlier.
As an exercise, try finding graph cuts using Julia. Use the previous lecture's formulation and then use Cheeger's inequality to make cuts.
\newpage
\section{Random Walks}
%Random walk graph
\begin{wrapfigure}{r}{0.20\textwidth}
\centering
\includegraphics[width=0.20\textwidth]{images/randomwalk}
\caption{Example of random walk graph}
\label{fig.randomwalk}
\end{wrapfigure}
We will now be looking at undirected graphs that capture reversible Markov chains. Consider the graph in Figure \ref{fig.randomwalk}. Start with a peg on vertex $a$ at time $0$. Then, at time $1$, pick uniformly between neighbours to move to; in this case, $\pr[\text{$v=b$ at $t=1 \ |\ v=a$ at $t=0$}] = \pr[\text{$v=c$ at $t=1 \ |\ v=a$ at $t=0$}] = 0.5$. If the graph is weighted, scale the transition probabilities using the weights. The resulting sequence of vertices that have been visited over time is called the transcript of the random walk.
Given a graph and initial vertex, we are interested in answering questions such as,
\begin{itemize}
\item What is the distribution over vertices after a given number of steps?
\item Is there a stationary distribution?
\item How quickly do we converge?
\end{itemize}
The state evolution can be quantified using knowledge of the graph structure. At time $t$, define $\V{p}_t \in \mathbb{R}^V, \V{p}_t \geq \V{0},$ and $\mathbbm{1}^{\top}\V{p}_t = \sum_\V{v}\V{p}_t(v)=1$; let
\begin{align*}
\V{p}_0(v) &= \begin{cases}1 \quad &\text{if} \,\,\, v=a\\0 &\text{else}\end{cases} \quad \text{(starting at a)}, \\
\V{p}_t(v) &= \pr[\text{the random walk is at vertex $v$ at time $t$}], \\
\V{p}_{t+1}(v) &= \sum_{u:(u,v)} \dfrac{1}{\text{deg}(u)}\V{p}_t(u) \quad \text{(unweighted)}.
\end{align*}
For weighted graphs, deriving the following is left as an exercise:
\begin{align*}
\pr[\text{$v$ at $t+1$ $|$ $u$ at $t$}] &= \dfrac{w(u,v)}{\sum_{z:(u,z)}w(u,z)} = \dfrac{w(u,v)}{\underbrace{\text{deg}(u)}_{\footnotesize\text{weighted deg}}}, \\
\V{p}_{t+1}(v) &= \sum_{u:(u,v)} \dfrac{w(u,v)}{\text{deg}(u)}\V{p}_t(u), \\
\V{p}_{t+1} &= \V{A}\V{D}^{-1}\V{p}_t.
\end{align*}
\end{document}
%%% Local Variables:
%%% mode: latex
%%% TeX-master: t
%%% End: