Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric...

124
Metric embeddings and geometric inequalities Lectures by Professor Assaf Naor Scribes: Holden Lee and Kiran Vodrahalli April 27, 2016

Transcript of Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric...

Page 1: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Metric embeddings and geometric inequalities

Lectures by Professor Assaf NaorScribes: Holden Lee and Kiran Vodrahalli

April 27, 2016

Page 2: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

2

Page 3: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Contents

1 The Ribe Program 71 The Ribe Program . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72 Bourgain’s Theorem implies Ribe’s Theorem . . . . . . . . . . . . . . . . . . 14

2.1 Bourgain’s discretization theorem . . . . . . . . . . . . . . . . . . . . 152.2 Bourgain’s Theorem implies Ribe’s Theorem . . . . . . . . . . . . . . 17

2 Restricted invertibility principle 191 Restricted invertibility principle . . . . . . . . . . . . . . . . . . . . . . . . . 19

1.1 The first restricted invertibility principles . . . . . . . . . . . . . . . . 191.2 A general restricted invertibility principle . . . . . . . . . . . . . . . . 21

2 Ky Fan maximum principle . . . . . . . . . . . . . . . . . . . . . . . . . . . 233 Finding big subsets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

3.1 Little Grothendieck inequality . . . . . . . . . . . . . . . . . . . . . . 263.1.1 Tightness of Grothendieck’s inequality . . . . . . . . . . . . 29

3.2 Pietsch Domination Theorem . . . . . . . . . . . . . . . . . . . . . . 303.3 A projection bound . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323.4 Sauer-Shelah Lemma . . . . . . . . . . . . . . . . . . . . . . . . . . . 33

4 Proof of RIP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 344.1 Step 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 344.2 Step 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

3 Bourgain’s Discretization Theorem 471 Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

1.1 Reduction to smooth norm . . . . . . . . . . . . . . . . . . . . . . . . 482 Bourgain’s almost extension theorem . . . . . . . . . . . . . . . . . . . . . . 49

2.1 Lipschitz extension problem . . . . . . . . . . . . . . . . . . . . . . . 502.2 John ellipsoid . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 512.3 Proof . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

3 Proof of Bourgain’s discretization theorem . . . . . . . . . . . . . . . . . . . 573.1 The Poisson semigroup . . . . . . . . . . . . . . . . . . . . . . . . . . 57

4 Upper bound on π›Ώπ‘‹β†’π‘Œ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 664.1 Fourier analysis on the Boolean cube and Enflo’s inequality . . . . . . 69

5 Improvement for 𝐿𝑝 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70

3

Page 4: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

6 Kirszbraun’s Extension Theorem . . . . . . . . . . . . . . . . . . . . . . . . 79

4 Grothendieck’s Inequality 911 Grothendieck’s Inequality . . . . . . . . . . . . . . . . . . . . . . . . . . . . 912 Grothendieck’s Inequality for Graphs (by Arka and Yuval) . . . . . . . . . . 933 Noncommutative Version of Grothendieck’s Inequality - Thomas and Fred . . 974 Improving the Grothendieck constant . . . . . . . . . . . . . . . . . . . . . . 1025 Application to Fourier series . . . . . . . . . . . . . . . . . . . . . . . . . . . 1046 ? presentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105

5 Lipschitz extension 1111 Introduction and Problem Setup . . . . . . . . . . . . . . . . . . . . . . . . . 111

1.1 Why do we care? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1121.2 Outline of the Approach . . . . . . . . . . . . . . . . . . . . . . . . . 113

2 π‘Ÿ-Magnification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1133 Wasserstein-1 norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1144 Graph theoretic lemmas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116

4.1 Expanders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1164.2 An Application of Menger’s Theorem . . . . . . . . . . . . . . . . . . 118

5 Main Proof . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118

4

Page 5: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Introduction

The topic of this course is geometric inequalities with applications to metric embeddings; andwe are actually going to do things more general than metric embeddings: Metric geometry.Today I will roughly explain what I want to cover and hopefully start proving a first majortheorem. The strategy for this course is to teach novel work. Sometimes topics will becovered in textbooks, but a lot of these things will be a few weeks old. There are also somethings which may not have been even written yet. I want to give you a taste for what’s goingon in the field. Notes from the course I taught last spring are also available.

One of my main guidelines in choosing topics will be topics that have many accessibleopen questions. I will mention open questions as we go along. I’m going to really choosetopics that have proofs which are entirely self-contained. I’m trying to assume nothing. Myintention is to make it completely clear and there should be no steps you don’t understand.

Now this is a huge area. I will explain some of the background today. I’m going to useproofs of some major theorems as excuses to see the lemmas that go into the proofs. Someof the theorems are very famous and major, and you’re going to see some improvements, butalong the way, we will see some lemmas which are immensely powerful. So we will always beproving a concrete theorem. But actually somewhere along the way, there are lemmas whichhave wide applicability to many many areas. These are excuses to discuss methods, thoughthe theorems are important.

The course can go in many directions: If some of you have some interests, we can alwayschange the direction of the course, so express your interests as we go along.

5

Page 6: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

6

Page 7: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Chapter 1

The Ribe Program

2/1/16

1 The Ribe Program

The main motivation for most of what we will discuss is called the Ribe Program, which is aresearch program many hundreds of papers large. We will see some snapshots of it, and it allcomes from a theorem from 1975, Ribe’s rigidity theorem 1.1.9, which we will state nowand prove later in a modern way. This theorem was Martin Ribe’s dissertation, which starteda whole direction of mathematics, but after he wrote his dissertation he left mathematics.He’s apparently a government official in Sweden. The theorem is in the context of Banachspaces; a relation between their linear structure and their structure as metric spaces. Nowfor some terminology.

Definition 1.1.1 (Banach space): A Banach space is a complete, normed vector space.Therefore, a Banach space is equipped with a metric which defines vector length and distancesbetween vectors. It is complete, so every Cauchy sequence of converges to a limit definedinside the space.

Definition 1.1.2: Let (𝑋, ‖·‖𝑋), (π‘Œ, β€–Β·β€–π‘Œ ) be Banach spaces. We say that 𝑋 is (crudely)finitely representable in π‘Œ if there exists some constant 𝐾 > 0 such that for every finite-dimensional linear subspace 𝐹 βŠ† 𝑋, there is a linear operator 𝑆 : 𝐹 β†’ π‘Œ such that for everyπ‘₯ ∈ 𝐹 ,

β€–π‘₯‖𝑋 ≀ ‖𝑆π‘₯β€–π‘Œ ≀ 𝐾 β€–π‘₯‖𝑋 .

Note 𝐾 is decided once and for all, before the subspace 𝐹 is chosen.(Some authors use β€œfinitely representable” to mean that this is true for any 𝐾 = 1 + πœ€.

We will not follow this terminology.)Finite representability is important because it allows us to conclude that 𝑋 has the

same finite dimensional linear properties (local properties) as π‘Œ . That is, it preserves anyinvariant involves finitely many vectors, their lengths, etc.

7

Page 8: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Let’s introduce some local properties like type. To motivate the definition, consider thetriangle inequality, which says

‖𝑦1 + Β· Β· Β·+ π‘¦π‘›β€–π‘Œ ≀ ‖𝑦1β€–π‘Œ + Β· Β· Β·+ β€–π‘¦π‘›β€–π‘Œ .

In what sense can we improve the triangle inequality? In 𝐿1 this is the best you can say. Inmany spaces there are ways to improve it if you think about it correctly.

For any choice πœ€1, . . . , πœ€π‘› ∈ {Β±1},π‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘Œ

β‰€π‘›βˆ‘π‘–=1

β€–π‘¦π‘–β€–π‘Œ .

Definition 1.1.3: df:type Say that 𝑋 has type 𝑝 if there exists 𝑇 > 0 such that for every𝑛, 𝑦1, . . . , 𝑦𝑛 ∈ π‘Œ ,

Eπœ€βˆˆ{Β±1}𝑛

π‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘Œ

≀ 𝑇

[π‘›βˆ‘π‘–=1

β€–π‘¦π‘—β€–π‘π‘Œ

] 1𝑝

.

The 𝐿𝑝 norm is always at most the 𝐿1 norm; if the lengths are spread out, this is asymptot-ically much better. Say π‘Œ has nontrivial type if 𝑝 > 1.

For example, 𝐿𝑝(πœ‡) has type min(𝑝, 2).

Later we’ll see a version of β€œtype” for metric spaces. How far is the triangle inequalityfrom being an equality is a common theme in many questions. In the case of normed spaces,this controls a lot of the geometry. Proving a result for 𝑝 > 1 is hugely important.

Proposition 1.1.4: pr:finrep-type If 𝑋 is finitely representable and π‘Œ has type 𝑝 then also 𝑋has type 𝑝.

Proof. Let π‘₯1, . . . , π‘₯𝑛 ∈ 𝑋. Let 𝐹 = span{π‘₯1, . . . , π‘₯𝑛}. Finite representability gives me𝑆 : 𝐹 β†’ π‘Œ . Let 𝑦𝑖 = 𝑆π‘₯𝑖. What can we say about

βˆ‘πœ€π‘–π‘¦π‘–?

Eπœ€

π‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘Œ

= Eπœ€

𝑆(

π‘›βˆ‘π‘–=1

πœ€π‘–π‘₯𝑖)

π‘Œ

β‰₯ Eπœ€

π‘›βˆ‘π‘–=1

πœ€π‘–π‘‹π‘–

𝑋

Eπœ€

π‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘Œ

≀ 𝑇

(π‘›βˆ‘π‘–=1

‖𝑆π‘₯𝑖‖𝑝) 1

𝑝

≀ 𝑇𝐾

(π‘›βˆ‘π‘–=1

β€–π‘₯𝑖‖𝑝) 1

𝑝

.

Putting these two inequalities together gives the result.

8

Page 9: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Theorem 1.1.5 (Kahane’s inequality). For any normed space π‘Œ and π‘ž β‰₯ 1, for all 𝑛,𝑦1, . . . , 𝑦𝑛 ∈ π‘Œ ,

Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

&π‘ž

(E[

π‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘žπ‘Œ

]) 1π‘ž

.

Here &π‘ž means β€œup to a constant”; subscripts say what the constant depends on. The constanthere does not depend on the norm π‘Œ .

Kahane’s Theorem tells us that the LHS of Definition 1.1.3 can be replaced by any norm,if we change ≀ to .. We get that having type 𝑝 is equivalent to

Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘π‘Œ

. 𝑇 π‘π‘›βˆ‘π‘–=1

β€–π‘¦π‘–β€–π‘π‘Œ .

Recall the parallelogram identity in a Hilbert space 𝐻:

Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

2=

π‘›βˆ‘π‘–=1

‖𝑦𝑖‖2𝐻 .

A different way to understand the inequality in the definition of β€œtype” is: how far is a givennorm from being an Euclidean norm? The Jordan-von Neumann Theorem says that ifparallelogram identity holds then it’s a Euclidean space. What happes if we turn it in aninequality?

Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

2𝐻

β‰₯≀ 𝑇

π‘›βˆ‘π‘–=1

‖𝑦𝑖‖2𝐻 .

Either inequality still characterizes a Euclidean space.What happens if we add constants or change the power? We recover the definition for

type and cotype (which has the inequality going the other way):

Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘žπ»

&.

π‘›βˆ‘π‘–=1

β€–π‘¦π‘–β€–π‘žπ» .

Definition 1.1.6: Say it has cotype π‘ž if

π‘›βˆ‘π‘–=1

β€–π‘¦π‘–β€–π‘žπ‘Œ . πΆπ‘žEπ‘›βˆ‘π‘–=1

πœ€π‘–π‘¦π‘–

π‘žπ‘Œ

R. C. James invented the local theory of Banach spaces, the study of geometry thatinvolves properties involving finitely many vectors (βˆ€π‘₯1, . . . , π‘₯𝑛, 𝑃 (π‘₯1, . . . , π‘₯𝑛) holds). As acounterexample, reflexivity cannot be characterized using finitely many vectors (this is atheorem).

Ribe discovered link between metric and linear spaces.First, terminology.

9

Page 10: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Definition 1.1.7: Two Banach spaces are uniformly homeomorphic if there exists 𝑓 :𝑋 β†’ π‘Œ that is 1-1 and onto and 𝑓, π‘“βˆ’1 are uniformly continuous.

Without the word β€œuniformly”, if you think of the spaces as topological spaces, all ofthem are equivalent. Things become interesting when you quantify! β€œUniformly” meansyou’re controlling the quantity.

Theorem 1.1.8 (Kadec). Any two infinite-dimensional separable Banach spaces are home-omorphic.

This is a amazing fact: these spaces are all topologically equivalent to Hilbert spaces!Over time people people found more examples of Banach spaces that are homeomorphic

but not uniformly homeomorphic. Ribe’s rigidity theorem clarified a big chunk of what washappening.

Theorem 1.1.9 (Rigidity Theorem, Martin Ribe (1975)). thm:ribe Suppose that 𝑋, π‘Œ are uni-formly homeomorphic Banach spaces. Then 𝑋 is finitely representable in π‘Œ and π‘Œ is finitelyrepresentable in 𝑋.

For example, for 𝐿𝑝 and πΏπ‘ž, for 𝑝 = π‘ž it’s always that case that one is not finitely repre-sentable in the other, and hence by Ribe’s Theorem, 𝐿𝑝, πΏπ‘ž are not uniformly homeomorphic.(When I write 𝐿𝑝, I mean 𝐿𝑝(R).)

Theorem 1.1.10. For every 𝑝 β‰₯ 1, 𝑝 = 2, 𝐿𝑝 and ℓ𝑝 are finitely representable in each other,yet not uniformly homeomorphic.

(Here ℓ𝑝 is the sequence space.)

Exercise 1.1.11: Prove the first part of this theorem: 𝐿𝑝 is finitely representable in ℓ𝑝.

Hint: approximate using step functions. You’ll need to remember some measure theory.When 𝑝 = 2, 𝐿𝑝, ℓ𝑝 are separable and isometric.The theorem in various cases was proved by:

1. 𝑝 = 1: Enflo

2. 1 < 𝑝 < 2: Bourgain

3. 𝑝 > 2: Gorelik, applying the Brouwer fixed point theorem (topology)

Every linear property of a Banach signs which is local (type, cotype, etc.; involvingsumming, powers, etc.) is preserved under a general nonlinear deformation.

After Ribe’s rigidity theorem, people wondered: can we reformulate the local theoryof Banach spaces without mentioning anything about the linear structure? Ribe’s rigiditytheorem is more of an existence statement, we can’t see anything about an explicit dictionarywhich maps statements about linear sums into statements about metric spaces. So peoplestarted to wonder whether we could reformulate the local theory of Banach spaces, but only

10

Page 11: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

looks at distances between pairs instead of summing things up. Local theory is one of thehugest subjects in analysis. If you could actually find a dictionary which takes one lineartheorem at a time, and restate it with only distances, there is a huge potential here! Becausethe definition of type only involves distances between points, we can talk about a metricspace’s type or cotype. So maybe we can use the intution given by linear arguments, andthen state things for metric spaces which often for very different reasons remain true from thelinear domain. And then now maybe you can apply these arguments to graphs, or groups!We could be able to prove things about the much more general metric spaces. Thus, we endup applying theorems on linear spaces in situations with a priori nothing to do with linearspaces. This is massively powerful.

There are very crucial entries that are missing in the dictionary. We don’t even now howto define many of the properties! This program has many interesting proofs. Some of themost interesting conjectures are how to define things!

Corollary 1.1.12. cor:uh-type If 𝑋, π‘Œ are uniformly homeomorphic and if one of them is oftype 𝑝, then the other does.

This follows from Ribe’s Theorem 1.1.9 and Proposition 1.1.4. Can we prove somethinglike this theorem without using Ribe’s Theorem 1.1.9? We want to reformulate the definitionof type using only the distance, so this becomes self-evident.

Enflo had an amazing idea. Suppose 𝑋 is a Banach space, π‘₯1, . . . , π‘₯𝑛 ∈ 𝑋. The type 𝑝inequality says

eq:type-pE[

π‘›βˆ‘π‘–=1

πœ€π‘–π‘₯𝑖

𝑝].𝑋

π‘›βˆ‘π‘–=1

β€–π‘₯𝑖‖𝑝 . (1.1)

Let’s rewrite this in a silly way. Define 𝑓 : {Β±1}𝑛 β†’ 𝑋 by

𝑓(πœ€1, . . . , πœ€π‘›) =π‘›βˆ‘π‘–=1

πœ€π‘–π‘₯𝑖.

Write πœ€ = (πœ€1, . . . , πœ€π‘›). Multiplying by 2𝑛, we can write the inequality (1.1) as

eq:type-genE [‖𝑓(πœ€)βˆ’ 𝑓(βˆ’πœ€)‖𝑝] .𝑋

π‘›βˆ‘π‘–=1

E [‖𝑓(πœ€)βˆ’ 𝑓(πœ€1, . . . , πœ€π‘–βˆ’1,βˆ’πœ€π‘–, πœ€π‘–+1, . . . , πœ€π‘›)‖𝑝] . (1.2)

This inequality just involves distances between points 𝑓(πœ€), so it is the reformulation weseek.

Definition 1.1.13: df:enflo A metric space (𝑋, 𝑑𝑋) has Enflo type 𝑝 if there exists 𝑇 > 0such that for every 𝑛 and every 𝑓 : {Β±1}𝑛 β†’ 𝑋,

E[𝑑𝑋(𝑓(πœ€), 𝑓(βˆ’πœ€))𝑝] ≀ 𝑇 π‘π‘›βˆ‘π‘–=1

E[𝑑𝑋(𝑓(πœ€), 𝑓(πœ€1, . . . , πœ€π‘–βˆ’1,βˆ’πœ€π‘–, πœ€π‘–+1, . . . , πœ€π‘›))𝑝].

11

Page 12: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

This is bold. It wasn’t true before for a general function! The discrete cube (Booleanhypercube) {Β±1}𝑛 is all the πœ– vectors, of which there are 2𝑛. Our function just assigns 2𝑛

points arbitrarily. No structure whatsoever. As they are indexed this way, you see nothing.But you’re free to label them by vertices of the cube however you want. But there are manylabelings! In (1.2), the points had to be vertices of a cube, but in Definition 1.1.13, theyare arbitrary. The moment you choose the labelings, you impose a cube structure betweenthe points. Some of them are diagonals of the cube, some of them are edges. πœ– and βˆ’πœ– areantipodal points. But it’s not really a diagonal. They are points on a manifold, and area function of how you decided to label them. What this sum says is that the sum over alldiagonals, the length of the diagonals to the power 𝑝 is less than the sum over edges to theπ‘π‘‘β„Ž powers (these are the points where one πœ–π‘– is different). Thus we can seeβˆ‘

diag𝑝 .𝑋

βˆ‘edge𝑝.

The total 𝑝th power of lengths of diagonals is up to a constant, at most the same thing overall edges.

This is a vast generalization of type; we don’t even know a Banach space satisfies this.The following is one of my favorite conjectures.

Conjecture 1.1.14 (Enflo). If a Banach space has type 𝑝 then it also has Enflo type 𝑝.

This has been open for 40 years. We will prove the following.

Theorem 1.1.15 (Bourgain-Milman-Wolfson, Pisier). If 𝑋 is a Banach space of type 𝑝 > 1then 𝑋 also has type π‘βˆ’ πœ€ for every πœ€ > 0.

If you know the type inequality for parallelograms, you get it for arbitrary sets of points,up to πœ€. Basically, you’re getting arbitrarily close to 𝑝 instead of getting the exact result.We also know that the conjecture stated before is true for a lot of specific Banach spaces,though we do not yet have the general result. For instance, this is true for the 𝐿4 norm.Index functions by vertices; some pairs are edges, some are diagonals; then the 𝐿4 norm ofthe diagonals is at most that of the edges.

12

Page 13: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

How do you go from knowing this for a linear function to deducing this for an arbitraryfunction?

Once you do this, you have a new entry in the hidden Ribe dictionary. If 𝑋 and π‘Œare uniformly homeomorphic Banach spaces and π‘Œ has Enflo type 𝑝, then so is 𝑋. Theminute you throw away the linear structure, Corollary 1.1.12 becomes easy. It requires atiny argument. Now you can take a completely arbitrary function 𝑓 : {Β±1}𝑛 β†’ 𝑋. Thereexists a homeomorphism πœ“ : 𝑋 β†’ π‘Œ such that πœ“, πœ“βˆ’1 are uniformly continuous. Now wewant to deduce that the same inequality in π‘Œ gives the same inequality in 𝑋.

Proposition 1.1.16: pr:uh-enflo If 𝑋, π‘Œ are uniformly homeomorphic Banach spaces and π‘Œhas Enflo type 𝑝, then so does 𝑋.

This is an example of making the abstract Ribe theorem explicit.

Lemma 1.1.17 (Corson-Klee). lem:corson-klee If 𝑋, π‘Œ are Banach spaces and πœ“ : 𝑋 β†’ π‘Œ areuniformly continuous, then for every π‘Ž > 0 there exists 𝐿(π‘Ž) such that

β€–π‘₯1 βˆ’ π‘₯2‖𝑋 β‰₯ π‘Ž =β‡’ β€–πœ“(π‘₯1)βˆ’ πœ“(π‘₯2)β€– ≀ 𝐿 β€–π‘₯1 βˆ’ π‘₯2β€– .Proof sketch of 1.1.16 given Lemma 1.1.17. By definition of uniformly homeomorphic, thereexists a homeomorphism πœ“ : 𝑋 β†’ π‘Œ such that πœ“, πœ“βˆ’1 are uniformly continuous. Lemma 1.1.17tells us that πœ“ perserves distance up to a constant. Dividing so that the smallest nonzerodistance you see is at least 1, we get the same inequality in the image and the preimage.

Proof details. Let πœ€π‘– denote πœ€ with the 𝑖th coordinate flipped. We need to prove

E(𝑑𝑋(𝑓(πœ€), 𝑓(βˆ’πœ€))𝑝) ≀ 𝑇 𝑝𝑓

π‘›βˆ‘π‘–=1

E(𝑑𝑋(𝑓(πœ€), 𝑓(πœ€π‘–))𝑝)

Without loss of generality, by scaling 𝑓 we may assume that all the points 𝑓(πœ€) are distanceat least 1 apart. (𝑋 is a Banach space, so distance scales linearly; this doesn’t affect whetherthe inequality holds.)

Let πœ“ : 𝑋 β†’ π‘Œ be such that πœ“, πœ“βˆ’1 are uniform homeomorphisms. Because πœ“βˆ’1 isuniformly homeomorphic, there is 𝐢 such that π‘‘π‘Œ (𝑦1, 𝑦2) ≀ 1 implies 𝑑𝑋(πœ“

βˆ’1(𝑦1), πœ“βˆ’1(𝑦2)) <

𝐢. WLOG, by scaling 𝑓 we may assume that all the points 𝑓(πœ€) are max(1, 𝐢) apart, sothat the points πœ“ ∘ 𝑓(πœ€) are at least 1 apart.

We know that for any 𝑔 : {Β±1}𝑛 β†’ π‘Œ that

E(𝑑𝑋(𝑔(πœ€), 𝑔(βˆ’πœ€))𝑝) ≀ 𝑇 𝑝𝑔

π‘›βˆ‘π‘–=1

E(𝑑𝑋(𝑔(πœ€), 𝑔(πœ€π‘–))𝑝).

We apply this to 𝑔 = πœ“ ∘ 𝑓 ,𝑋

πœ“

οΏ½οΏ½

{Β±1}𝑛

𝑓;;

𝑔=πœ“βˆ˜π‘“##

π‘Œ

13

Page 14: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

to get

E(𝑑𝑋(𝑓(πœ€), 𝑓(βˆ’πœ€))𝑝) = E(𝑑𝑋(πœ“βˆ’1 ∘ 𝑔(πœ€), πœ“βˆ’1 ∘ 𝑔(βˆ’πœ€))𝑝)≀ πΏπœ“βˆ’1(1)E(π‘‘π‘Œ (𝑔(πœ€), 𝑔(βˆ’πœ€))𝑝)

≀ πΏπœ“βˆ’1(1)𝑇 𝑝𝑔

π‘›βˆ‘π‘–=1

E(π‘‘π‘Œ (𝑔(πœ€), 𝑔(πœ€π‘–)))

≀ πΏπœ“βˆ’1(1)πΏπœ“(1)𝑇𝑝𝑔

π‘›βˆ‘π‘–=1

E(𝑑𝑋(𝑔(πœ€), 𝑔(πœ€π‘–))𝑝)

as needed.

The parallelogram inequality for exponent 1 instead of 2 follows from using the triangleinequality on all possible paths for all paths of diagonals. Type 𝑝 > 1 is a strengthening ofthe triangle inequality. For which metric spaces does it hold?

What’s an example of a metric space where the inequality doesn’t hold with 𝑝 > 1? Thecube itself (with 𝐿1 distance).

𝑛𝑝 οΏ½ 𝑛.

I will prove to you that this is the only obstruction: given a metric space that doesn’t containbi-Lipschitz embeddings of arbitrary large cubes, the inequality holds.

We know an alternative inequality involving distance equivalent to type; I can prove it.It is, however, not a satisfactory solution to the Ribe program. There are other situationswhere we have complete success.

We will prove some things, then switch gears, slow down and discuss Grothendieck’sinequality and applications. They will come up in the nonlinear theory later.

2 Bourgain’s Theorem implies Ribe’s Theorem

2-3-16We will use the Corson-Klee Lemma 1.1.17.

Proof of Lemma 1.1.17. Suppose π‘₯, 𝑦 ∈ 𝑋, β€–π‘₯βˆ’ 𝑦‖ β‰₯ π‘Ž. Break up the line segment fromπ‘₯, 𝑦 into intervals of length π‘Ž; let π‘₯ = π‘₯0, π‘₯1, . . . , π‘₯π‘˜ = 𝑦 be the endpoints of those intervals,with

β€–π‘₯𝑖+1 βˆ’ π‘₯𝑖‖ ≀ π‘Ž.

The modulus of continuity is defined as

π‘Šπ‘“ (𝑑) = sup {‖𝑓(𝑒)βˆ’ 𝑓(𝑣)β€– : 𝑒, 𝑣 ∈ 𝑋, β€–π‘’βˆ’ 𝑣‖ ≀ 𝑑} .

Uniform continuity says lim𝑑→0π‘Šπ‘“ (𝑑) = 0. The number of intervals is

π‘˜ ≀ β€–π‘₯βˆ’ π‘¦β€–π‘Ž

+ 1 ≀ 2 β€–π‘₯βˆ’ π‘¦β€–π‘Ž

.

14

Page 15: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Then

‖𝑓(π‘₯)βˆ’ 𝑓(𝑦)β€– β‰€π‘˜βˆ‘π‘–=1

‖𝑓(π‘₯𝑖)βˆ’ 𝑓(π‘₯π‘–βˆ’1)β€–

≀ πΎπ‘Šπ‘“ (π‘Ž) ≀2π‘Šπ‘“ (π‘Ž)

π‘Žβ€–π‘₯βˆ’ 𝑦‖ ,

so we can let 𝐿(π‘Ž) =2π‘Šπ‘“ (π‘Ž)

π‘Ž.

2.1 Bourgain’s discretization theorem

There are 3 known proofs of Ribe’s Theorem.

1. Ribe’s original proof, 1987.

2. HK, 1990, a genuinely different proof.

3. Bourgain’s proof, a Fourier analytic proof which gives a quantitative version. This isthe version we’ll prove.

Bourgain uses the Discretization Theorem 1.2.4. There is an amazing open problem in thiscontext.

Saying 𝛿 is big says there is a not-too-fine net, which is enough. Therefore we areinterested in lower bounds on 𝛿.

Definition 1.2.1 (Discretization modulus): Let 𝑋 be a finite-dimensional normed spacedim(𝑋) = 𝑛 < ∞. Let the target space π‘Œ be an arbitrary Banach space. Consider theunit ball 𝐡𝑋 in 𝑋. Take a maximal 𝛿-net 𝒩𝛿 in 𝐡𝑋 . Suppose we can embed 𝒩𝛿 into π‘Œ via𝑓 : 𝒩𝛿 β†’ π‘Œ . Suppose we know in π‘Œ that

β€–π‘₯βˆ’ 𝑦‖ ≀ ‖𝑓(π‘₯)βˆ’ 𝑓(𝑦)β€– ≀ 𝐷 β€–π‘₯βˆ’ 𝑦‖ .

for all π‘₯, 𝑦 ∈ 𝑁𝛿. (We say that 𝒩𝛿 embeds with distortion 𝐷 into π‘Œ .)

You can prove using a nice compactness argument that if this holds for 𝛿 is small enough,then the entire space 𝑋 embeds into π‘Œ with rough the same distortion. Bourgain’s dis-cretization theorem 1.2.4 says that you can choose 𝛿 = 𝛿𝑛 to be independent of the geometryof 𝑋 and π‘Œ such that if you give a 𝛿-approximation of the unit-ball in the 𝑛-dimensionalnorm, you succeed in embedding the whole space.

I often use this theorem in this way: I use continuous methods to show embedding 𝑋into π‘Œ requires big distortion; immediately I get an example with a finite object. Let us nowmake the notion of distortion more precise.

Definition 1.2.2 (Distortion): Suppose (𝑋, 𝑑𝑋), (π‘Œ, π‘‘π‘Œ ) are metric spaces 𝐷 β‰₯ 1. We saythat 𝑋 embeds into π‘Œ with distortion 𝐷 if there exists 𝑓 : 𝑋 β†’ π‘Œ and 𝑠 > 0 such that forall π‘₯, 𝑦 ∈ 𝑋,

𝑆𝑑𝑋(π‘₯, 𝑦) ≀ π‘‘π‘Œ (𝑓(π‘₯), 𝑓(𝑦)) ≀ 𝐷𝑆𝑑𝑋(π‘₯, 𝑦).

15

Page 16: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

The infimum over those 𝐷 β‰₯ 1 such that 𝑋 embeds into π‘Œ with distortion is denotedπΆπ‘Œ (𝑋). This is a measure of how far 𝑋 is being from a subgeometry of π‘Œ .

Definition 1.2.3 (Discretization modulus): Let be a 𝑛-dimensional normed space andπ‘Œ be any Banach space, πœ€ ∈ (0, 1). Let 𝛿𝑋 β†’Λ“π‘Œ (πœ€) be the supremum over all those 𝛿 > 0 suchthat for every 𝛿-net 𝒩𝛿 in 𝐡𝑋 ,

πΆπ‘Œ (𝒩𝛿) β‰₯ (1βˆ’ πœ€)πΆπ‘Œ (𝑋).

Here 𝐡𝑋 := {π‘₯ ∈ 𝑋 : β€–π‘₯β€– ≀ 1}.

In other words, the distortion of the 𝛿-net is not much larger than the distortion of thewhole space. That is, the discrete 𝛿-ball encodes almost all information about the spacewhen it comes to embedding into π‘Œ : If you got πΆπ‘Œ (𝒩𝛿) to be small, then the distortion ofthe entire object is not much larger.

Theorem 1.2.4 (Bourgain’s discretization theorem). For every 𝑛, πœ€ ∈ (0, 1), for every 𝑋, π‘Œ ,dim𝑋 = 𝑛,

𝛿𝑋 β†’Λ“π‘Œ (πœ€) β‰₯ π‘’βˆ’(π‘›πœ€ )

𝐢𝑛

.

Moreover for 𝛿 = π‘’βˆ’(2𝑛)𝐢𝑛, we have πΆπ‘Œ (𝑋) ≀ 2πΆπ‘Œ (𝒩𝛿).

Thus there is a 𝛿 dependending on the dimension such that in any 𝑛-dimensional normspace, the unit ball it encodes all the information of embedding 𝑋 into anything else. It’sonly a function of the dimension, not of any of the relevant geometry.

The theorem says that if you look at a 𝛿-net in the unit ball, it encodes all the informationabout𝑋 when it comes to embedding into everything else. The amount you have to discretizeis just a function of the dimension, and not of any of the other relevant geometry.

Remark 1.2.5: The proof is via a linear operator. All the inequality says is that you canfind a function with the given distortion. The proof will actually give a linear operator.

The best known upper bound is

𝛿𝑋 β†’Λ“π‘Œ

(1

2

).

1

𝑛.

The latest progress was 1987, there isn’t a better bound yet. You have a month to thinkabout it before you get corrupted by Bourgain’s proof.

There is a better bound when the target space is a 𝐿𝑝 space.

Theorem 1.2.6 (Gladi, Naor, Shechtman). For every 𝑝 β‰₯ 1, if dim𝑋 = 𝑛,

𝛿𝑋 →˓𝐿𝑝(πœ€) &πœ€2

𝑛52

(We still don’t know what the right power is.) The case 𝑝 = 1 is important for appli-cations. There are larger classes for spaces where we can write down axioms for where thisholds. There are crazy Banach spaces which don’t belong to this class, so we’re not done.We need more tools to show this: Lipschitz extension theorems, etc.

16

Page 17: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

2.2 Bourgain’s Theorem implies Ribe’s Theorem

With the β€œmoreover,” Bourgain’s theorem implies Ribe’s Theorem 1.1.9.

Proof of Ribe’s Theorem 1.1.9 from Bourgain’s Theorem 1.2.4. Let 𝑋, π‘Œ be Banach spacesthat are uniformly homeomorphic. By Corson-Klee 1.1.17, there exists 𝑓 : 𝑋 β†’ π‘Œ such that

β€–π‘₯βˆ’ 𝑦‖ β‰₯ 1 =β‡’ β€–π‘₯βˆ’ 𝑦‖ ≀ ‖𝑓(π‘₯)βˆ’ 𝑓(𝑦)β€– ≀ 𝐾 β€–π‘₯βˆ’ 𝑦‖ .

(Apply the Corson-Klee lemma for both 𝑓 and the inverse.)In particular, if 𝑅 > 1 and 𝒩 is a 1-net in

𝑅𝐡𝑋 = {π‘₯ ∈ 𝑋 : β€–π‘₯β€– ≀ 𝑅} ,

then πΆπ‘Œ (𝒩 ) ≀ 𝐾. Equivalently, for every 𝛿 > 0 every 𝛿-net in 𝐡𝑋 satisfies πΆπ‘Œ (𝒩 ) ≀ 𝐾. If𝐹 βŠ† 𝑋 is a finite dimension subspace and 𝛿 = π‘’βˆ’(2 dim𝐹 )𝐢 dim𝐹

, then by the β€œmoreover” partof Bourgain’s Theorem 1.2.4, there exists a linear operator 𝑇 : 𝐹 β†’ π‘Œ such that

β€–π‘₯βˆ’ 𝑦‖ ≀ ‖𝑇π‘₯βˆ’ 𝑇𝑦‖ ≀ 2𝐾 β€–π‘₯βˆ’ 𝑦‖

for all π‘₯, 𝑦 ∈ 𝐹 . This means that 𝑋 is finitely representable.

The motivation for this program comes in passing from continuous to discrete. Thetheory has many applications, e.g. to computer science whcih cares about finite things. Iwould like an improvement in Bourgain’s Theorem 1.2.4.

First we’ll prove a theorem that has nothing to do with Ribe’s Theorem. There arelemmas we will be using later. It’s an easier theorem. It looks unrelated to metric theory,but the lemmas are relevant.

17

Page 18: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

18

Page 19: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Chapter 2

Restricted invertibility principle

1 Restricted invertibility principle

1.1 The first restricted invertibility principles

We take basic facts in linear algebra and make things quantitative. This is the lesson of thecourse: when you make things quantitative, new mathematics appears.

Proposition 2.1.1: If 𝐴 : Rπ‘š β†’ R𝑛 is a linear operator, then there exists a linear subspace𝑉 βŠ† R𝑛 with dim(𝑉 ) = rank(𝐴) such that 𝐴 : 𝑉 β†’ 𝐴(𝑉 ) is invertible.

What’s the quantitative question we want to ask about this? Invertibility just says thatan inverse exists. Can we find a large subspace where not only is 𝐴 invertible, but the inversehas small norm?

We insist that the subspace is a coordinate subspace. Let 𝑒1, . . . , π‘’π‘š be the standardbasis of Rπ‘š, 𝑒𝑗 = (0, . . . , 1⏟ ⏞

𝑗

, 0, . . .). The goal is to find a β€œlarge” subset 𝜎 βŠ† {1, . . . ,π‘š} such

that 𝐴 is invertible on R𝜎 where

R𝜎 := {(π‘₯1, . . . , π‘₯𝑛) ∈ Rπ‘š : π‘₯𝑖 = 0 if 𝑖 ∈ 𝜎}

and the norm of π΄βˆ’1 : 𝐴(R𝜎) β†’ R𝜎 is small.A priori this seems a crazy thing to do; take a small multiple of the identity. But we can

find conditions that allow us to achieve this goal.

Theorem 2.1.2 (Bourgain-Tzafriri restricted invertibility principle, 1987). thm:btrip Let 𝐴 :Rπ‘š β†’ Rπ‘š be a linear operator such that

‖𝐴𝑒𝑗‖2 = 1

for every 𝑗 ∈ {1, . . . ,π‘š}. Then there exist 𝜎 βŠ† {1, . . . ,π‘š} such that

1. |𝜎| β‰₯ π‘π‘šβ€–π΄β€–2 , where ‖𝐴‖ is the operator norm of 𝐴.

19

Page 20: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

2. 𝐴 is invertible on R𝜎 and the norm of π΄βˆ’1 : 𝐴(R𝜎) β†’ R𝜎 is at most 𝐢 β€² (i.e.,β€–π΄π½πœŽβ€–π‘†βˆž ≀ 𝐢 β€², to use the notation introduced below).

Here 𝑐, 𝐢 β€² are universal constants.

Suppose the biggest eigenvalue is at most 100. Then you can always find a coordinatesubset of proportional size such that on this subset, 𝐴 is invertible and the inverse has normbounded by a universal constant.

All of the proofs use something very amazing.This proof is from 3 weeks ago. This has been reproved many times. I’ll state a theorem

that gives better bound than the entire history.This was extended to rectangular matrices. (The extension is nontrivial.)Given 𝑉 βŠ† Rπ‘š a linear subspace with dim𝑉 = π‘˜ and 𝐴 : 𝑉 β†’ Rπ‘š a linear operator, the

singular values of 𝐴𝑠1(𝐴) β‰₯ 𝑠2(𝐴) β‰₯ Β· Β· Β· β‰₯ π‘ π‘˜(𝐴)

are the eigenvalues of (𝐴*𝐴)12 . We can decompose

𝐴 = π‘ˆπ·π‘‰

where 𝐷 is a matrix with 𝑠𝑖(𝐴)’s on the diagonal, and π‘ˆ, 𝑉 are unitary.

Definition 2.1.3: For 𝑝 β‰₯ 1 the Schatten-von Neumann 𝑝-norm of 𝐴 is

‖𝐴‖𝑆𝑝:=

(π‘˜βˆ‘π‘–=1

𝑠𝑖(𝐴)𝑝

) 1𝑝

= (Tr((𝐴*𝐴)𝑝2 ))

1𝑝

= (Tr((𝐴𝐴*)𝑝2 ))

1𝑝

The cases 𝑝 = ∞, 2 give the operator and Frobenius norm,

β€–π΄β€–π‘†βˆž= operator norm

‖𝐴‖𝑆2=ÈTr(𝐴*𝐴) =

(βˆ‘π‘Ž2𝑖𝑗� 1

2 .

Exercise 2.1.4: ‖·‖𝑆𝑝is a norm on β„³π‘›Γ—π‘š(R). You have to prove that given 𝐴,𝐡,

(Tr([(𝐴+𝐡)*(𝐴+𝐡)]𝑝2 ))

1𝑝 ≀ (Tr((𝐴*𝐴)

𝑝2 ))

1𝑝 + (Tr((𝐡*𝐡)

𝑝2 ))

1𝑝 .

This requires an idea. Note if 𝐴,𝐡 commute this is trivial. Apparently von Neumannwrote a paper called β€œMetric Spaces” in the 1930s in which he just proves this inequalityand doesn’t know what to do with it, so it got forgotten for a while until the 1950s, whenSchatten wrote books on applications. When I was a student in grad school, I was takinga class on random matrices. There was two weeks break, I was certain that it was trivialbecause the professor had not said it was not, and it completely ruined my break though Icame up with a different proof of it. It’s short, but not trivial: It’s not typical linear algebra!.This is like another triangle inequality, which we may need later on.

Spielman and Srivastava have a beautiful theorem.

20

Page 21: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Definition 2.1.5: Stable rank.Let 𝐴 : Rπ‘š β†’ R𝑛. The stable rank is defined as

srank(𝐴) =

(‖𝐴‖𝑆2

β€–π΄β€–π‘†βˆž

)2

.

The numerator is the sum of squares of the singular values, and the denominator is themaximal value. Large stable rank means that many singular values are nonzero, and infact large on average. Many people wanted to get the size of the subset in the RestrictedInvertibility Principle to be close to the stable rank.

Theorem 2.1.6 (Spielman-Srivastava). thm:ss For every linear operator 𝐴 : Rπ‘š β†’ R𝑛, πœ€ ∈(0, 1), there exists 𝜎 βŠ† {1, . . . ,π‘š} with |𝜎| β‰₯ (1βˆ’ πœ€)srank(𝐴) such that

(𝐴𝐽𝜎)βˆ’1π‘†βˆž

.

βˆšπ‘š

πœ€ ‖𝐴‖𝑆2

.

Here, 𝐽𝜎 is the identity function restricted to R𝜎, 𝐽 : R𝜎 β†’Λ“ Rπ‘š.

This is stronger than Bourgain-Tzafriri. In Bourgain-Tzafriri the columns were unitvectors.

Proof of Theorem 2.1.2 from Theorem 2.1.6. Let 𝐴 be as in Theorem 2.1.2. Then ‖𝐴‖𝑆2=È

Tr(𝐴*𝐴) =βˆšπ‘š and srank(𝐴) = π‘š

‖𝐴‖2π‘†βˆž. We obtain the existence of

|𝜎| β‰₯ (1βˆ’ πœ€)π‘š

‖𝐴‖2π‘†βˆž

with β€–(𝐴𝐽𝜎)βˆ’1β€–π‘†βˆž.

βˆšπ‘š=

1πœ€.

This is a sharp dependence on πœ€.The proof introduces algebraic rather than analytic methods; it was eye-opening. Marcus

even got sets bigger than the stable rank and looked at 𝑝𝑓‖𝐴‖𝑆2‖𝐴‖𝑆4

2, which is muchstronger.

1.2 A general restricted invertibility principle

I’ll show a theorem that implies all these intermediate theorems. We use (classical) analysisand geometry instead of algebra. What matters is not the ratio of the norms, but the tailof the distribution of 𝑠1(𝐴)

2, . . . , π‘ π‘š(𝐴)2.

Theorem 2.1.7. thm:gen-srank Let 𝐴 : Rπ‘š β†’ R𝑛 be a linear operator. If π‘˜ < rank(𝐴) then thereexist 𝜎 βŠ† {1, . . . ,π‘š} with |𝜎| = π‘˜ such that

(𝐴𝐽𝜎)βˆ’1π‘†βˆž

. minπ‘˜<π‘Ÿβ‰€rank(𝐴)

ΓŠπ‘šπ‘Ÿ

(π‘Ÿ βˆ’ π‘˜)βˆ‘π‘šπ‘–=π‘Ÿ 𝑠𝑖(𝐴)

2.

21

Page 22: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

You have to optimize over π‘Ÿ. You can get the ratio of 𝐿𝑝 norms from the tail bounds. Thisimplies all the known theorems in restricted invertibility. The subset can be as big as youwant up to the rank, and we have sharp control in the entire range. This theorem general-izes Spielman-Srivasta (Theorem 2.1.6), which had generalized Bourgain-Tzafriri (Theorem2.1.2). 2-8-16

Now we will go backwards a bit, and talk about a less general result. After Theorem 2.1.6,a subsequent theorem gave the same theorem but instead of the stable rank, used somethingbetter.

Theorem 2.1.8 (Marcus, Spielman, Srivastava). thm:mss4 If

π‘˜ <1

4

(‖𝐴‖𝑆2

‖𝐴‖𝑆4

)4

,

there exists 𝜎 βŠ† {1, . . . ,π‘š}, |𝜎| = π‘˜ such that(𝐴𝐽𝜎)

βˆ’1π‘†βˆž

.

βˆšπ‘š

‖𝐴‖𝑆2

.

A lot of these quotients of norms started popping up in people’s results. The correctgeneralization is the following notion.

Definition 2.1.9: For 𝑝 > 2, define the stable 𝑝th rank by

srank𝑝(𝐴) =

�‖𝐴‖𝑆2

‖𝐴‖𝑆𝑝

οΏ½ 2π‘π‘βˆ’2

.

Exercise 2.1.10: Show that if 𝑝 β‰₯ π‘ž > 2, then

srank𝑝(𝐴) ≀ srankπ‘ž(𝐴).

(Hint: Use Holder’s inequality.)

Now we would like to prove how Theorem 2.1.7 generalizes the previously listed results:

Proof of Generalizability of Theorem 2.1.7. Using Holder’s inequality with 𝑝2,

‖𝐴‖2𝑆2=

π‘šβˆ‘π‘—=1

𝑠𝑗(𝐴)2

=π‘Ÿβˆ’1βˆ‘π‘—=1

𝑠𝑗(𝐴)2 +

π‘šβˆ‘π‘—=π‘Ÿ

𝑠𝑗(𝐴)2

≀ (π‘Ÿ βˆ’ 1)1βˆ’2𝑝

οΏ½π‘Ÿβˆ’1βˆ‘π‘—=1

𝑠𝑗(𝐴)𝑝

οΏ½ 2𝑝

+π‘šβˆ‘π‘—=π‘Ÿ

𝑠𝑗(𝐴)2

22

Page 23: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

≀ (π‘Ÿ βˆ’ 1)1βˆ’2𝑝 ‖𝐴‖2𝑆𝑝

+π‘šβˆ‘π‘—=π‘Ÿ

𝑠𝑗(𝐴)2

π‘šβˆ‘π‘—=π‘Ÿ

𝑠𝑗(𝐴)2 β‰₯ ‖𝐴‖2𝑆2

οΏ½1βˆ’ (π‘Ÿ βˆ’ 1)βˆ’

2𝑝

‖𝐴‖2𝑆𝑝

‖𝐴‖2𝑆2

οΏ½= ‖𝐴‖2𝑆2

οΏ½1βˆ’

οΏ½π‘Ÿ βˆ’ 1

srank𝑝(𝐴)

οΏ½1βˆ’ 2𝑝

οΏ½Now we can plug the previous calculation into Theorem 2.1.7 to demonstrate the way thenew theorem generalizes the previous results:

(𝐴𝐽𝜎)βˆ’1. min

π‘˜+1β‰€π‘Ÿβ‰€rank(𝐴)

Γπ‘šπ‘Ÿ

(π‘Ÿ βˆ’ π‘˜) ‖𝐴‖2𝑆2

(1βˆ’

(π‘Ÿβˆ’1

srank𝑝(𝐴)

)1βˆ’ 2𝑝

)=

βˆšπ‘š

β€–π΄β€–βˆžmin

π‘˜+1β‰€π‘Ÿβ‰€rank(𝐴)

Γπ‘Ÿ

(π‘Ÿ βˆ’ π‘˜)(1βˆ’

(π‘Ÿβˆ’1

srank𝑝(𝐴)

)1βˆ’ 2𝑝

)This equation implies all the earlier theorems.

To optimize, fix the stable rank, differentiate in π‘Ÿ, and set to 0. All theorems in theliterature follow from this theorem; in particular, we get all the bounds we got before. Therewas nothing special about the number 4 in Theorem 2.1.8; this is about the distribution ofthe eigenvalues.

2 Ky Fan maximum principle

sec:kf We’ll be doing linear algebra. It’s mostly mechanical, except we’ll need this lemma.

Lemma 2.2.1 (Ky Fan maximum principle). lem:kf Suppose that 𝑃 : Rπ‘š β†’ Rπ‘š is a rank π‘˜orthogonal projection. Then

Tr(𝐴*𝐴𝑃 ) β‰€π‘˜βˆ‘π‘–=1

𝑠𝑖(𝐴)2

where 𝑠𝑖(𝐴) are the singular values, i.e., 𝑠𝑖(𝐴)2 are the eigenvalues of 𝐡 := 𝐴*𝐴.

This material was given in class on 2-15. This lemma follows from the following generaltheorem.

Theorem 2.2.2 (Convex function of dot products acheives maximum at eigenvectors).thm:eignmax Let 𝐡 : R𝑛 β†’ R𝑛 be symmetric, and let 𝑓 : R𝑛 β†’ R be a convex function. Then forevery orthonormal basis 𝑒1, . . . , 𝑒𝑛 ∈ R𝑛, there exists a permutation πœ‹ such that

eq:kf0𝑓(βŸ¨π΅π‘’1, 𝑒1⟩, . . . , βŸ¨π΅π‘’π‘›, π‘’π‘›βŸ©) ≀ 𝑓(πœ†πœ‹(1), . . . , πœ†πœ‹(𝑛)) (2.1)

23

Page 24: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Essentially, we’re saying using the eigenbasis maximizes the convex function.

Remark 2.2.3: We can weaken the condition on 𝑓 to the following: for every 𝑖 < 𝑗,𝑑 ↦→ 𝑓(π‘₯1, . . . , π‘₯𝑖 + 𝑑, π‘₯𝑖+1, . . . , π‘₯π‘—βˆ’1, π‘₯𝑗 βˆ’ 𝑑, π‘₯𝑗+1, . . . , π‘₯𝑛) is convex as a function of 𝑑. If 𝑓 issmooth, this is equivalent to the second derivative in 𝑑 being β‰₯ 0:

eq:kf1

πœ•2𝑓

πœ•π‘₯2𝑖+πœ•2𝑓

πœ•π‘₯2π‘—βˆ’ 2

πœ•2𝑓

πœ•π‘₯π‘–πœ•π‘₯𝑗β‰₯ 0 (2.2)

In other words, you just need that the Hessian is positive semidefinite. (Above, we wrotethe determinant of the Hessian on the π‘₯𝑖 βˆ’ π‘₯𝑗 plane. This being true for all pairs 𝑖, 𝑗 is thesame as the Hessian being positive definite.)

Proof. We may assume without loss of generality that

1. 𝑓 is smooth. If not, convolute with a good kernel.

2. Strict inequality holds:πœ•2𝑓

πœ•π‘₯2𝑖+πœ•2𝑓

πœ•π‘₯2π‘—βˆ’ 2

πœ•2𝑓

πœ•π‘₯π‘–πœ•π‘₯𝑗> 0.

To see this, note we can take any πœ€ > 0 and perturb 𝑓(π‘₯) into 𝑓(π‘₯) + πœ€β€–π‘₯β€–22. Theinequality to prove (2.1) is also perturbed by this slight change, and taking πœ€ β†’ 0gives the desired inequality.

Now let 𝑒1, . . . , 𝑒𝑛 be an orthonormal basis at which 𝑓(βŸ¨π΅π‘’1, 𝑒1⟩, . . . , βŸ¨π΅π‘’π‘›, π‘’π‘›βŸ©) attainsits maximum. Then for 𝑒𝑖, 𝑒𝑗, we want to rotate in the 𝑖 βˆ’ 𝑗 plane by angle πœƒ. Since 𝑒𝑖, 𝑒𝑗span a two dimensional subspace, recall the 2-dimensional rotation matrix. Let

π‘…πœƒ =

οΏ½cos(πœƒ) sin(πœƒ)sin(πœƒ) βˆ’ cos(πœƒ)

οΏ½;𝑒𝑖;𝑗 =

�𝑒𝑖𝑒𝑗

οΏ½Multiplying, we get

π‘…πœƒπ‘’π‘–;𝑗 =

οΏ½cos(πœƒ) sin(πœƒ)sin(πœƒ) βˆ’ cos(πœƒ)

οΏ½ �𝑒𝑖𝑒𝑗

οΏ½=

οΏ½cos(πœƒ)𝑒𝑖 + sin(πœƒ)𝑒𝑗sin(πœƒ)𝑒𝑖 βˆ’ cos(πœƒ)𝑒𝑗

οΏ½=

οΏ½(π‘…πœƒπ‘’π‘–;𝑗)1(π‘…πœƒπ‘’π‘–;𝑗)2

οΏ½Then, we replace 𝑓 with 𝑔(πœƒ) =

𝑓(βŸ¨π΅π‘’1, 𝑒1⟩, . . . , ⟨𝐡 (π‘…πœƒπ‘’π‘–;𝑗)1 , (π‘…πœƒπ‘’π‘–;𝑗)1⟩, ⟨𝐡 (π‘…πœƒπ‘’π‘–;𝑗)2 , (π‘…πœƒπ‘’π‘–;𝑗)2⟩, . . . , βŸ¨π΅π‘’π‘›, π‘’π‘›βŸ©

οΏ½where we keep all other dot products the same. By assumption, 𝑔 attains its maximum atπœƒ = 0, so 𝑔′(0) = 0, 𝑔′′(0) ≀ 0. Expanding out the rotated dot products explicitly in 𝑔(πœƒ), weget that the 𝑖th argument is

cos2(πœƒ)βŸ¨π΅π‘’π‘–, π‘’π‘–βŸ©+ sin2(πœƒ)βŸ¨π΅π‘’π‘—, π‘’π‘—βŸ©+ sin(2πœƒ)βŸ¨π΅π‘’π‘–, π‘’π‘—βŸ©

24

Page 25: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

and the 𝑗th argument is

sin2(πœƒ)βŸ¨π΅π‘’π‘–, π‘’π‘–βŸ©+ cos2(πœƒ)βŸ¨π΅π‘’π‘—, π‘’π‘—βŸ© βˆ’ sin(2πœƒ)βŸ¨π΅π‘’π‘–, π‘’π‘—βŸ©

Then we can mechanically take the derivatives at πœƒ = 0 to get

0 = 𝑔′(0) = 2βŸ¨π΅π‘’π‘–, π‘’π‘—βŸ©(𝑓π‘₯𝑖 βˆ’ 𝑓π‘₯𝑗)

0 β‰₯ 𝑔′′(0) = 2 (βŸ¨π΅π‘’π‘—, π‘’π‘—βŸ© βˆ’ βŸ¨π΅π‘’π‘–, π‘’π‘–βŸ©) (𝑓π‘₯𝑖 βˆ’ 𝑓π‘₯𝑗)⏟ ⏞ =0 if βŸ¨π΅π‘’π‘–,π‘’π‘—βŸ©=0

+4βŸ¨π΅π‘’π‘–, π‘’π‘—βŸ©2(𝑓π‘₯𝑖π‘₯𝑖 + 𝑓π‘₯𝑗π‘₯𝑗 βˆ’ 2𝑓π‘₯𝑖π‘₯𝑗

�⏟ ⏞ β‰₯0

.

This implies that for all 𝑖 = 𝑗 βŸ¨π΅π‘’π‘–, π‘’π‘—βŸ© = 0, which implies that for all 𝑖, 𝐡𝑒𝑖 = πœ‡π‘–π‘’π‘– for someπœ‡π‘–. Thus any function applied to a vector of dot products is maximized at eigenvalues.

Exercise 2.2.4: exr:kf If 𝑓 : R𝑛 β†’ R satisfies the conditions in Theorem 2.2.2 and (𝑒1, . . . , 𝑒𝑛),(𝑣1, . . . , 𝑣𝑛) are two orthonormal bases of R𝑛, then for every 𝐴 : R𝑛 β†’ R𝑛, there exists πœ‹ ∈ 𝑆𝑛,(πœ€1, . . . , πœ€π‘›) ∈ {Β±1}𝑛 such that

𝑓(βŸ¨π΄π‘’1, 𝑣1⟩, βŸ¨π΄π‘’2, 𝑣2⟩, . . . , βŸ¨π΄π‘’π‘›, π‘£π‘›βŸ©) ≀ 𝑓(πœ€1π‘ πœ‹(1)(𝐴), . . . , πœ€π‘›π‘ πœ‹(𝑛)(𝐴))

Show that choosing 𝑒, 𝑣 as the singular vectors maximizes 𝑓 (over all pairs of orthonormalbases).

To solve this problem, you can rotate both vectors in the same direction and take deriva-tives, and also rotate them in opposite directions and take derivatives to get enough infor-mation to prove that the singular values are the maximum.

Essentially, a lot of the inequalities you find in books follow from this. For instance, ifyou want to prove that the Schatten 𝑝-norm is a norm, it follows directly from this fact.

Corollary 2.2.5. Let β€– Β· β€– be a norm on R𝑛 that is invariant under premutations and sign:

β€–(π‘₯1, . . . , π‘₯𝑛)β€– = β€–(πœ€1π‘₯πœ‹(1), . . . , πœ€π‘›π‘₯πœ‹(𝑛))β€–

for all πœ€ ∈ {Β±1}𝑛 and πœ‹ ∈ 𝑆𝑛 (In the literature, we call this a symmetric norm). Thisinduces a norm on matrices π‘€π‘šΓ—π‘›(R) with

‖𝐴‖ = β€–(π‘ πœ‹(1)(𝐴), . . . , π‘ πœ‹(𝑛)(𝐴)β€–)

Then the triangle inequality holds for matrices 𝐴,𝐡:

‖𝐴+𝐡‖ ≀ ‖𝐴‖+ ‖𝐡‖.

Proof. We have by Exercise 2.2.4

‖𝐴+𝐡‖ = max(𝑒𝑖)βŠ₯,(𝑣𝑖)βŠ₯

β€–(⟨(𝐴+𝐡)𝑒𝑖, π‘£π‘–βŸ©)𝑛𝑖=1β€–

≀ max(𝑒𝑖)βŠ₯,(𝑣𝑖)βŠ₯

β€–(⟨(𝐴𝑒𝑖, π‘£π‘–βŸ©)𝑛𝑖=1β€–+ max(𝑒𝑖)βŠ₯,(𝑣𝑖)βŠ₯

β€–(βŸ¨π΅π‘’π‘–, π‘£π‘–βŸ©)𝑛𝑖=1β€–

≀ ‖𝐴‖+ ‖𝐡‖ .

25

Page 26: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Remember Theorem 2.2.2! For many results, you simply need to apply the right convexfunction to get the result.

Our lemma follows from setting 𝑓(π‘₯) =βˆ‘π‘˜π‘–=1 π‘₯𝑖.

Proof of Ky Fan Maximum Principle (Lemma 2.2.1). Take an orthonormal basis 𝑒1, . . . , 𝑒𝑛of 𝑃 such that 𝑒1, . . . , π‘’π‘˜ is a basis of the range of 𝑃 . Then

Tr(𝐡𝑃 ) =π‘˜βˆ‘π‘—=1

βŸ¨π΅π‘’π‘—, π‘’π‘—βŸ© β‰€π‘˜βˆ‘π‘–=1

𝑠𝑖(𝐡) =π‘˜βˆ‘π‘–=1

𝑠𝑖(𝐴)2

3 Finding big subsets

We’ll present 4 lemmas for finding big subsets with certain properties. We’ll put themtogether at the end.

3.1 Little Grothendieck inequality

Theorem 2.3.1 (Little Grothendieck inequality). thm:lgi Fix π‘˜,π‘š, 𝑛 ∈ N. Suppose that 𝑇 :Rπ‘š β†’ R𝑛 is a linear operator. Then for every π‘₯1, . . . , π‘₯π‘˜ ∈ Rπ‘š,

π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22 β‰€πœ‹

2‖𝑇‖2β„“π‘šβˆžβ†’β„“π‘›2

π‘˜βˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘–

for some 𝑖 ∈ {1, . . . ,π‘š} where π‘₯π‘Ÿ = (π‘₯π‘Ÿ1, . . . , π‘₯π‘Ÿπ‘š).

Later we will show πœ‹2is sharp.

If we had only 1 vector, what does this say?

‖𝑇π‘₯1β€–2 β‰€βˆšπœ‹

2β€–π‘‡β€–β„“π‘šβˆžβ†’β„“π‘›2

β€–π‘₯1β€–βˆž

We know the inequality is true for π‘˜ = 1 with constant 1, by definition of the operator norm.The theorem is true for arbitrary many vectors, losing an universal constant (πœ‹

2). After we

see the proof, the example where πœ‹2is attained will be natural.

We give Grothendieck’s original proof.The key claim is the following.

Claim 2.3.2. clm:lgi

eq:lgi1

π‘šβˆ‘π‘—=1

(π‘˜βˆ‘π‘Ÿ=1

(𝑇 *𝑇π‘₯π‘Ÿ)2𝑗

) 12

β‰€βˆšπœ‹

2β€–π‘‡β€–β„“π‘šβˆžβ†’β„“π‘›2

(π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–2) 1

2

. (2.3)

26

Page 27: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof of Theorem 2.3.1. Assuming Claim 2.3.2, we prove the theorem.

π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22 =π‘˜βˆ‘π‘Ÿ=1

βŸ¨π‘‡π‘₯π‘Ÿ, 𝑇π‘₯π‘ŸβŸ©

=π‘˜βˆ‘π‘Ÿ=1

⟨π‘₯π‘Ÿ, 𝑇 *𝑇π‘₯π‘ŸβŸ©

=π‘˜βˆ‘π‘Ÿ=1

π‘šβˆ‘π‘—=1

π‘₯π‘Ÿπ‘—(𝑇*𝑇π‘₯π‘Ÿ)𝑗

β‰€π‘šβˆ‘π‘—=1

(π‘˜βˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘—

) 12(

π‘˜βˆ‘π‘Ÿ=1

(𝑇 *𝑇π‘₯π‘Ÿ)2𝑗

) 12

by Cauchy-Schwarz

≀

οΏ½max1β‰€π‘—β‰€π‘š

(π‘˜βˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘—

) 12

οΏ½οΏ½π‘šβˆ‘π‘—=1

π‘˜βˆ‘π‘Ÿ=1

(𝑇 *𝑇π‘₯π‘Ÿ)2𝑗

οΏ½ 12

≀ max1β‰€π‘—β‰€π‘š

(π‘˜βˆ‘π‘–=1

π‘₯2𝑖𝑗

) 12 βˆšπœ‹

2β€–π‘‡β€–β„“π‘šβˆžβ†’β„“π‘›2

(π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22

) 12

π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22 β‰€πœ‹

2‖𝑇‖2β„“π‘šβˆžβ†’β„“π‘›2

max𝑗

π‘˜βˆ‘π‘Ÿ=1

π‘₯2𝑖𝑗.

We bounded by a square root of the multiple of the same term, a bootstrapping argument.In the last step, divide and square.

Proof of Claim 2.3.2. Let 𝑔1, . . . , π‘”π‘˜ be iid standard Gaussian random variables. For everyfixed 𝑗 ∈ {1, . . . ,π‘š},

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿ(𝑇*𝑇π‘₯π‘Ÿ)𝑗.

This is a Gaussian random variable with mean 0 and varianceβˆ‘π‘˜π‘Ÿ=1(𝑇

*𝑇π‘₯π‘Ÿ)2𝑗 . Taking the

expectation,1

Eπ‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿ(𝑇*𝑇π‘₯π‘Ÿ)𝑗

=

(π‘˜βˆ‘π‘Ÿ=1

(𝑇 *𝑇π‘₯π‘Ÿ)2𝑗

) 12Ê2

πœ‹.

Sum these over 𝑗:

E

⎑⎣ π‘šβˆ‘π‘—=1

𝑇 *(

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ)𝑗

⎀⎦ =

Ê2

πœ‹

π‘šβˆ‘π‘—=1

(π‘˜βˆ‘π‘Ÿ=1

(𝑇 *𝑇π‘₯π‘Ÿ)2𝑗

) 12

π‘šβˆ‘π‘—=1

(π‘˜βˆ‘π‘Ÿ=1

(𝑇 *𝑇π‘₯π‘Ÿ)2𝑗

) 12

=

βˆšπœ‹

2E

⎑⎣ π‘šβˆ‘π‘—=1

𝑇 *

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿ(𝑇π‘₯π‘Ÿ)𝑗

⎀⎦ .eq:lgi2 (2.4)

1È

12πœ‹

βˆ«βˆžβˆ’βˆž |π‘₯|π‘’βˆ’ π‘₯2

2 = βˆ’2È

12πœ‹ [𝑒

βˆ’ π‘₯2

2 ]∞0 =È

2πœ‹

27

Page 28: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Define a random sign vector 𝑧 ∈ {Β±1}π‘š by

𝑧𝑗 = sign

οΏ½(𝑇 *

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ

)𝑗

οΏ½Then

π‘šβˆ‘π‘—=1

(𝑇 *

π‘˜βˆ‘π‘Ÿ=1

𝑔𝑇π‘₯π‘Ÿ)𝑗

=

βŸ¨π‘§, 𝑇 *

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ

⟩=

βŸ¨π‘‡π‘§,

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ

βŸ©β‰€ ‖𝑇𝑧‖2

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ

2

≀ β€–π‘‡β€–β„“π‘šβˆžβ†’β„“π‘›2

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ

2

This is a pointwise inequality. Taking expectations and using Cauchy-Schwarz,

eq:lgi3E

⎑⎣ π‘šβˆ‘π‘—=1

(𝑇 *

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ

)𝑗

⎀⎦ ≀ β€–π‘‡β€–β„“π‘šβˆžβ†’β„“π‘›2

οΏ½Eπ‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿπ‘‡π‘₯π‘Ÿ

22

οΏ½ 12

. (2.5)

What is the second moment? Expand:

eq:lgi4Eπ‘˜βˆ‘π‘Ÿ=1

𝑔𝑖𝑇π‘₯π‘Ÿ

22

= E

βŽ‘βŽ£βˆ‘π‘–π‘—

𝑔𝑖𝑔𝑗 βŸ¨π‘‡π‘₯𝑖, 𝑇π‘₯π‘—βŸ©

⎀⎦ =π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22 . (2.6)

Chaining together (2.4), (2.5), (2.6) gives the result.

Why use the Gaussians? The identity characterizes the Gaussians using rotation invari-ance. Using other random variables gives other constants that are not sharp.

There will be lots of geometric lemmas:

βˆ™ A fact about restricting matrices.

βˆ™ Another geometric argument to give a different method for selecting subsets.

βˆ™ A combinatorial lemma for selecting subsets.

Finally we’ll put them together in a crazy induction.2-10-16: We were in the process of proving three or four subset selection principles, which

we will somehow use to prove the RIP.The little Grothendieck inequality (Theorem 2.3.1) is part of an amazing area of mathe-

matics with many applications. It’s little, but very useful. The proof is really Grothendieck’soriginal proof, re-organized. For completeness, we’ll show the fact that the inequality is sharp(cannot be improved).

28

Page 29: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

3.1.1 Tightness of Grothendieck’s inequality

Corollary 2.3.3.Γˆπœ‹/2 is the best constant in Theorem 2.3.1.

From the proof, we reverse engineer vectors that make the inequality sharp. They aregiven in the following example.

Example 2.3.4: Let 𝑔1, 𝑔2, . . . , π‘”π‘˜ be iid Gaussians on the probability space (Ξ©, 𝑃 ). Let𝑇 : 𝐿∞(Ξ©, 𝑃 ) β†’ β„“π‘˜2 be

𝑇𝑓 = (E[𝑓𝑔1], . . . ,E[π‘“π‘”π‘˜]).

Let π‘₯π‘Ÿ ∈ 𝐿∞(Ξ©, 𝑃 ),

π‘₯π‘Ÿ =π‘”π‘Ÿ(βˆ‘π‘˜

𝑖=1 𝑔2𝑖

οΏ½ 12

.

Proof. Le 𝑔1, . . . , π‘”π‘˜;𝑇 ;π‘₯1, . . . π‘₯π‘˜ be as in the example. Note the π‘₯π‘Ÿ are nothing more thanvectors on the π‘˜-dimensional unit sphere, so they are bounded functions on the measurespace Ξ©. We can also write

eq:sum-xr

π‘˜βˆ‘π‘Ÿ=1

π‘₯π‘Ÿ(πœ”)2 =

π‘˜βˆ‘π‘Ÿ=1

π‘”π‘Ÿ(πœ”)2βˆ‘π‘Ÿ

𝑖=1 𝑔𝑖(πœ”)2= 1 (2.7)

We use the Central Limit Theorem in order to replace the Ξ© by a discrete space. Let πœ€π‘Ÿ,𝑖 beΒ±1 random variables. Then letting

π‘”π‘Ÿ =πœ€π‘Ÿ,1 + . . .+ πœ€π‘Ÿ,π‘βˆš

𝑁

instead, we have that π‘”π‘Ÿ approaches a standard Gaussian in distribution, so the statementswe make will be asymptotically true. With this discretization, the random variables {π‘”π‘Ÿ}live in Ξ© = {Β±1}𝑁𝐾 . So 𝐿∞(Ξ©) = 𝑙2

𝑁𝐾

∞ , which is in a large but finite dimension. So πœ” willreally be a coordinate in Ξ©.

Now we show two things; they are nothing more than computations.

1. β€–π‘‡β€–πΏβˆž(Ξ©,P)β†’π‘™π‘˜2=È2/πœ‹,

2. We also showβˆ‘π‘˜π‘Ÿ=1 ‖𝑇π‘₯π‘Ÿβ€–22

π‘˜β†’βˆžβˆ’βˆ’βˆ’β†’ 1.

From (2.7) and the 2 items, the little Grothendieck inequality is sharp in the limit.

29

Page 30: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

For (1), we have

β€–π‘‡β€–β„“βˆžβ†’β„“2 = supβ€–π‘“β€–βˆžβ‰€1

(π‘˜βˆ‘π‘Ÿ=1

E [π‘“π‘”π‘Ÿ]2

)1/2

= supβ€–π‘“β€–βˆžβ‰€1supβˆ‘π‘˜

π‘Ÿ=1𝛼2π‘Ÿ=1

βˆ‘π‘Ÿ=1

π›Όπ‘ŸE [π‘“π‘”π‘Ÿ]

= supβˆ‘π‘˜

π‘Ÿ=1𝛼2π‘Ÿ=1

supβ€–π‘“β€–βˆžβ‰€1E[𝑓

π‘˜βˆ‘π‘–=1

π›Όπ‘Ÿπ‘”π‘Ÿ

]= supβˆ‘π‘˜

π‘Ÿ=1

Eπ‘˜βˆ‘π‘Ÿ=1

π›Όπ‘Ÿπ‘”π‘Ÿ

= E|𝑔1| =

Ê2

πœ‹

(2.8)

as we claimed, since ‖𝛼‖2 = 1 impliesβˆ‘π‘˜π‘Ÿ=1 π›Όπ‘Ÿπ‘”π‘Ÿ is also a gaussian.

Now for (2),

π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22 =π‘˜βˆ‘π‘Ÿ=1

οΏ½E

⎑⎣ 𝑔2π‘Ÿ(βˆ‘π‘˜π‘–=1 𝑔

2𝑖

�1/2⎀⎦�2

= 𝐾

οΏ½E

⎑⎣ 𝑔21(βˆ‘π‘˜π‘–=1 𝑔

2𝑖

�1/2⎀⎦�2

= 𝐾

οΏ½1

𝐾E

⎑⎣ π‘˜βˆ‘π‘Ÿ=1

𝑔2π‘Ÿ(βˆ‘π‘˜π‘–=1 𝑔

2𝑖

�1/2⎀⎦�2

=1

𝐾

οΏ½E

⎑⎣( π‘˜βˆ‘π‘–=1

𝑔2𝑖

)1/2⎀⎦�2

(2.9)

and you can use Stirling to finish. This is just a πœ’2-distribution.In this case E 𝑔1𝑔2

(βˆ‘

𝑖𝑔2𝑖 )

1/2 = E 𝑔1(βˆ’π‘”2)

(βˆ‘

𝑖𝑔2𝑖 )

1/2 . Also note that if (𝑔1, . . . , π‘”π‘˜) ∈ Rπ‘˜ is a standard

Gaussian, then (𝑔1,...,π‘”π‘˜)(βˆ‘π‘˜

𝑖=1𝑔2𝑖

οΏ½1/2 and(βˆ‘π‘˜

𝑖=1 𝑔2𝑖 )

1/2οΏ½are independent. In other words, the length

and angle are independent: This is just polar coordinates, you can check this.

Now, how does this relate to the Restricted Invertibility Problem?

3.2 Pietsch Domination Theorem

Theorem 2.3.5 (Pietsch Domination Theorem). thm:pdt Fix π‘š,𝑛 ∈ N and 𝑀 > 0. Supposethat 𝑇 : Rπ‘š β†’ R𝑛 is a linear operator such that for every π‘₯1, . . . , π‘₯π‘˜ ∈ Rπ‘š have

eq:pdt1

(π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22

)1/2

≀𝑀 max1β‰€π‘—β‰€π‘š

(π‘˜βˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘—

)1/2

(2.10)

30

Page 31: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Then there exist πœ‡ = (πœ‡1, . . . , πœ‡π‘š) ∈ Rπ‘š with πœ‡1 β‰₯ 0 andβˆ‘π‘šπ‘–=1 πœ‡π‘– = 1 such that for every

π‘₯ ∈ Rπ‘š

eq:pdt2‖𝑇π‘₯β€–2 ≀𝑀

(π‘€βˆ‘π‘–=1

πœ‡π‘– β€–π‘₯𝑖‖2)1/2

(2.11)

The theorem says that you can come up with a probability measure such that the normof T as an operator as a standard norm from π‘™βˆž to 𝑙2 (?), is bounded by 𝑀 .

Remark 2.3.6: The theorem really an iff: (2.11) is a stronger statement than (2.10), andin fact they are equivalent.

Proof. Define 𝐾 βŠ† Rπ‘š with

𝐾 =

{𝑦 ∈ Rπ‘š : 𝑦𝑖 =

π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22 βˆ’π‘€2π‘šβˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘– for some π‘˜, π‘₯1, . . . , π‘₯π‘˜ ∈ Rπ‘š}

Basically we cleverly select a convex set. Every 𝑛-tuple of vectors in Rπ‘š gives you a newvector in Rπ‘š. Let’s check that 𝐾 is convex. We have to check if two vectors 𝑦, 𝑧 ∈ 𝐾,then all points on the line between them are in 𝐾. 𝑦, 𝑧 ∈ 𝐾 means that there exist (π‘₯𝑖)

π‘˜π‘–=1,

(𝑀𝑖)𝑙𝑖=1,

𝑦𝑖 =π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22 βˆ’π‘€2π‘šβˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘–

𝑧𝑖 =π‘™βˆ‘

π‘Ÿ=1

β€–π‘‡π‘€π‘Ÿβ€–22 βˆ’π‘€2π‘™βˆ‘

π‘Ÿ=1

𝑀2π‘Ÿπ‘–

for all 𝑖. Then 𝛼𝑦𝑖 + (1 βˆ’ 𝛼)𝑧𝑖 comes from (βˆšπ›Όπ‘₯1, . . . ,

βˆšπ›Όπ‘₯π‘˜,

√1βˆ’ 𝛼𝑀1, . . .

√1βˆ’ π›Όπ‘€π‘˜). So

by design, 𝐾 is a convex set.Now, the assumption of the theorem says that(

π‘˜βˆ‘π‘Ÿ=1

‖𝑇π‘₯π‘Ÿβ€–22

)1/2

≀𝑀max1β‰€π‘—β‰€π‘š

(π‘˜βˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘—

)1/2

which implies

‖𝑇π‘₯π‘Ÿβ€–22 βˆ’π‘€2max1β‰€π‘—β‰€π‘šπ‘šβˆ‘π‘Ÿ=1

π‘₯2π‘Ÿπ‘— ≀ 0

which implies 𝐾 ∩ (0,∞)π‘š = βˆ…. By the hyperplane separation theorem (for two disjointconvex sets in Rπ‘š with at least one compact, there is a hyperplane between them), thereexists 0 = πœ‡ = (πœ‡1, . . . , πœ‡π‘š) ∈ Rπ‘š with

βŸ¨πœ‡, π‘¦βŸ© ≀ βŸ¨πœ‡, π‘§βŸ©

for all 𝑦 ∈ 𝐾 and 𝑧 ∈ (0,∞)π‘š. By renormalizing, we may assumeβˆ‘π‘šπ‘–=1 πœ‡π‘– = 1. Moreover πœ‡

cannot have any strictly negative coordinate: Otherwise you could take 𝑧 to have arbitrarily

31

Page 32: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

large value at a strictly negative coordinate with zeros everywhere else, implying βŸ¨π‘’, π‘§βŸ© is nolonger bounded from below, a contradiction. Therefore, πœ‡ is a probability vector and βŸ¨πœ‡, π‘§βŸ©can be arbitrarily small. So for every 𝑦 ∈ 𝐾,

βˆ‘π‘šπ‘–=1 πœ‡π‘–π‘¦π‘– ≀ 0. Write

𝑦𝑖 = ‖𝑇π‘₯β€–22 βˆ’π‘€2 β€–π‘₯𝑖‖2 ∈ 𝐾.

Expanding this out,

‖𝑇π‘₯β€–22 βˆ’π‘€2π‘›βˆ‘π‘–=1

πœ‡π‘– β€–π‘₯𝑖‖2 ≀ 0,

which is exactly what we wanted.

3.3 A projection bound

Lemma 2.3.7. lem:projbound π‘š,𝑛 ∈ N, πœ€ ∈ (0, 1), 𝑇 : R𝑛 β†’ Rπ‘š a linear operator. ThenβˆƒπœŽ βŠ‚ {1, . . . ,π‘š} with |𝜎| β‰₯ (1βˆ’ πœ€)π‘š such that

β€–ProjRπœŽπ‘‡β€–π‘†βˆž β‰€βˆš

πœ‹

2πœ€π‘šβ€–π‘‡β€–π‘™π‘›2β†’π‘™π‘š1

We will find ways to restrict a matrix to a big submatrix. We won’t be able to controlits operator norm, but we will be able to control the norm from 𝑙𝑛2 to π‘™π‘š1 . Then we pass toa further subset, which this becomes an operator norm on, which is an improvement whichGrothendieck gave us. This is the first very useful tool to start finding big submatrices.

Proof. We have 𝑇 : 𝑙𝑛2 β†’ π‘™π‘š1 , 𝑇* : π‘™π‘šβˆž β†’ 𝑙𝑛2 . Now some abstract nonsense gives us that for Ba-

nach spaces, the norm of an operator and its adjoint are equal, i.e. ‖𝑇‖𝑙𝑛2β†’π‘™π‘š1= ‖𝑇 *β€–π‘™π‘šβˆžβ†’π‘™π‘›2

.This statement follows from the Hahn-Banach theorem (come see me if you haven’t seenthis before, I’ll tell you what book to read). From the Little Grothendieck inequality(Theorem 2.3.1), 𝑇 * satisfies the assumption of the Pietsch domination theorem 2.3.5 with𝑀 =

Γˆπœ‹2‖𝑇‖𝑙𝑛2β†’π‘™π‘š1

(we’re applying it to 𝑇 *). By the theorem, there exists a probabilityvector (πœ‡1, . . . , πœ‡π‘š) such that for every 𝑦 ∈ Rπ‘š,

‖𝑇 *𝑦‖2 =𝑀

(π‘šβˆ‘π‘–=1

πœ‡π‘–π‘¦2𝑖

)1/2

with 𝑀 =È

πœ‹2‖𝑇‖𝑙𝑛2β†’π‘™π‘š1

. Define 𝜎 =βŒ‹π‘– ∈ {1, . . . ,π‘š} : πœ‡π‘– ≀ 1

π‘šπœ€

{; then |𝜎| β‰₯ (1 βˆ’ πœ€)π‘š by

Markov’s inequality. We can also see this by writing

1 =π‘šβˆ‘π‘–=1

πœ‡π‘– =βˆ‘π‘–βˆˆπœŽ

πœ‡π‘– +βˆ‘π‘– ∈𝜎

πœ‡π‘– >βˆ‘π‘–βˆˆπœŽ

πœ‡π‘– +π‘šβˆ’ |𝜎|π‘šπœ€

which follows since for 𝑗 ∈ 𝜎, πœ‡π‘— >1π‘šπœ€

. Continuing,

π‘šπœ€βˆ’π‘š+ |𝜎|π‘šπœ€

β‰₯βˆ‘π‘–βˆˆπœŽ

πœ‡π‘–

|𝜎| β‰₯ (π‘šπœ€)βˆ‘π‘–βˆˆπœŽ

πœ‡π‘– +π‘š(1βˆ’ πœ€)

32

Page 33: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Then, because πœ‡ is a probability distribution, (π‘šπœ€)βˆ‘π‘–βˆˆπœŽ πœ‡π‘– β‰₯ 0 and we have

|𝜎| β‰₯ π‘š(1βˆ’ πœ€)

Now take π‘₯ ∈ R𝑛 and choose 𝑦 ∈ Rπ‘š with ‖𝑦‖2 = 1. Then

βŸ¨π‘¦,ProjRπœŽπ‘‡π‘₯⟩2 = βŸ¨π‘‡ *ProjRπœŽπ‘¦, π‘₯⟩2 ≀ ‖𝑇 *ProjRπœŽπ‘¦β€–22 Β· β€–π‘₯β€–22

≀ πœ‹

2‖𝑇‖𝑙𝑛2β†’π‘™π‘š1

(βˆ‘π‘–βˆˆπœŽ

πœ‡π‘–π‘¦2𝑖

)β€–π‘₯β€–22 ≀

πœ‹

2‖𝑇‖2𝑙𝑛2β†’π‘™π‘š1

1

π‘šπœ€β€–π‘₯β€–22

by Cauchy-Schwarz. Taking square roots gives the desired result.

In the previous proof, we used a lot of duality to get an interesting subset.

Remark 2.3.8: In Lemma 2.3.7, I think that either the constant πœ‹/2 is sharp (no subsetare bigger; it could come from the Gaussians), or there is a different constant here. If theconstant is 1, I think you can optimize the previous argument and get the constant to bearbitrarily close to 1, which would have some nice applications: In other words, getting

Γˆπœ‹

2πœ€π‘š

as close to 1 as possible would be good. I didn’t check before class, but you might wantto check if you can carry out this argument using the Gaussian argument we made for thesharpness of πœ‹

2in Grothendieck’s inequality (Theorem 2.3.1). It’s also possible that there is

a different universal constant.

3.4 Sauer-Shelah Lemma

Now we will give another lemma which is very easy and which we will use a lot.

Lemma 2.3.9 (Sauer-Shelah). lem:saushel Take integers π‘š,𝑛 ∈ N and suppose that we have alarge set Ξ© βŠ† {Β±1}𝑛 with

|Ξ©| >π‘šβˆ’1βˆ‘π‘˜=0

(𝑛

π‘˜

)Then βˆƒπœŽ βŠ† {1, . . . , 𝑛} such that with |𝜎| = π‘š, if you project onto R𝜎 the set of vectors, you getthe entire cube: ProjR𝜎(Ξ©) = {Β±1}𝜎.2 For every πœ€ ∈ {Β±1}𝜎, there exists 𝛿 = (𝛿1, . . . , 𝛿𝑛) ∈ Ξ©such that 𝛿𝑗 = πœ€π‘— for 𝑗 ∈ 𝜎.

Note that Lemma 2.3.9 is used in the proof of the Fundamental Theorem of StatisticalLearning Theory.

Proof. We want to prove by induction on 𝑛. First denote the shattering set

sh(Ξ©) = {𝜎 βŠ† {1, . . . , 𝑛} : ProjR𝜎Ω = {Β±1}𝜎}2I.e., the VC dimension of Ξ© is β‰₯ π‘š.

33

Page 34: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

The claim is that the number of sets shattered by a given set is |sh(Ξ©)| β‰₯ |Ξ©|. The emptyset case is trivial. What happens when 𝑛 = 1? Ξ© βŠ‚ {βˆ’1, 1}, and thus the set is shattered.Assume that our claim holds for 𝑛, and now set Ξ© βŠ† {Β±1}𝑛+1 = {Β±1}𝑛 Γ— {Β±1}. Define

Ξ©+ = {πœ” ∈ {Β±1}𝑛 : (πœ”, 1) ∈ Ξ©}

Ξ©βˆ’ = {πœ” ∈ {Β±1}𝑛 : (πœ”,βˆ’1) ∈ Ξ©}

Then, letting Ξ©+ = {(πœ”, 1) ∈ {Β±1}𝑛+1 : πœ” ∈ Ξ©+} and Ξ©βˆ’ similarly, we have |Ξ©| = |Ξ©+| +|Ξ©βˆ’| = |Ξ©+|+ |Ξ©βˆ’|. By our inductive step, we have sh(Ξ©+) β‰₯ |Ξ©+| and sh(Ξ©βˆ’) β‰₯ |Ξ©βˆ’|. Notethat any subset that shatters Ξ©+ also shatters Ξ©, and likewise for Ξ©βˆ’. Note that if a set Ξ©β€²

shatters both of them, we are allowed to add on an extra coordinate to get Ξ©β€²Γ—{Β±1} whichshatters Ξ©. Therefore,

sh(Ξ©+) βˆͺ sh(Ξ©βˆ’) βˆͺ {𝜎 βˆͺ {𝑛+ 1} : 𝜎 ∈ sh(Ξ©+) ∩ sh(Ξ©βˆ’)} βŠ† sh(Ξ©)

where the last union is disjoint since the dimensions are different. Therefore, we can nowuse this set inclusion to complete the induction using the principle of inclusion-exclusion:

|sh(Ξ©)| β‰₯ |sh(Ξ©+) βˆͺ sh(Ξ©βˆ’)|+ |sh(Ξ©+) ∩ sh(Ξ©βˆ’)| (disjoint sets)

= |sh(Ξ©+)|+ |sh(Ξ©βˆ’)| βˆ’ |sh(Ξ©+) ∩ sh(Ξ©βˆ’)|+ |sh(Ξ©+) ∩ sh(Ξ©βˆ’)|= |sh(Ξ©+)|+ |sh(Ξ©βˆ’)|β‰₯ |Ξ©+|+ |Ξ©βˆ’| = |Ξ©|

which completes the induction as desired.

We will primarily use the theorem as the following corollary, which says that if you havehalf of the points in terms of cardinality, you get half of the dimension.

Corollary 2.3.10. If |Ξ©| β‰₯ 2π‘›βˆ’1 then there exists 𝜎 βŠ† {1, . . . , 𝑛} with |𝜎| β‰₯ βŒˆπ‘›+12βŒ‰ β‰₯ 𝑛

2such

that ProjR𝜎Ω = {±1}𝜎.

2-15-16: Last time we left off with the proof of the Sauer-Shelah lemma. To remind you,we were finding ways to find interesting subsets where matrices behave well. Now recall wehad a linear algebraic fact which I owe you; I will prove it in an analytic way. The proof hasbeen moved to Section 2.

4 Proof of RIP

4.1 Step 1

Now we need another geometric lemma for the proof of Theorem 2.1.7, the restricted invert-ibility principle.

34

Page 35: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Lemma 2.4.1 (Step 1). lem:step1 Fix π‘š,𝑛, π‘Ÿ ∈ N. Let 𝐴 : Rπ‘š β†’ R𝑛 be a linear operator withrank(𝐴) β‰₯ π‘Ÿ. For every 𝜏 βŠ† {1, . . . ,π‘š}, denote

𝐸𝜏 = (span((𝐴𝑒𝑗)π‘—βˆˆπœ ))βŠ₯ .

Then there exists 𝜏 βŠ† {1, . . . ,π‘š} with |𝜏 | = π‘Ÿ such that for all 𝑗 ∈ 𝜏 ,

β€–ProjπΈπœβˆ–{𝑗}𝐴𝑒𝑗‖2 β‰₯

1βˆšπ‘š

(π‘šβˆ‘π‘–=1

𝑠𝑖(𝐴)2

)1/2

.

Basically we’re taking the projection of the π‘—π‘‘β„Ž column onto the orthogonal completementof the span of the subspace of all columns in the set except for the π‘—π‘‘β„Ž one, and bounding thenorm of that by a dimension term and the square root of the sum of the eigenvalues. (This issharp asymptotically, and may in fact even be sharp as written tooβ€”I need to check. Checkthis?)

Proof. For every 𝜏 βŠ† {1, . . . ,π‘š}, denote

𝐾𝜏 = conv ({±𝐴𝑒𝑗}π‘—βˆˆπœ )

The idea is to make the convex hull have big volume. Once we do that, wewill get all theseinequalities for free. Let 𝜏 βŠ† {1, . . . ,π‘š} be the subset of size π‘Ÿ that maximizes volπ‘Ÿ(𝐾𝜏 ).We know that volπ‘Ÿ(𝐾𝜏 ) > 0. Observe that for any 𝛽 βŠ† {1, . . . ,π‘š} of size π‘Ÿ βˆ’ 1 and 𝑖 ∈ 𝛽,we have

𝐾𝛽βˆͺ{𝑖} = conv (𝐾𝛽 βˆͺ {±𝐴𝑒𝑖}) ,

which is a double cone.

What is the height of this cone? It is β€–Proj𝐸𝛽𝐴𝑒𝑖‖2, as 𝐸𝛽 is the orthogonal complement

of the space spanned by 𝛽. Therefore, the π‘Ÿ-dimensional volume is given by

volπ‘Ÿ(𝐾𝛽βˆͺ{𝑖}) = 2 Β·volπ‘Ÿβˆ’1(𝐾𝛽) Β· β€–Proj𝐸𝛽

𝐴𝑒𝑖‖2π‘Ÿ

35

Page 36: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Because |𝜏 | = π‘Ÿ is the maximizing subset of 𝐹Ω, for any 𝑗 ∈ 𝜏 and 𝑖 ∈ {1, . . . ,π‘š}, choosing𝛽 = 𝜏 βˆ– {𝑗}, we get

volπ‘Ÿ(𝐾𝛽βˆͺ{𝑗}) β‰₯ volπ‘Ÿ(𝐾𝛽βˆͺ{𝑖})

=β‡’ β€–ProjπΈπœβˆ–{𝑗}𝐴𝑒𝑗‖2 β‰₯ β€–ProjπΈπœβˆ–{𝑗}

𝐴𝑒𝑖‖2.

for every 𝑗 ∈ 𝜏 and 𝑖 ∈ {1, . . . ,π‘š}. Summing,

π‘šβ€–ProjπΈπœβˆ–{𝑗}𝐴𝑒𝑗‖22 β‰₯

π‘šβˆ‘π‘–=1

β€–ProjπΈπœβˆ–{𝑗}𝐴𝑒𝑖‖22 = β€–ProjπΈπœβˆ–{𝑗}

𝐴‖2𝑆2.

Then, for all 𝑗 ∈ 𝜏 ,

eq:rip-s1-1β€–ProjπΈπœβˆ–{𝑗}𝐴𝑒𝑗‖2 β‰₯

1βˆšπ‘šβ€–ProjπΈπœβˆ–{𝑗}

𝐴‖𝑆2 (2.12)

Let’s denote 𝑃 = ProjπΈπœβˆ–{𝑗}. Note 𝑃 is an orthogonal projection of rank π‘Ÿ βˆ’ 1. Then,

eq:rip-s1-2

‖𝑃𝐴‖2𝑆2= Tr((𝑃𝐴)*(𝑃𝐴)) = Tr(𝐴*𝑃 *𝑃𝐴) = Tr(𝐴*𝑃𝐴) = Tr(𝐴𝐴*𝑃 )

= Tr(𝐴𝐴*)βˆ’ Tr(𝐴𝐴*(𝐼 βˆ’ 𝑃 )) β‰₯π‘šβˆ‘π‘–=1

𝑠𝑖(𝐴)2 βˆ’

π‘Ÿβˆ’1βˆ‘π‘–=1

𝑠𝑖(𝐴)2 =

π‘šβˆ‘π‘–=π‘Ÿ

𝑠𝑖(𝐴)2

(2.13)

using the Ky Fan maximal principle 2.2.1, since 𝐼 βˆ’ 𝑃 is a projection of rank π‘šβˆ’ π‘Ÿ + 1.Putting (2.12) and (2.13) together gives the result.

4.2 Step 2

In our proof of the restricted invertibility principle, this is the first step. Before proving it,let me just tell you what the second step looks like.

Lemma 2.4.2 (Step 2). lem:step2 Let π‘˜,π‘š, 𝑛 ∈ N, 𝐴 : Rπ‘š β†’ R𝑛, rank(𝐴) > π‘˜. Let πœ” βŠ†{1, . . . ,π‘š} with |πœ”| = rank(𝐴) such that {𝐴𝑒𝑗}π‘—βˆˆπœ” are linearly independent. Denote forevery 𝑗 ∈ Ξ©

𝐹𝑗 = πΈπœ”βˆ–{𝑗} =(span(𝐴𝑒𝑖)π‘–βˆˆπœ”βˆ–{𝑗}

οΏ½.

Then there exists 𝜎 βŠ† πœ” with |𝜎| = π‘˜ such that

β€– (𝐴𝐽𝜎)βˆ’1 β€–π‘†βˆž .

Èrank(𝐴)È

rank(𝐴)βˆ’ π‘˜Β·maxπ‘—βˆˆπœ”

1

β€–Proj𝐹𝑗𝐴𝑒𝑗‖

Most of the work is in the second step. First we pass to a subset where we have someinformation about the shortest possible orthogonal project. But Step 1 saves us by boundingwhat this can be. Here we use the Grothendieck inequality, Sauer-Shelah, etc. Everything:It’s simple, but it kills the restricted invertibility principle.

36

Page 37: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof of Theorem 2.1.7 given Step 1 and 2. Take 𝐴 : Rπ‘š β†’ R𝑛. By Step 1 (Lemma 2.4.1),we can find subset 𝜏 βŠ† {1, . . . ,π‘š} with |𝜏 | = π‘Ÿ such that for all 𝑗 ∈ 𝜏 ,

eq:step1β€–ProjπΈπœβˆ–{𝑗}𝐴𝑒𝑗‖2 β‰₯

1βˆšπ‘š

(π‘šβˆ‘π‘–=π‘Ÿ

𝑠𝑖(𝐴)2

)1/2

. (2.14)

Now we apply Step 2 (Lemma 2.4.2) to 𝐴𝐽𝜏 , using πœ” = 𝜏 , and find a further subset 𝜎 βŠ† 𝜏such that

β€– (𝐴𝐽𝜎)βˆ’1 β€–π‘†βˆž ≀ minπ‘˜<π‘Ÿ<rank(𝐴)

Ìrank(𝐴)

rank(𝐴)βˆ’ π‘Ÿmaxπ‘—βˆˆπœ”

1Proj𝐹𝑗

𝐴𝑒𝑗

≀ minπ‘˜<π‘Ÿ<rank(𝐴)

ΓŠπ‘šπ‘Ÿ

(π‘Ÿ βˆ’ π‘˜)βˆ‘π‘šπ‘–=π‘Ÿ 𝑠𝑖(𝐴)

2

which we get by plugging directly in π‘Ÿ for the rank and using Step 1 (2.14) to get thedenominator.

2-17Now we prove Step 2 (Lemma 2.4.2). Note we can assume πœ” = {1, . . . ,π‘š} and that the

rank is π‘š.First we need some lemmas.

Lemma 2.4.3. lem:rip-step2-1 Let 𝐴 : Rπ‘š β†’ R𝑛 be such that {𝐴𝑒𝑗}π‘šπ‘—=1 are linearly independent,and 𝜎 βŠ† {1, . . . ,π‘š}, 𝑑 ∈ N. Then there exists 𝜏 βŠ† 𝜎 with

|𝜏 | β‰₯(1βˆ’ 1

2𝑑

)|𝜎|,

such that denoting πœƒ = 𝜏βˆͺ({1, . . . ,π‘š}βˆ–πœŽ), 𝐹𝑗 = (span({𝐴𝑒𝑗}𝑖 =𝑗))βŠ₯, and𝑀 = maxπ‘—βˆˆπœ”1

β€–Proj𝐹𝑗𝐴𝑒𝑗‖ ,

we have that for all 𝜎 ∈ Rπœƒ, βˆ‘π‘–βˆˆπœ

|π‘Žπ‘–| ≀ 2𝑑2𝑀

È|𝜎|βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

.

37

Page 38: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

This is proved by a nice inductive argument.

Proof. TODO next time

Lemma 2.4.4. lem:rip-step2-2 Let π‘š,𝑛, 𝑑 ∈ N and 𝛽 βŠ† {1, . . . ,π‘š}. Let 𝐴 : Rπ‘š β†’ R𝑛 be alinear operator such that {𝐴𝑒𝑗}π‘šπ‘—=1 are linearly independent. Then there exist two subsets

𝜎 βŠ† 𝜏 βŠ† 𝛽 such that |𝜏 | β‰₯(1βˆ’ 1

2𝑑

οΏ½|𝛽|, |πœβˆ–πœŽ| ≀ |𝛽|

4, and if we denote πœƒ = 𝜏 βˆͺ ({1, . . . ,π‘š}βˆ–π›½),

𝑀 = maxπ‘—βˆˆπœ”1

β€–Proj𝐹𝑗𝐴𝑒𝑗‖ , then

ProjR𝜎(π΄π½πœƒ)βˆ’1π‘†βˆž

. 2𝑑2𝑀.

Proof of Lemma 2.4.4 from Lemma 2.4.3. Apply Lemma 2.4.3 with 𝜎 = 𝛽. Basically, we’regoing to inductively construct a subset 𝜎 = 𝛽 of {1, Β· Β· Β· ,π‘š} onto which to project, so thatwe can control the 𝑙1 norm of the subset 𝜏 by the 𝑙2 norm on a slightly greater set πœƒ viaLemma 2.4.3.

We find 𝜏 βŠ† 𝛽 with |𝜏 | β‰₯(1βˆ’ 1

2𝑑

οΏ½|𝛽| such thatβˆ‘

π‘–βˆˆπœ|π‘Žπ‘–| ≀ 2

𝑑2𝑀

È|𝛽|

βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

.

Rewriting gives that

βˆ€π‘Ž ∈ R𝜎, β€–ProjR𝜏 π‘Žβ€– ≀ 2𝑑2𝑀

È|𝛽| β€–π΄π½πœƒπ‘Žβ€–2

=β‡’ProjR𝜏 (π΄π½πœƒ)

βˆ’1β„“πœƒ2β†’β„“πœ1

. 2𝑑2𝑀

È|𝛽|.

Denote πœ€ = |𝛽|4|𝜏 | . By Lemma 2.3.7, there exists 𝜎 βŠ† 𝜏 , |𝜎| β‰₯ (1βˆ’ πœ€)|𝜏 | such thatProjR𝜎(π΄π½πœƒ)

βˆ’1π‘†βˆž

=ProjR𝜎 ProjR𝜏 (π΄π½πœƒ)

βˆ’1π‘†βˆž

.Ê

πœ‹

2πœ€|𝜏 |2

𝑑2𝑀

È|𝛽|

. 2𝑑2𝑀.

38

Page 39: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Now, we have basically finished making use of Lemma 2.4.3 in the proof of Lemma 2.4.4.It remains to use Lemma 2.4.4 to prove the full second step in our proof of the RestrictedInvertibility Principle.

Proof of Lemma 2.4.2 (Step 2). Fix an integer π‘Ÿ such that 122π‘Ÿ+1 ≀ 1 βˆ’ π‘˜

π‘šβ‰€ 1

22π‘Ÿ. Proceed

inductively as follows. First set

𝜏0 = {1, . . . ,π‘š}𝜎0 = πœ‘.

Suppose 𝑒 ∈ {0, . . . , π‘Ÿ + 1} and we constructed πœŽπ‘˜, πœπ‘˜ βŠ† {1, . . . ,π‘š} such that if we denote𝛽𝑒 = πœπ‘’βˆ–πœŽπ‘’, πœƒπ‘’ = πœπ‘’ βˆͺ ({1, . . . ,π‘š}βˆ–π›½π‘’βˆ’1), then

1. πœŽπ‘’ βŠ† πœπ‘’ βŠ† π›½π‘’βˆ’1

2. |πœπ‘’| β‰₯(1βˆ’ 1

22π‘Ÿβˆ’π‘’+4

οΏ½|π›½π‘’βˆ’1|

3. |𝛽𝑒| ≀ 14|π›½π‘’βˆ’1|

4. β€–ProjRπœŽπ‘’ (π΄π½πœƒπ‘’)βˆ’1β€–π‘†βˆž

. 2π‘Ÿβˆ’π‘’2𝑀 .

Let 𝐻 = 2π‘Ÿ βˆ’ 𝑒+ 4. For instance, |𝜏1| β‰₯(1βˆ’ 1

22π‘Ÿ+3

οΏ½|𝛽0|. What is the new 𝛽?

For the inductive step, apply Lemma 2.4.4 on π›½π‘’βˆ’1 with 𝑑 = 2π‘Ÿβˆ’ 𝑒+ 4 to get πœŽπ‘’ βŠ† πœπ‘’ βŠ†π›½π‘’βˆ’1 such that |πœπ‘’| β‰₯

(1βˆ’ 1

22π‘Ÿβˆ’π‘’+4

οΏ½|π›½π‘’βˆ’1|, |πœπ‘’βˆ–πœŽπ‘’| ≀ |π›½π‘’βˆ’1|

4As we induct, we are essentially

building up more πœŽπ‘’ to eventually produce the invertible subset over which we will project.Note that the size of the πœƒ set is decreasing as we proceed inductively.

eq:rip-s2-1

ProjRπœŽπ‘’ (π΄π½πœƒπ‘’)

βˆ’1. 2π‘Ÿβˆ’π‘’/2𝑀. (2.15)

39

Page 40: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

We know |π›½π‘’βˆ’1| ≀ π‘š4π‘’βˆ’1 ,

π›½π‘’βˆ’1 = 𝛽𝑒 βŠ” πœŽπ‘’ βŠ” (π›½π‘’βˆ’1βˆ–πœπ‘’)|π›½π‘’βˆ’1| = |𝛽𝑒|+ |πœŽπ‘’|+ (|π›½π‘’βˆ’1| βˆ’ |πœπ‘’|)|πœŽπ‘’| = |π›½π‘’βˆ’1| βˆ’ |𝛽𝑒| βˆ’ (|π›½π‘’βˆ’1| βˆ’ |πœπ‘’|)

β‰₯ |π›½π‘’βˆ’1| βˆ’ |𝛽𝑒| βˆ’|π›½π‘’βˆ’1|22π‘Ÿβˆ’π‘’+4

β‰₯ |π›½π‘’βˆ’1| βˆ’ |𝛽𝑒| βˆ’π‘š

22π‘Ÿ+𝑒+2.

Our choice for the invertible subset is

𝜎 =π‘Ÿ+1⨆𝑒=1

πœŽπ‘’.

Telescoping gives

|𝜎| =π‘Ÿ+1βˆ‘π‘’=1

|πœŽπ‘’| β‰₯ |𝛽0| βˆ’ |π›½π‘Ÿ+1| βˆ’π‘š

22π‘Ÿ+2

βˆžβˆ‘π‘’=1

1

2𝑒

β‰₯ π‘šβˆ’ π‘š

4π‘Ÿ+1βˆ’ π‘š

22π‘Ÿ+2

= π‘š(1βˆ’ 1

22π‘Ÿ+1

)β‰₯ π‘š

π‘˜

π‘š= π‘˜.

Observe that 𝜎 βŠ† β‹‚π‘Ÿ+1𝑒=1 πœƒπ‘’ and for every 𝑒,

πœŽπ‘’, . . . , πœŽπ‘Ÿ+1 βŠ† πœπ‘’

𝜎1, . . . , πœŽπ‘’βˆ’1 βŠ† {1, . . . ,π‘š}βˆ–π›½π‘’βˆ’1

This allows us to use the conclusion for all the πœŽπ‘’β€™s at once.For π‘Ž ∈ R𝜎, π½πœŽπ‘Ž βŠ† π½πœƒπ‘’Rπœƒπ‘’ ,

ProjRπœŽπ‘’ (π΄π½πœƒπ‘’)βˆ’1(𝐴𝐽𝜎)π‘Ž = ProjRπœŽπ‘’ π½πœŽπ‘Ž.

since π΄βˆ’1𝐴 = 𝐼 and projecting a subset onto its containing set is just the subset itself. Then,breaking π½πœŽπ‘Ž into orthogonal components,

β€–π½πœŽπ‘Žβ€–22 =π‘Ÿ+1βˆ‘π‘’=1

β€–ProjRπœŽπ‘’ π½πœŽπ‘Žβ€–22

=π‘Ÿ+1βˆ‘π‘’=1

ProjRπœŽπ‘’ (π΄π½πœŽπ‘’)

βˆ’1(𝐴𝐽𝜎)π‘Ž22

.π‘Ÿ+1βˆ‘π‘’=1

22π‘Ÿβˆ’π‘’π‘€2 β€–π΄π½πœŽπ‘Žβ€–22 by (2.15)

. 22π‘Ÿπ‘€2 β€–π΄π½πœŽπ‘Žβ€–22

40

Page 41: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

≀ 𝑀2

1βˆ’ π‘˜π‘š

β€–π΄π½πœŽπ‘Žβ€–22

β€–π½πœŽπ‘Žβ€–2 β‰€ΓŠ

π‘š

π‘šβˆ’ π‘˜π‘€ β€–π΄π½πœŽπ‘Žβ€–2

Since this is true for all π‘Ž ∈ R𝜎,

=β‡’(𝐴𝐽𝜎)

βˆ’1π‘†βˆž

.Ê

π‘š

π‘šβˆ’ π‘˜π‘€.

An important aspect of this whole proof to realize is the way we took a first subset togive one bound, and then took another subset in order to apply the Little Grothendieckinequality. In other words, we first took a subset where we could control 𝑙1 to 𝑙2, and thentook a further subset to get the operator norm bounds.

Also, notice that in this approach, we construct our β€œoptimal subset” of the space in-ductively: However, doing this in general is inefficient. It would be nice to construct the setgreedily instead for algorithmic reasons, if we wanted to come up with a constructive proof.I have a feeling that if we were to avoid this kind of induction, it would involve not usingSauer-Shelah at all.

How can we make this theorem algorithmic?The way the Pietsch Domination Theorem 2.3.5 worked was by duality. We look at a

certain explicitly defined convex set. We found a separating hyperplane which must be aprobability measure. Then we had a probabilistic construction. This part is fine.

The bottleneck for making this an algorithm (I do believe this will become an algorithm)consists of 2 parts:

1. Sauer-Shelah lemma 2.3.9: We have some cloud of points in the boolean cube, andwe know there is some large subset of coordinates (half of them) such that when youproject to it you see the full cube. I’m quite certain that it’s NP-hard in general. (Workis necessary to formulate the correct algorithmic Sauer-Shelah lemma. How is the setgiven to you?). In fact, formulating the right question for an algorithmi Sauer-Shelahis the biggest difficulty.

We only need to answer the algorithmic question tailored to our sets, which havea simple description: the intersection of an ellipsoid with a cube. There is a goodchance that there is a polynomial time algorithm in this case. This question has otherapplications as well (perhaps one could generalize to the intersection of the cube withother convex bodies). 3

2. The second bottleneck is finding a subset with maximal volume.

It’s another place where we chose a subset, the subset that maximizes the volume outof all subsets of a given size (Lemma 2.4.1). Specifically, we want the set of columns

3There could be an algorithm out there, but I couldn’t find one in the literature.

41

Page 42: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

of a matrix that maximizes the volume of the convex hull. Computing the volume ofthe convex hull requires some thought. Also, there are

(π‘›π‘Ÿ

οΏ½subsets of size π‘Ÿ; if π‘Ÿ = 𝑛/2

there are exponentially many. We need a way to find subsets with maximum volumefast. There might be a replacement algorithm which approximately maximizes thisvolume.

2-22: We are at the final lemma, Lemma 2.4.3 in the Restricted Invertiblity Theorem.

The proof of this is an inductive application of the Sauer-Shelah lemma. A very importantidea comes from Giannopoulos. If you naively try to use Sauer-Shelah, it won’t work out.We will give a stronger statement of the previous lemma which we can prove by induction.

Lemma 2.4.5 (Stronger version of Lemma 2.4.3). lem:SS-induct-stronger Take π‘š,𝑛 ∈ N, 𝐴 : Rπ‘š β†’R𝑛 a linear operator such that {𝐴𝑒𝑗}π‘šπ‘—=1 are linearly independent. Suppose that π‘˜ β‰₯ 0 is aninteger and 𝜎 βŠ† {1, . . . ,π‘š}. Then there exists 𝜏 βŠ† 𝜎 with |𝜏 | β‰₯ (1 βˆ’ 1

2π‘˜)|𝜎| such that for

every πœƒ βŠ‡ 𝜏 for all π‘Ž ∈ Rπ‘š we have

eq:rip21-1

βˆ‘π‘–βˆˆπœ

|π‘Žπ‘–| β‰€π‘€Γˆ|𝜎|(

π‘˜βˆ‘π‘Ÿ=1

2π‘Ÿ/2) βˆ‘

π‘–βˆˆπœƒπ‘Žπ‘–π΄π‘’π‘–

2

+ (2π‘˜ βˆ’ 1)βˆ‘

π‘–βˆˆπœƒβˆ©(πœŽβˆ–πœ)|π‘Žπ‘–| (2.16)

42

Page 43: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Lemma 2.4.3 (all we need to complete the proof of the Restricted Invertibility Principle)is the case where πœƒ = 𝜏 βˆͺ {1, Β· Β· Β· ,π‘š} βˆ– 𝜎, 𝑑 = π‘˜.

We prove the stronger version via induction on π‘˜.

Proof. As π‘˜ becomes bigger, we’re allowed to take more of the set 𝜎. The idea is to useSauer-Shelah, taking half of 𝜎, and then half of what remains at each step.

For π‘˜ = 0, the statement is vacuous, because we can take 𝜏 to be the empty set. Byinduction, assume that the statement holds for π‘˜: we have found 𝜏 βŠ† 𝜎 such that |𝜏 | β‰₯(1βˆ’ 1

2π‘˜)|𝜎| and (2.16) holds for every 𝜏 βŠ† πœƒ. If 𝜎 = 𝜏 already, then 𝜏 satisfies (2.16) for π‘˜+1

as well, so WLOG |𝜎 βˆ– 𝜏 | > 0. Now define 𝑣𝑗 is the projection

𝑣𝑗 =Proj𝐹𝑗

𝐴𝑒𝑗Proj𝐹𝑗

𝐴𝑒𝑗22

Then βŸ¨π‘£π‘–, π΄π‘’π‘—βŸ© = 𝛿𝑖𝑗 by definition since (𝑣𝑖) is a dual basis for the 𝐴𝑒𝑗’s.Now we want to user Sauer-Shelah so we’re going to define a certain subset of the cube.

Define

Ξ© =

βŽ§βŽ¨βŽ©πœ– ∈ {Β±1}πœŽβˆ–πœ :

βˆ‘π‘–βˆˆπœŽβˆ–πœ πœ–π‘–π‘£π‘–

2

β‰€π‘€Γˆ2|𝜎 βˆ– 𝜏 |

⎫⎬⎭which is an ellipsoid intersected with the cube (it is not a sphere since the 𝑣𝑖s are notorthogonal). Then we have

𝑀2|𝜎 βˆ– 𝜏 | β‰₯βˆ‘π‘–βˆˆπœŽβˆ–πœ

1

β€–Proj𝐹𝑗𝐴𝑒𝑗‖2

=βˆ‘π‘—βˆˆπœŽβˆ–πœ

‖𝑣𝑗‖22

=1

2|πœŽβˆ–πœ |βˆ‘

πœ–βˆˆ{Β±1}πœŽβˆ–πœ

βˆ‘π‘—βˆˆπœŽβˆ–πœ πœ–π‘—π‘£π‘—

2

2

.

where the last step is true for any vectors (sum the squares and the pairwise correlationsdisappear).

Using Markov’s inequality this is

β‰₯ 1

2|πœŽβˆ–πœ |(2|πœŽβˆ–πœ | βˆ’ |Ξ©|

�𝑀22|𝜎 βˆ– 𝜏 |

which gives|Ξ©| > 2|πœŽβˆ–πœ |βˆ’1

Then by Sauer-Shelah (Lemma 2.3.9, there exists 𝛽 βŠ† 𝜎 βˆ– 𝜏 such that

ProjR𝛽 Ξ© = {Β±1}𝛽

and

|𝛽| β‰₯ 1

2|𝜎 βˆ– 𝜏 |

43

Page 44: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Define 𝜏 * = 𝜏 βˆͺ 𝛽. We will show that 𝜏 * satisfies the inductive hypothesis with π‘˜ + 1.Each time we find a certain set of coordinates to add to what we have before. |𝜏 *| is thecorrect size because

|𝜏 *| = |𝜏 |+ |𝛽| β‰₯ |𝜏 |+ |𝜎| βˆ’ |𝜏 |2

=|𝜏 |+ |𝜎|

2β‰₯(1βˆ’ 1

2π‘˜+1

)|𝜎|

where we used that |𝜏 | β‰₯(1βˆ’ 1

2π‘˜

�|𝜎|. So at least 𝜏 * is the right size.

Now, suppose πœƒ βŠ‡ 𝜏 *. For every π‘Ž ∈ Rπ‘š, we claim there exists some πœ– ∈ Ξ© such thatβˆ€π‘— ∈ 𝛽 such that πœ–π‘— = sign(π‘Žπ‘—). For any 𝛽, we can find some vector in the cube that has thesign pattern of our given vector π‘Ž. What does being in Ξ© mean? It means that at least thedual basis is small there. πœ– ∈ Ξ© says that

eq:step21-2β€–βˆ‘π‘–βˆˆπœŽβˆ–πœ

πœ–π‘–π‘£π‘–β€–2 β‰€π‘€Γˆ2|𝜎 βˆ– 𝜏 | ≀

π‘€Γˆ2|𝜎|

2π‘˜/2(2.17)

That was how we chose our ellipsoid.

βˆ‘π‘–βˆˆπ›½

|π‘Žπ‘–| =

βˆ‘π‘–βˆˆπ›½

π‘Žπ‘–π΄π‘’π‘–,βˆ‘π‘–βˆˆπœŽβˆ–πœ

πœ–π‘£π‘–

βŒ‹because the 𝑣𝑖’s are a dual basis so βŸ¨π΄π‘’π‘–, π‘£π‘—βŸ© = 𝛿𝑖𝑗, and 𝛽 βŠ† πœŽβˆ–πœ . We only know the πœ–π‘– arethe signs when you’re inside 𝛽. This equals

=

βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–,βˆ‘π‘–βˆˆπœŽβˆ–πœ

πœ–π‘£π‘–

βŒ‹βˆ’

βˆ‘π‘–βˆˆ(πœƒβˆ–π›½)∩(πœŽβˆ–πœ)

πœ–π‘–π‘Žπ‘–

Note that (πœƒ βˆ– 𝛽) ∩ (𝜎 βˆ– 𝜏) = πœƒ ∩ (𝜎 βˆ– 𝜏 *). In this set we can’t control the signs πœ–π‘–. ByCauchy-Schwarz and (2.17), this is

β‰€βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

Β·

βˆ‘π‘–βˆˆπœŽβˆ–πœ πœ–π‘£π‘–

2

+βˆ‘

π‘–βˆˆπœƒβˆ©(πœŽβˆ–πœ*)|π‘Žπ‘–|

β‰€βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

Β·π‘€Γˆ2|𝜎|

2π‘˜/2+

βˆ‘π‘–βˆˆπœƒβˆ©(πœŽβˆ–πœ*)

|π‘Žπ‘–|.

Because the conclusion of Sauer-Shelah told us nothing about the signs of πœ–π‘– outside 𝛽, sowe just take the worst possible thing.

Summarizing, βˆ‘π‘–βˆˆπ›½

|π‘Žπ‘–| β‰€π‘€Γˆ2|𝜎|

2π‘˜/2

βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

+βˆ‘

π‘–βˆˆπœƒβˆ–(πœŽβˆ–πœ*)|π‘Žπ‘–|

44

Page 45: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Using the inductive step,βˆ‘π‘–βˆˆπœ*

|π‘Žπ‘–| =βˆ‘π‘–βˆˆπœ

|π‘Žπ‘–|+βˆ‘π‘–βˆˆπ›½

|π‘Žπ‘–|

β‰€π‘€Γˆ|𝜎|π›Όπ‘˜

βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

+(2π‘˜ βˆ’ 1

οΏ½ βˆ‘π‘–βˆˆπœƒβˆ©(πœŽβˆ–πœ)

|π‘Žπ‘–|+βˆ‘π‘–βˆˆπ›½

|π‘Žπ‘–|

= π›Όπ‘˜Γˆ|𝜎|βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

+(2π‘˜ βˆ’ 1

οΏ½ βˆ‘π‘–βˆˆπœƒβˆ©(πœŽβˆ–πœ*)

|π‘Žπ‘–|+ 2π‘˜βˆ‘π‘–βˆˆπ›½

|π‘Žπ‘–|

In the last step, we moved 𝛽 from the second to the third term. Now use what we got beforefor the bound on

βˆ‘π‘–βˆˆπ›½ |π‘Žπ‘–| and plug it in to get

≀(π›Όπ‘˜ + 2(π‘˜+1)/2

�È|𝜎|βˆ‘π‘–βˆˆπœƒ

π‘Žπ‘–π΄π‘’π‘–

2

+(2π‘˜+1 βˆ’ 1

οΏ½ βˆ‘π‘–βˆˆπœƒβˆ©(πœŽβˆ–πœ*)

|π‘Žπ‘–|

which is exactly the inductive hypothesis.4

Remark 2.4.6 (Algorithmic Sauer-Shelah): We only used Sauer-Shelah 2.3.9 for intersect-ing cubes with an ellipsoid, so to make RIP algorithmic, we need an algorithm for findingthese particular intersections. Moreover, the ellipsoids are big is because we ask that the2-norm is at most

√2 times the expectation. An efficient algorithm probably exists.

Afterwards, wecould ask for higher dimensional shapes. I’ve seen some references thatworked for Sauer-Shelah when sets were of a special form, namely of size π‘œ(𝑛). This is some-thing more geometric. I don’t think there’s literature about Sauer-Shelah for intersectionof surfaces with small degree. This is a tiny motivation to do it, but it’s still interestingindependently.

4I looked through the original Gianpopoulous paper, and it was clear he tried out many many thingsto find which inductive hypothesis makes everything go through cleanly. You want to bound an 𝑙1 sumfrom above, so you want to use duality, and then use Sauer-Shelah to get signs such that the norm of thedual-basis is small.

45

Page 46: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

46

Page 47: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Chapter 3

Bourgain’s Discretization Theorem

1 Outline

We will prove Bourgain’s Discretization Theorem. This will take maybe two weeks, and hasmany interesting ingredients along the way. By doing this, we will also prove Ribe’s theorem,which is what we stated at the beginning.

Let’s remind ourselves of the definition.

Definition (Definition 1.2.3, discretization modulus): (𝑋, β€– Β· ‖𝑋), (π‘Œ, β€– Β· β€–π‘Œ ) are Banachspaces. Let πœ– ∈ (0, 1). Then 𝛿𝑋 β†’Λ“π‘Œ (πœ–) is the supremum over 𝛿 > 0 such that for every 𝛿-net𝒩𝛿 of the unit ball of 𝑋, the distortion

πΆπ‘Œ (𝑋) ≀ πΆπ‘Œ (𝒩𝛿)

1βˆ’ πœ–

πΆπ‘Œ is smallest bi-Lipschitz distortion by which you can embed 𝑋 into π‘Œ . There are ideasrequired to get 1 βˆ’ πœ–. For example, for πœ– = 1

2, the equation says is that if we succeed in

embedding a 𝛿-net into π‘Œ , then we succeeded in the full space with a distortion twice asmuch. A priori it’s not even clear that there exists such a 𝛿. There’s a nontrivial compactnessargument needed to prove its existence. We will just prove bounds on it assuming it exists.

Now, Bourgain’s discretization theorem says

Theorem (Bourgain’s discretization theorem 1.2.4). If dim(𝑋) = 𝑛, dim(π‘Œ ) = ∞, then

𝛿𝑋 β†’Λ“π‘Œ (πœ–) β‰₯ π‘’βˆ’(π‘›πœ– )

𝐢𝑛

for 𝐢 a universal constant.

Remark 3.1.1: It doesn’t matter what the maps are for the 𝒩𝛿, the proof will give a linearmapping and we won’t end up needing them. Assuming linearity in the definition is notnecessary.

Rademacher’s theorem says that any mapping R𝑛 β†’ R𝑛 is differentiable almost every-where with bi-Lipschitz derivative. Yu can extend this to 𝑛 = ∞, but you need some

47

Page 48: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

additional properties: π‘Œ must be the dual space, and the limit needs to be in the weak-*topology. In this topology there exists a sequence of norm-1 vectors that tend to 0. Almosteverywhere, this doesn’t happen. The Principle of Local Reflexivity says π‘Œ ** and π‘Œ arenot the same for infinite dimensional π‘Œ . The double dual of all sequences which tend to 0is 𝐿∞, a bigger space, but the difference between these never appears in finite dimensionalphenomena.

1.1 Reduction to smooth norm

From now on, 𝐡𝑋 = {π‘₯ ∈ 𝑋 : ‖𝑋‖𝑋 ≀ 1} denotes the ball, and 𝑆𝑋 = πœ•π΅ = {π‘₯ ∈ 𝑋 : ‖𝑋‖𝑋 = 1}denotes the boundary.

Later on we will be differentiating things without thinking about it, so we first prove thatwe can assume ‖·‖𝑋 is smooth on 𝑋 βˆ– {0}.

Lemma 3.1.2. For all 𝛿 ∈ (0, 1) there exists some 𝛿-net of 𝑆𝑋 with |𝒩𝛿| ≀(1 + 2

𝛿

�𝑛.

Proof. Let 𝒩𝛿 βŠ† 𝑆𝑋 be maximal with respect to inclusion such that β€–π‘₯βˆ’ 𝑦‖𝑋 > 𝛿 for everydistinct π‘₯, 𝑦 ∈ 𝒩𝛿. Maximality means that if 𝑧 ∈ 𝑆𝑋 , there exists π‘₯ ∈ 𝒩𝛿, β€–π‘₯ βˆ’ 𝑦‖𝑋 ≀ 𝛿(otherwise 𝑁𝛿 βˆͺ {𝑧} is a bigger 𝛿-net). The balls {π‘₯ + 𝛿

2𝐡𝑋}π‘₯βˆˆπ’©π›Ώ

are pairwise disjoint andcontained in 1 + 𝛿

2𝐡𝑋 . Thus

vol

οΏ½οΏ½1 +

𝛿

2

�𝐡𝑋

οΏ½β‰₯

βˆ‘π‘‹βˆˆπ‘€π›Ώ

vol

οΏ½π‘₯+

𝛿

2𝐡𝑋

οΏ½οΏ½1 +

𝛿

2

�𝑛vol(𝐡𝑋) = |𝒩𝛿|

�𝛿

2

�𝑛vol(𝐡𝑋)

1

Reduction to smooth norm. Let 𝒩𝛿 be a 𝛿-net of 𝑆𝑋 with |𝒩𝛿| = 𝑁 ≀ (1 + 2𝛿)𝑛. Given

𝑧 ∈ 𝒩𝛿, by Hahn-Banach we can choose 𝑧* such that

1. βŸ¨π‘§, 𝑧*⟩ = 1

2. ‖𝑧*‖𝑋* = 1.

Let π‘˜ be an integer such that 𝑁1/(2π‘˜) ≀ 1 + 𝛿. Then define

β€–π‘₯β€– :=

οΏ½βˆ‘π‘§βˆˆπ’©π›Ώ

βŸ¨π‘§*, π‘₯⟩2π‘˜οΏ½1/(2π‘˜)

Each term satisfies |βŸ¨π‘§*, π‘₯⟩| ≀ β€–π‘₯‖𝑋 , so we know

β€–π‘₯β€– ≀ 𝑁1/2π‘˜β€–π‘₯‖𝑋 ≀ (1 + 𝛿)β€–π‘₯‖𝑋 .1There are non-sharp bounds for the smallest size of a 𝛿-net. There is a lot of literature about the relations

between these things, but we just need an upper bound.

48

Page 49: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

For the lower bound, if π‘₯ ∈ 𝑆𝑋 , then choose 𝑧 ∈ 𝒩𝛿 such that β€–π‘₯ βˆ’ 𝑧‖𝑋 ≀ 𝛿. Then1βˆ’ βŸ¨π‘§*, π‘₯⟩ = βŸ¨π‘§*, 𝑧 βˆ’ π‘₯⟩ ≀ ‖𝑧 βˆ’ π‘₯β€– ≀ 𝛿, so βŸ¨π‘§*, π‘₯⟩ β‰₯ 1βˆ’ 𝛿, giving

β€–π‘₯β€– β‰₯ (1βˆ’ 𝛿) β€–π‘₯‖𝑋 .

Thus any norm is up to 1+𝛿 some equal to a smooth norm. 𝛿 was arbitrary, so without loss ofgenerality we can assume in the proof of Bourgain embedding that the norm is smooth.

The strategy to prove Bourgain’s discretization theorem is as follows. Given a 𝛿-net 𝒩𝛿 βŠ†π΅π‘‹ with 𝛿 ≀ π‘’βˆ’(𝑛/πœ–)𝐢𝑛

, and we know βˆƒ 𝑓 : 𝒩𝛿 β†’ π‘Œ such that 1𝐷‖π‘₯βˆ’π‘¦β€–π‘‹ ≀ ‖𝑓(π‘₯)βˆ’π‘“(𝑦)β€–π‘Œ ≀

β€–π‘₯βˆ’ 𝑦‖𝑋 , i.e., we can embed with distortion 𝐷. Our goal is to show that if 𝛿 ≀ π‘’βˆ’π·πΆπ‘›

thenthis implies that there exists 𝑇 : 𝑋 β†’ π‘Œ linear operator invertible ‖𝑇‖ β€–π‘‡βˆ’1β€– . 𝐷.

The steps are as follows.

1. Find the correct coordinate system (John ellipsoid), which will give us a dot prod-uct structure and a natural Laplacian. (We will need a little background in convexgeometry. )

2. A priori 𝑓 is defined on the net. Extend 𝑓 to the whole space in a nice way (Bourgain’sextension theorem) which doesn’t coincide with the function on the net, but is not toofar away from it.

3. Solve the Laplace equation. Start with some initial condition, and then evolve 𝑓according to the Poisson semigroup. This extended function is smooth the instant itflows a little bit away from the discrete function.

4. There is a point where the derivative satisfies what we want: The point exists by apigeonhole style argument, but we won’t be able to give it explicitly. This comes fromestimates of the Poisson kernel. We will use Fourier analysis.

2-24

2 Bourgain’s almost extension theorem

Theorem 3.2.1 (Bourgain’s almost extension theorem). Let 𝑋 be a 𝑛-dimensional normedspace, π‘Œ a Banach space, 𝒩𝛿 βŠ† 𝑆𝑋 a 𝛿-net of 𝑆𝑋 , 𝜏 β‰₯ 𝐢𝛿. Suppose 𝑓 : 𝒩𝛿 β†’ π‘Œ is𝐿-Lipschitz. Then there exists 𝐹 : 𝑋 β†’ π‘Œ such that

1. ‖𝐹‖Lip . (1 + π›Ώπ‘›πœ)𝐿.

2. ‖𝐹 (π‘₯)βˆ’ 𝑓(π‘₯)β€–π‘Œ ≀ 𝜏𝐿 for all π‘₯ ∈ 𝒩𝛿.

3. Supp(𝐹 ) βŠ† (2 + 𝜏)𝐡𝑋 .

4. 𝐹 is smooth.

Parts 3 and 4 will come β€œfor free.”

49

Page 50: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

2.1 Lipschitz extension problem

Theorem 3.2.2 (Johnson-Lindenstrauss-Schechtman, 1986). Let 𝑋 be a 𝑛-dimensionalnormed space 𝐴 βŠ† 𝑋, π‘Œ be a Banach space, and 𝑓 : 𝐴 β†’ π‘Œ be 𝐿-Lipschitz. There ex-ists 𝐹 : 𝑋 β†’ π‘Œ such that 𝐹 |𝐴 = 𝑓 and ‖𝐹‖Lip . 𝑛𝐿.

We know a lower bound ofβˆšπ‘›; losing

βˆšπ‘› is sometimes needed. (The lower bound for

nets on the whole space is 4βˆšπ‘›.) A big open problem is what the true bound is.

This was done a year before Bourgain. Why didn’t he use this theorem? This theoremis not sufficient because the Lipschitz constant grows with 𝑛.

We want to show that ‖𝑇‖ β€–π‘‡βˆ’1β€– . 𝐷. We can’t bound the norm of π‘‡βˆ’1 with anythingless than the distortion 𝐷, so to prove Bourgain embedding we can’t lose anything in theLipschitz constant of 𝑇 ; the Lipschitz constant can’t go to ∞ as 𝑛→ ∞.

Bourgain had the idea of relaxing the requirement that the new function be strictlyan extension (i.e., agree with the original function where it is defined). What’s extremelyimportant is that the new function be Lipschitz with a constant independent of 𝑛.

We need ‖𝐹‖Lip . 𝐿. Let’s normalize so 𝐿 = 1.When the parameter is 𝜏 = 𝑛𝛿, we want ‖𝑓(π‘₯)βˆ’ 𝐹 (π‘₯)β€– . 𝑛𝛿. Note 𝛿 is small (less than

the inverse of any polynomial), so losing 𝑛 is nothing.2

How sharp is Theorem 3.2.1? Given a 1-Lipschitz function on a 𝛿-net, if we want toalmost embed it without losing anything, how close can close can we guarantee it to be fromthe original function

Theorem 3.2.3. There exists a 𝑛-dimensional normed space 𝑋, Banach space π‘Œ , 𝛿 > 0,𝒩𝛿 βŠ† 𝑆𝑋 a 𝛿-net, 1-Lipschitz function 𝑓 : 𝒩𝛿 β†’ π‘Œ such that if 𝑓 : 𝑋 β†’ π‘Œ , ‖𝐹‖Lip . 1 thenthere exists π‘₯ ∈ 𝒩𝛿 such that

‖𝐹 (π‘₯)βˆ’ 𝑓(π‘₯)β€– &𝑛

π‘’π‘βˆšln𝑛

.

Thus, what Bourgain proved is essentially sharp. This is a fun construction with Grothendieck’sinequality.

Our strategy is as follows. Consider 𝑃𝑑 * 𝐹 , where

𝑃𝑑(π‘₯) =𝐢𝑛𝑑

(𝑑2 + β€–π‘₯β€–22)𝑛+12

, 𝐢𝑛 =Ξ“(𝑛+12

οΏ½πœ‹

𝑛+12

,

and 𝑃𝑑 is the Poisson kernel. Let (𝑇𝐹 )(π‘₯) = (𝑃𝑑 * 𝐹 )β€²(π‘₯); 𝑇 is linear and ‖𝑇‖ . 1. As𝑑 β†’ 0, this becomes closer to 𝐹 . We hope when 𝑑 β†’ 0 that it is invertible. This is true.We give a proof by contradiction (we won’t actually find what 𝑑 is) using the pigeonholeprinciple.

The Poisson kernel depends on an Euclidean norm in it. In order for the argument towork, we have to choose the right Euclidean norm.

2If people get 𝛿 to be polynomial, then we’ll have to start caring about 𝑛.

50

Page 51: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

2.2 John ellipsoid

Theorem 3.2.4 (John’s theorem). Let 𝑋 be a 𝑛-dimensional normed space, identified withR𝑛. Then there exists a linear operator 𝑇 : R𝑛 β†’ R𝑛 such that 1√

𝑛‖𝑇π‘₯β€–2 ≀ β€–π‘₯‖𝑋 ≀ ‖𝑇π‘₯β€–2

for all π‘₯ ∈ 𝑋.

John’s Theorem says we can always find 𝑇 which sandwiches the 𝑋-norm up to a factorof

βˆšπ‘›. β€œEverything is a Hilbert space up to a

βˆšπ‘› factor.”

Proof. Without loss of generality 𝐷 ≀ 𝑛. Let β„° be an ellipsoid of maximal volume containedin 𝐡𝑋 .

maxβŒ‹vol(𝑆𝐡ℓ𝑛2

) : 𝑆 βˆˆπ‘€π‘›(R), 𝑆𝐡ℓ𝑛2βŠ† 𝐡𝑋

{.

β„° exists by compactness. Take 𝑇 = π‘†βˆ’1.The goal is to show that

βˆšπ‘›β„° βŠ‡ 𝐡𝑋 . We can choose the coordinate system such that

β„° = 𝐡ℓ𝑛2= 𝐡𝑛

2 .Suppose by way of contradiction that

βˆšπ‘›π΅π‘›

2 βŠ‡ 𝐡𝑋 . Then there exists π‘₯ ∈ 𝐡𝑋 such thatβ€–π‘₯‖𝑋 ≀ 1 yet β€–π‘₯β€–2 >

βˆšπ‘›.

Denote 𝑑 = β€–π‘₯β€–2 , 𝑦 = π‘₯𝑑. Then ‖𝑦‖2 = 1 and 𝑑𝑦 ∈ 𝐡𝑋 .

By applying a rotation, we can assume WLOG 𝑦 = 𝑒1.Claim: Define for π‘Ž > 1, 𝑏 < 1

β„°π‘Ž,𝑏 :={

π‘›βˆ‘π‘–=1

𝑑𝑖𝑒𝑖 :(𝑑1π‘Ž

)2

+π‘›βˆ‘π‘–=2

(𝑑𝑖𝑏

)2

≀ 1

}.

Stretching by π‘Ž and squeezing by 𝑏 can make the ellipse grow and stay inside the body,contradicting that it is the minimizer. More precisely, we show that there exists π‘Ž > 1 and𝑏 < 1 such that β„°π‘Ž,𝑏 βŠ† 𝐡𝑋 and Vol(β„°π‘Ž,𝑏) > Vol(𝐡𝑛

2 ). This is equivalent to π‘Žπ‘π‘›βˆ’1 > 1.

We need a lemma with some trigonometry.

Lemma 3.2.5 (2-dimensional lemma). Suppose π‘Ž > 1, 𝑏 ∈ (0, 1), and π‘Ž2

𝑑2+ 𝑏2

(1βˆ’ 1

𝑑2

�≀ 1.

Then β„°π‘Ž,𝑏 βŠ† 𝐡𝑋 .

51

Page 52: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. Let 𝑏 =√

𝑑2βˆ’π‘Ž2𝑑2βˆ’1

. Then πœ“(π‘Ž) = π‘Žπ‘π‘›βˆ’1 = π‘Ž(𝑑2βˆ’π‘Ž2𝑑2βˆ’1

οΏ½π‘›βˆ’12 , πœ“(1) = 1. It is enough to show

πœ“β€²(1) > 0. Now

πœ“β€²(π‘Ž) =

�𝑑2 βˆ’ π‘Ž2

𝑑2 βˆ’ 1

οΏ½π‘›βˆ’12

βˆ’1𝑑2 βˆ’ π‘›π‘Ž2

𝑑2 βˆ’ 1,

which is indeed > 0 for 𝑑 >βˆšπ‘›. Note this is really a 2-dimensional argument; there is 1

special direction. We shrunk the 𝑦 direction and expanded the π‘₯ direction.It’s enough to show the new ellipse β„°π‘Ž,𝑏 is in the rhombus in the picture. Calculations

are left to the reader.

2.3 Proof

Proof of Theorem 3.2.1. By translating 𝑓 (so that it is 0 at some point), without loss ofgenerality we can assume that for all π‘₯ ∈ 𝒩𝛿, ‖𝑓(π‘₯)β€–π‘Œ ≀ 2𝐿.Step 1: (Rough extension 𝐹1 on 𝑆𝑋 .) We show that there exists 𝐹1 : 𝑆𝑋 β†’ π‘Œ such that forall π‘₯ ∈ 𝒩𝛿 β†’ π‘Œ such that for all π‘₯ ∈ 𝒩𝛿,

1. ‖𝐹1(π‘₯)βˆ’ 𝑓(π‘₯)β€–π‘Œ ≀ 2𝐿𝛿.

2. βˆ€π‘₯, 𝑦 ∈ 𝑆𝑋 , ‖𝐹1(π‘₯)βˆ’ 𝐹1(𝑦)β€– ≀ 𝐿(β€–π‘₯βˆ’ 𝑦‖𝑋 + 4𝛿).

This is a partition of unity argument. Write 𝒩𝛿 = {𝑋1, . . . , 𝑋𝑁}. Consider

{(π‘₯𝑝 + 2𝛿𝐡𝑋) ∩ 𝑆𝑋}𝑁𝑝=1,

which is an open cover of 𝑆𝑋 .Let {πœ‘π‘ : 𝑆𝑋 β†’ [0, 1]}𝑁𝑝=1 be a partition of unity subordinted to this open cover. This

means that

52

Page 53: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

1. Suppπœ‘π‘ βŠ† π‘₯𝑝 + 2𝛿𝐡𝑋 ,

2.βˆ‘π‘π‘=1 πœ‘π‘(π‘₯) = 1 for all π‘₯ ∈ 𝑆𝑋 .

For all π‘₯ ∈ 𝑆𝑋 , define 𝐹1(π‘₯) =βˆ‘π‘π‘=1 πœ‘π‘(π‘₯)𝑓(π‘₯𝑝) ∈ π‘Œ . Then as 𝐹1 is a weighted sum of

𝑓(π‘₯𝑝)’s, ‖𝐹1β€–βˆž ≀ 2𝐿. If π‘₯ ∈ 𝒩𝛿, because πœ‘π‘(π‘₯) is 0 when |π‘₯βˆ’ π‘₯𝑝| > 2𝛿,

‖𝐹1(π‘₯)βˆ’ 𝑓(π‘₯)β€–π‘Œ =

βˆ‘π‘:β€–π‘₯βˆ’π‘₯𝑝‖𝑋≀2𝛿

πœ‘π‘(π‘₯)(𝑓(π‘₯𝑝)βˆ’ 𝑓(π‘₯))

≀ βˆ‘

β€–π‘₯βˆ’π‘₯𝑝‖≀2𝛿

πœ‘π‘(π‘₯)𝐿 β€–π‘₯βˆ’ π‘₯𝑝‖𝑋 ≀ 2𝐿𝛿.

For π‘₯, 𝑦 ∈ 𝑆𝑋 ,

‖𝐹1(π‘₯)βˆ’ 𝐹1(𝑦)β€– =

βˆ‘

β€–π‘₯βˆ’ π‘₯𝑝‖ ≀ 2𝛿‖𝑦 βˆ’ π‘₯π‘žβ€– ≀ 2𝛿

(𝑓(π‘₯𝑝)βˆ’ 𝑓(π‘₯π‘ž))πœ‘π‘(π‘₯)πœ‘π‘ž(π‘₯)

β‰€βˆ‘

β€–π‘₯βˆ’ π‘₯𝑝‖ ≀ 2𝛿‖𝑦 βˆ’ π‘₯π‘žβ€– ≀ 2𝛿

𝐿 β€–π‘₯𝑝 βˆ’ π‘₯π‘žβ€–πœ‘π‘(π‘₯)πœ‘π‘(𝑦)

≀ 𝐿(β€–π‘₯βˆ’ 𝑦‖+ 4𝛿).

2-29: We continue proving Bourgain’s almost extension theorem.Step 2: Extend 𝐹1 to 𝐹2 on the whole space such that

1. βˆ€π‘₯ ∈ 𝒩𝛿, ‖𝐹2(π‘₯)βˆ’ 𝑓(π‘₯)β€–π‘Œ . 𝐿𝛿.

2. ‖𝐹2(π‘₯)βˆ’ 𝐹2(𝑦)β€–π‘Œ . 𝐿(β€–π‘₯βˆ’ 𝑦‖𝑋 + 𝛿).

3. Supp(𝐹2) βŠ† 2𝐡𝑋 .

4. 𝐹2 is smooth.

Denote 𝛼(𝑑) = max{1βˆ’ |1βˆ’ 𝑑|, 0}.

53

Page 54: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Let

𝐹2(π‘₯) = 𝛼(β€–π‘₯‖𝑋)𝐹1 (π‘₯) β€–π‘₯‖𝑋 .

𝐹2 still satisfies condition 1. As for condition 2,

‖𝐹2(π‘₯)βˆ’ 𝐹2(𝑦)β€–π‘Œ =

𝛼(β€–π‘₯‖𝑋)𝐹1

οΏ½π‘₯

β€–π‘₯‖𝑋

οΏ½βˆ’ 𝛼(‖𝑦‖𝑋)𝐹1

�𝑦

‖𝑦‖𝑋

�≀ |𝛼(β€–π‘₯β€–)βˆ’ 𝛼(‖𝑦‖)|

𝐹1

οΏ½π‘₯

β€–π‘₯‖𝑋

�⏟ ⏞

≀2𝐿

+𝛼(‖𝑦‖)𝐹1

οΏ½π‘₯

β€–π‘₯‖𝑋

οΏ½βˆ’ 𝐹1

�𝑦

‖𝑦‖𝑋

οΏ½

≀ (β€–π‘₯β€– βˆ’ ‖𝑦‖)2𝐿+ 𝛼(‖𝑦‖)𝐿(

π‘₯

β€–π‘₯β€–βˆ’ 𝑦

‖𝑦‖

+ 4𝛿

)≀ 2𝐿 β€–π‘₯βˆ’ 𝑦‖+ 𝐿𝛼(‖𝑦‖)

(β€–π‘₯β€–

1

β€–π‘₯β€–βˆ’ 1

‖𝑦‖

+

β€–π‘₯βˆ’ 𝑦‖‖𝑦‖

+ 4𝛿

)≀ 2𝐿 β€–π‘₯βˆ’ 𝑦‖+ 𝐿𝛼(‖𝑦‖)

οΏ½β€–π‘₯βˆ’ 𝑦‖‖𝑦‖

+β€–π‘₯βˆ’ 𝑦‖‖𝑦‖

+ 4𝛿

οΏ½. 𝐿(β€–π‘₯βˆ’ 𝑦‖+ 𝛿),

where in the last step we used 𝛼(‖𝑦‖) ≀ ‖𝑦‖ and 𝛼(‖𝑦‖) ≀ 1.Note 𝐹2 is smooth because the sum for 𝐹1 was against a partition of unity and ‖·‖𝑋 is

smooth, although we don’t have uniform bounds on smoothness for 𝐹2.Step 3: We make 𝐹 smoother by convolving.

Lemma 3.2.6 (Begun, 1999). Let 𝐹2 : 𝑋 β†’ π‘Œ satisfy ‖𝐹2(π‘₯)βˆ’ 𝐹2(𝑦)β€–π‘Œ ≀ 𝐿(β€–π‘₯βˆ’ 𝑦‖𝑋+𝛿).Let 𝜏 β‰₯ 𝑐𝛿. Define

𝐹 (π‘₯) =1

Vol(πœπ΅π‘‹)

βˆ«πœπ΅π‘‹

𝐹2(π‘₯+ 𝑦) 𝑑𝑦.

Then

‖𝐹‖Lip ≀ 𝐿

οΏ½1 +

𝛿𝑛

2𝜏

οΏ½.

The lemma proves the almost extension theorem as follows. We passed from 𝑓 : 𝒩𝛿 β†’ π‘Œto 𝐹1 to 𝐹2 to 𝐹 . If π‘₯ ∈ 𝒩𝛿,

‖𝐹 (π‘₯)βˆ’ 𝑓(π‘₯)β€–π‘Œ =

1

Vol(πœπ΅π‘‹)

βˆ«π΅π‘‹

(𝐹2(π‘₯+ 𝑦)βˆ’ 𝑓(π‘₯)) 𝑑𝑦

≀ 1

Vol(πœπ΅π‘‹)

βˆ«πœπ΅π‘‹

‖𝐹2(π‘₯+ 𝑦)βˆ’ 𝐹2(π‘₯)β€–π‘Œ + ‖𝐹2(π‘₯)βˆ’ 𝑓(π‘₯)β€–π‘ŒβŸ ⏞ 𝛿𝐿

𝑑𝑦

≀ 1

Vol(πœπ΅π‘‹)

βˆ«πœπ΅π‘‹

(𝐿(β€–π‘¦β€–π‘‹βŸ ⏞ β‰€πœ

+𝛿𝐿)) 𝑑𝑦 . 𝐿𝜏.

Now we prove the lemma.

54

Page 55: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. We need to show

‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–π‘Œ ≀ 𝐿

οΏ½1 +

𝛿𝑛

2𝜏

οΏ½β€–π‘₯βˆ’ 𝑦‖𝑋 .

Without loss of generality 𝑦 = 0, Vol(πœπ΅π‘‹) = 1. Denote

𝑀 = πœπ΅π‘‹βˆ–(π‘₯+ πœπ΅π‘‹)

𝑀 β€² = (π‘₯+ πœπ΅π‘‹)βˆ–πœπ΅π‘‹ .

We have

𝐹 (0)βˆ’ 𝐹 (π‘₯) =βˆ«π‘€πΉπ‘§(𝑦) 𝑑𝑦 βˆ’

βˆ«π‘€ ′𝐹𝑧(𝑦) 𝑑𝑦.

Define πœ”(𝑧) to be the Euclidean length of the interval (𝑧 + Rπ‘₯) ∩ (πœπ΅π‘‹). By Fubini,∫Proj

𝑋βŠ₯ (πœπ΅π‘‹)πœ”(𝑧) 𝑑𝑧 = Vol𝑛(πœπ΅π‘‹) = 1.

Denote

π‘Š = {𝑧 ∈ πœπ΅π‘‹ : (𝑧 + Rπ‘₯) ∩ (πœπ΅π‘‹) ∩ (π‘₯+ πœπ΅π‘‹) = πœ‘}𝑁 = πœπ΅π‘‹βˆ–π‘Š.

Define 𝐢 : 𝑀 β†’ 𝑀 β€² a shift in direction 𝑋 on every fiber that maps the interval (𝑧 + Rπ‘₯) βˆ©π‘€ β†’ (𝑧 + Rπ‘₯) βˆ©π‘€ β€².

55

Page 56: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

𝐢 is a measure preserving transformation with

‖𝑧 βˆ’ 𝐢(𝑧)‖𝑋 =

βŽ§βŽ¨βŽ©β€–π‘₯‖𝑋 , 𝑧 ≀ 𝑁

πœ”(𝑧)β€–π‘₯‖𝑋‖π‘₯β€–2

, 𝑧 ∈ π‘Š βˆ©π‘€.

(In the second case we translate by an extra factor πœ”(𝑧)β€–π‘₯β€–2

.) Then

‖𝐹 (0)βˆ’ 𝐹 (π‘₯)β€–π‘Œ =∫

𝑀𝐹2(𝑦) 𝑑𝑦 βˆ’

βˆ«π‘€ ′𝐹2(𝑦) 𝑑𝑦

π‘Œ

=∫

𝑀(𝐹2(𝑦)βˆ’ 𝐹2(𝐢(𝑦))) 𝑑𝑦

π‘Œ

β‰€βˆ«π‘€πΏ(‖𝑦 βˆ’ 𝐢(𝑦)‖𝑋 + 𝛿) 𝑑𝑦

β‰€βˆ«π‘€πΏ(‖𝑦 βˆ’ 𝐢(𝑦)‖𝑋 + 𝛿) 𝑑𝑦

= 𝐿𝛿Vol(𝑀) + πΏβˆ«π‘€β€–π‘¦ βˆ’ 𝐢(𝑦)‖𝑋 π‘‘π‘¦βˆ«

𝑀‖𝑦 βˆ’ 𝐢(𝑦)‖𝑋 𝑑𝑦 =

βˆ«π‘β€–π‘₯‖𝑋 𝑑𝑦 +

βˆ«π‘Šβˆ©π‘€

πœ”(𝑦) β€–π‘₯‖𝑋‖π‘₯β€–2

𝑑𝑦

= β€–π‘₯‖𝑋 Vol(𝑁) +∫Proj(π‘Šβˆ©π‘€)

πœ”(𝑧) β€–π‘₯‖𝑋‖π‘₯β€–2

β€–π‘₯β€–2 𝑑𝑧 orthogonal decomposition

= β€–π‘₯‖𝑋 Vol(𝑁) + Vol(πœπ΅π‘‹βˆ–π‘)

= β€–π‘₯‖𝑋 Vol(πœπ΅π‘‹) = β€–π‘₯‖𝑋 .

We show 𝑀 = πœπ΅π‘‹βˆ–(π‘₯+ πœπ΅π‘‹) βŠ† πœπ΅π‘‹βˆ–(1βˆ’ β€–π‘₯β€–πœ)πœπ΅π‘‹ . Indeed, for 𝑦 βˆˆπ‘€ ,

‖𝑦 βˆ’ π‘₯‖𝑋 β‰₯ 𝜏

‖𝑦‖ β‰₯ 𝜏 βˆ’ β€–π‘₯β€– =

οΏ½1βˆ’ β€–π‘₯β€–

𝜏

�𝜏.

56

Page 57: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Then

Vol(𝑀) ≀ Vol(πœπ΅π‘‹)βˆ’ Vol

οΏ½οΏ½1βˆ’ β€–π‘₯β€–

𝜏

οΏ½πœπ΅π‘‹

οΏ½= 1βˆ’

οΏ½1βˆ’ β€–π‘₯β€–

𝜏

οΏ½.𝑛 β€–π‘₯β€–πœ

Bourgain did it in a more complicated, analytic way avoiding geometry. Begun noticesthat careful geometry is sufficient.

Later we will show this theorem is sharp.

3 Proof of Bourgain’s discretization theorem

At small distances there is no guarantee on the function 𝑓 . Just taking derivatives is danger-ous. It might be true that we can work with the initial function. But the only way Bourgainfigured out how to prove the theorem was to make a 1-parameter family of functions.

3.1 The Poisson semigroup

Definition 3.3.1: The Poisson kernel is 𝑃𝑑(π‘₯) : R𝑛 β†’ R given by

𝑃𝑑(π‘₯) =𝐢𝑛𝑑

(𝑑2 + β€–π‘₯β€–22)𝑛+12

, 𝐢𝑛 =Ξ“(𝑛+12

οΏ½πœ‹

𝑛+12

.

Proposition 3.3.2 (Properties of Poisson kernel): 1. For all 𝑑 > 0,∫R𝑛 𝑃𝑑(π‘₯) 𝑑π‘₯ = 1.

2. (Semigroup property) 𝑃𝑑 * 𝑃𝑠 = 𝑃𝑑+𝑠.

3. 𝑃𝑑(π‘₯) = π‘’βˆ’2πœ‹β€–π‘₯β€–2𝑑.

Lemma 3.3.3. Let 𝐹 be the function obtained from Bourgain’s almost extension theo-rem 3.2.1. For all 𝑑 > 0, ‖𝑃𝑑 * 𝐹‖Lip . 1.

Proof. We have

𝑃𝑑 * 𝐹 (π‘₯)βˆ’ 𝑃𝑑 * 𝐹 (𝑦) =∫R𝑛𝑃𝑑(𝑧)(𝐹 (π‘₯βˆ’ 𝑧)βˆ’ 𝐹 (π‘₯βˆ’ 𝑦)) 𝑑𝑧.

Now use the fact that 𝐹 is Lipschitz.

Our goal is to show there exists 𝑑0 > 0, π‘₯ ∈ 𝐡 such that if we define

eq:bdt-T𝑇 = (𝑃𝑑0 * 𝐹 )β€²(π‘₯) : 𝑋 β†’ π‘Œ, (3.1)

(i.e., π‘‡π‘Ž = πœ•π‘Ž(𝑃𝑑0 *𝐹 )(π‘₯)). We showed ‖𝑇‖ . 1. It remains to show β€–π‘‡βˆ’1β€– . 𝐷. Then 𝑇 hasdistortion at most 𝑂(𝐷), and hence 𝑇 gives the desired extension in Bourgain’s discretizatontheorem 1.2.4.

3-2: Today we continue with Bourgain’s Theorem. Summarizing our progress so far:

57

Page 58: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

1. Initially we had a function 𝑓 : 𝒩𝛿 β†’ π‘Œ and a 𝛿-net 𝒩𝛿 of 𝐡𝑋 .

2. From 𝑓 we got a new function 𝐹 : 𝑋 β†’ π‘Œ with supp(𝐹 ) βŠ† 3𝐡𝑋 , satisfying thefollowing. (None of the constants matter.)

(a) ‖𝐹‖𝐿𝑝 ≀ 𝐢 for some constant.

(b) For all π‘₯ ∈ 𝒩𝛿, ‖𝐹 (π‘₯)βˆ’ 𝑓(π‘₯)β€–π‘Œ ≀ 𝑛𝛿.

Let 𝑃𝑑 : R𝑛 β†’ R be the Poisson kernel. What can be said about the convolution 𝑃𝑑 * 𝐹?We know that for all 𝑑 and for all π‘₯, β€–(𝑃𝑑 *𝐹 )(π‘₯)β€–π‘‹β†’π‘Œ ≀ 1. The goal is to make this functioninvertible. More precisely, the norm of the inverted operator is not too small.

We’ll ensure this one direction at a time by investigating the directional derivatives. For𝑔 : R𝑛 β†’ π‘Œ , the directional derivative is defined by: for all π‘Ž ∈ 𝑆𝑋 ,

πœ•π‘Žπ‘”(π‘₯) = lim𝑑→0

𝑔(π‘₯+ π‘‘π‘Ž)βˆ’ 𝑔(π‘₯)

𝑑.

There isn’t any particular reason why we need to use the Poisson kernel; there are manyother kernels which satisfy this. We definitely need the semigroup property, in addition toall sorts of decay conditions.

We need a lot of lemmas.

Lemma 3.3.4 (Key lemma 1). lem:bdt-1 There are universal constants 𝐢,𝐢 β€², 𝐢 β€²β€² such that thefollowing hold. Suppose 𝑑 ∈ (0, 1

2], 𝑅 ∈ (0,∞), 𝛿 ∈ (0, 1

100𝑛) satisfy

𝛿 ≀ 𝐢𝑑 log(3/𝑑)√

𝑛≀ 𝐢 β€² 1

𝑛5/2𝐷2(3.2)

𝐢 ′′𝑛3/2𝐷2 log(3/𝑑) ≀ 𝑅 ≀ 𝐢 β€²β€²

π‘‘βˆšπ‘›. (3.3)

Then for all π‘₯ ∈ 12𝐡𝑋 , π‘Ž ∈ 𝑆𝑋 , we have

(β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ * 𝑃𝑅𝑑)(π‘₯) β‰₯𝐢 β€²β€²β€²

𝐷

This result says that given a direction π‘Ž and look at how long the vector πœ•π‘Ž(𝑃𝑑 *𝐹 ) is, itis not only large, but large on average. Here the average is over the time period 𝑅𝑑.

Note that I haven’t been very careful with constants; there are some points where theconstant is missing/needs to be adjusted.

The first condition says is that 𝑑 is not too small, but the lower bound is extremely small.

The upper bound is(1𝑛

οΏ½π‘˜for some π‘˜ whereas the lower bound is something like π‘’βˆ’π‘’

𝑛, so

there’s a huge gap.The second condition says that log

(1𝑑

οΏ½is still exponential.

This lemma will come up later on and it will be clear.From now on, let us assume this key lemma and I will show you how to finish. We will

go back and prove it later.

58

Page 59: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Lemma 3.3.5 (Key lemma 2). lem:bdt-2 There’s a constant 𝐢1 such that the following holds (wecan take 𝐢1 = 8). Let πœ‡ be any Borel probability measure on 𝑆𝑋 . For every 𝑅,𝐴 ∈ (0,∞),there exists 𝐴

(𝑅+1)π‘š+1 ≀ 𝑑 ≀ 𝐴 such thatβˆ«π‘†π‘‹

∫R𝑛

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯π‘‘πœ‡(π‘Ž) β‰€βˆ«π‘†π‘‹

∫R𝑛

β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹 )β€–π‘Œ 𝑑π‘₯⏟ ⏞ (*)

π‘‘πœ‡(π‘Ž) +𝐢1vol(3𝐡𝑋)

π‘š

Basically, under convolution, we can take the derivative under the integral sign. Thus(*) is an average of what we wrote on the left. Averaging only makes things smaller fornorms, because norms are convex. Thus the right integral is less than the left integral, fromJensen’s inequality.

The lemma says that adding that small factor, we get a bound in the opposite direction.We will argue that there must be a scale at which Jensen’s inequality stabilizes, i.e., β€–πœ•π‘Ž(𝑃𝑑 *𝐹 )β€–π‘Œ * 𝑃𝑅𝑑(π‘₯) stabilizes to a constant.

Note this is really the pigeonhole argument, geometrically.

Proof. Bourgain uses this technique a lot in his papers: find some point where the inequalityis equality, and then leverage that.

If the result does not hold, then for every 𝑑 in the range 𝑑(𝑅+1)π‘š+1 ≀ 𝑑 ≀ 𝐴, the reverse

holds. We’ll use the inequality at eveyr point in a geometric series. For all π‘˜ ∈ {0, . . . ,π‘š+1},βˆ«π‘†π‘‹

∫R𝑛

β€–πœ•π‘Ž(𝑃𝐴(𝑅+1)π‘˜βˆ’π‘šβˆ’1*𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯π‘‘πœ‡(π‘Ž) >βˆ«π‘†π‘‹

∫R𝑛

β€–πœ•π‘Ž(𝑃𝐴(𝑅+1)π‘˜βˆ’π‘š*𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯π‘‘πœ‡(π‘Ž)+𝐢1vol(3𝐡𝑋)

π‘š.

Summing up these inequalities and telescopingβˆ«π‘†π‘‹

∫R𝑛

β€–πœ•π‘Ž(𝑃𝐴(𝑅+1)βˆ’π‘šβˆ’1*𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯π‘‘πœ‡(π‘Ž) >βˆ«π‘†π‘‹

∫R𝑛

β€–πœ•π‘Ž(𝑃𝐴(𝑅+1)*𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯π‘‘πœ‡(π‘Ž)+8(π‘š+ 1)vol(3𝐡𝑋)

π‘š(3.4)

(recall that we’ve been using the semigroup properties of the Poisson kernel this whole time).Now why is this conclusion absurd? Take 𝐢1 to be the bound on the Lipschitz constant of𝐹 . Because β€–πœ•π‘ŽπΉβ€–π‘Œ ≀ 8, we haveπœ•π‘Ž(𝑃𝐴(𝑅+1)βˆ’π‘šβˆ’1 * 𝐹 )(π‘₯) ≀ 𝐢1. Since partial derivativescommute with the integral sign, we get∫

𝑆𝑋

∫R𝑛

β€–πœ•π‘Ž(𝑃𝐴(𝑅+1)βˆ’π‘šβˆ’1 * 𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯ π‘‘πœ‡(π‘Ž) =βˆ«π‘†π‘‹

∫R𝑛

β€–(𝑃𝐴(𝑅+1)βˆ’π‘šβˆ’1 * πœ•π‘ŽπΉ )(π‘₯)β€–π‘Œ 𝑑π‘₯ π‘‘πœ‡(π‘Ž)

(3.5)

β‰€βˆ«π‘†π‘‹

∫R𝑛𝐢1 ≀ 𝐢1vol(𝐡𝑋) (3.6)

because 𝐹 is Lipschitz and the Poisson semigroup integrates to unity. Together (3.4)and (3.6) give a contradiction.

Now assuming Lemma 3.3.4 (Key Lemma 1), let’s complete the proof of Bourgain’s

discretization theorem. Assume from now on 𝛿 <(

1𝑐𝐷

οΏ½(𝐢𝐷)2𝑛

where 𝐢 is a large enoughconstant that we will choose (𝐢 = 500 or something).

59

Page 60: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof of Bourgain’s Discretization Theorem 1.2.4. Let β„± βŠ† 𝑆𝑋 be a 1𝐢2𝐷

-net in 𝑆𝑋 . Then|β„±| ≀ (𝐢3𝐷)𝑛 for some 𝐢3. We will apply the Lemma 3.3.5 (Key Lemma 2) with πœ‡ theuniform measure on β„± ,

𝐴 = (1/𝐢𝐷)5𝑛

𝑅 + 1 = (𝐢𝐷)4𝑛

π‘š = ⌈(𝐢𝐷)𝑛+1βŒ‰.

Then there exists (1/(𝐢𝐷))(𝐢𝐷)2𝑛 ≀ 𝑑 ≀ (1/(𝐢𝐷))5𝑛 such that

eq:bdt-pf-lem2

βˆ‘π‘Žβˆˆβ„±

∫R𝑛

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯ β‰€βˆ‘π‘Žβˆˆβ„±

∫R𝑛

β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹 )(π‘₯)β€–π‘Œ 𝑑π‘₯+8vol(3𝐡𝑋)

π‘š|β„±|

(3.7)We check the conditions of Lemma 3.3.4 (Key Lemma 1).

1. For (3.2), note 𝑑 is exponentially small, so the RHS inequality is satisfied. For the LHS

inequality, note 𝛿 ≀ 𝑑 and𝐢 ln( 1

𝑑 )βˆšπ‘›

≀ 1. To see the second inequality, note𝐢 ln( 1

𝑑 )βˆšπ‘›

β‰₯𝐢5𝑛 ln(𝐢𝐷)√

𝑛β‰₯ 1 (for 𝑛 large enough).

2. For (3.2), note the LHS is dominated by ln(1𝑑

�≀ (𝐢𝐷)2𝑛 ln(𝐢𝐷) which is much less

than 𝑅 = (𝐢𝐷)4𝑛 βˆ’ 1, and 1𝑑, the dominating term on the RHS, is β‰₯ (𝐢𝐷)5𝑛.

In my paper (reference?), I wrote what the exact constants are. They’re written in thepaper, and are not that important. We’re choosing our constants big enough so that ourinequalities hold true with room to spare.

Now we can use Key Lemma 1, which says

eq:bdt-pf-lem1 (β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ * 𝑃𝑅𝑑) (π‘₯) β‰₯1

𝐷(3.8)

Using 𝑃(𝑅+1)𝑑 = 𝑃𝑅𝑑 * 𝑃𝑑 (semigroup property) and Jensen’s inequality on the norm, whichis a convex function, we have

eq:bdt-convβ€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 *𝐹 )(π‘₯)β€–π‘Œ = β€– (πœ•π‘Ž(𝑃𝑑 * 𝐹 ))*𝑃𝑅𝑑(π‘₯)β€–π‘Œ ≀ (β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ * 𝑃𝑅𝑑) (π‘₯). (3.9)

Since the norm is a convex function, by Jensen’s we get our inequality.Let

πœ“(π‘₯) := (β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ * 𝑃𝑅𝑑) (π‘₯)βˆ’ β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹

οΏ½(π‘₯)β€–π‘Œ

From (3.9), πœ“(π‘₯) β‰₯ 0 pointwise. Using Markov’s inequality,

vol{π‘₯ ∈ 1

2𝐡𝑋 : πœ“(π‘₯) >

1

𝐷

}≀ 𝐷

∫R𝑛

(β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ * 𝑃𝑅𝑑) (π‘₯)βˆ’ β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹

οΏ½(π‘₯)β€–π‘Œ )𝑑π‘₯.

because we can upper bound the integral over the ball by an integral over R𝑛. Note we areusing the probabilistic method.

60

Page 61: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

This inequality was for a fixed π‘Ž. We now use the union bound over β„± , a 𝛿-net of π‘Žβ€™s toget

vol{π‘₯ ∈ 1

2𝐡𝑋 : βˆƒπ‘Ž ∈ β„± , πœ“π‘Ž(π‘₯) > 1/𝐷

}(3.10)

β‰€βˆ‘π‘Žβˆˆβ„±

vol(π‘₯ ∈ 1

2𝐡𝑋 : πœ“π‘Ž(π‘₯) > 1/𝐷) (3.11)

≀ π·βˆ‘π‘Žβˆˆβ„±

∫R𝑛

(β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ * 𝑃𝑅𝑑)(π‘₯)βˆ’ β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹 )(π‘₯)β€–π‘Œ

�𝑑π‘₯ (3.12)

= π·βˆ‘π‘Žβˆˆβ„±

∫R𝑛

(β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ )(π‘₯)βˆ’ β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹 )(π‘₯)β€–π‘Œ

�𝑑π‘₯ (3.13)

≀ 𝐢1vol(3𝐡𝑋)

π‘š(𝐢3𝐷)𝑛 < vol

(1

2𝐡𝑋

). (3.14)

where (3.13) follows because when you convolve something with a probability measure, theintegral over R𝑛 does not change, and (3.14) follows from (3.7) and our choice of π‘š.

We’ve proved that there must exist a point in half the ball such that for every π‘Ž ∈ β„± ,our net, we have

1

𝐷β‰₯ β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€–π‘Œ * 𝑃𝑅𝑑(π‘₯)βˆ’ β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹 )(π‘₯)β€–π‘Œ

I think we want either this to be 12𝐷

, or (3.8) to be 1𝐷/2

, in order to get the following bound

(with 12𝐷

or 1𝐷). This involves changing some constants in the proof.

From this we can conclude that for this π‘₯,

β€–π‘‡π‘Žβ€–π‘Œ = β€–πœ•π‘Ž(𝑃(𝑅+1)𝑑 * 𝐹 )(π‘₯)β€–π‘Œ β‰₯ 1/𝐷

where π‘Ž ∈ 𝑆𝑋 we let 𝑑0 = (𝑅 + 1)𝑑. This shows β€–π‘‡βˆ’1β€– ≀ 𝐷.

For the exact constants, see the paper (reference).Note this is very much a probabiblistic existence statement or result. Usually we estimate

by hand the random variable we want to care about. Here we want to prove an existentialbound, so we estimate the probability of the bad case. But we estimate the probability ofthe bad case using another proof by contradiction.

It remains to estimate some integrals.3-7: Finishing Bourgain. There’s a remaining lemma about the Poisson semigroup that

we’re going to do today (Lemma 3.3.4); it’s nice, but it’s not as nice as what came before.The last remaining lemma to prove is Lemma 3.3.4.Remember that 1√

𝑛‖π‘₯β€–2 ≀ β€–π‘₯‖𝑋 ≀ β€–π‘₯β€–2 for all π‘₯ ∈ 𝑋. We need three facts about the

Poisson kernel.

Proposition 3.3.6 (Fact 1): pr:pois1∫Rπ‘›βˆ–(π‘Ÿπ΅π‘‹)

𝑃𝑑(π‘₯)𝑑π‘₯ ≀ π‘‘βˆšπ‘›

π‘Ÿ.

61

Page 62: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. First note that the Poisson semigroup has the property that

eq:pois-rescale𝑃𝑑(π‘₯) =1

𝑑𝑛𝑃1(π‘₯/𝑑), (3.15)

i.e., 𝑃𝑑 is just a rescaling of 𝑃1. So it’s enough to prove this for 𝑑 = 1. We haveβˆ«β€–π‘₯‖𝑋β‰₯π‘Ÿ

𝑃𝑑(π‘₯) 𝑑π‘₯ β‰€βˆ«β€–π‘₯β€–2β‰₯π‘Ÿ

𝑃𝑑(π‘₯) 𝑑π‘₯

=βˆ«β€–π‘₯β€–2β‰₯π‘Ÿ/𝑑

𝑃1(π‘₯) 𝑑π‘₯

by change of variables and (3.15)

= πΆπ‘›π‘†π‘›βˆ’1

∫ ∞

π‘Ÿ/𝑑

π‘ π‘›βˆ’1

(1 + 𝑠2)(𝑛+1)/2𝑑𝑠

polar coordinates, where π‘†π‘›βˆ’1 = vol(Sπ‘›βˆ’1)

≀ πΆπ‘›π‘†π‘›βˆ’1

∫ ∞

π‘Ÿ/𝑑

1

𝑠2𝑑𝑠

=𝑑

π‘ŸπΆπ‘›π‘†π‘›βˆ’1

≀ π‘‘βˆšπ‘›

π‘Ÿ,

where the last inequality follows from expanding in terms of the Gamma function and usingStirling’s formula.

Above we changed to polar coordinates usingβˆ«β€–π‘₯β€–2≀𝑅

𝑓(β€–π‘₯β€–) 𝑑π‘₯ =∫ 𝑅

0π‘†π‘›βˆ’1 β€–π‘₯β€–π‘›βˆ’1 𝑓(π‘Ÿ) π‘‘π‘Ÿ.

Proposition 3.3.7 (Fact 2): For all 𝑦 ∈ R𝑛,∫R𝑛

|𝑃𝑑(π‘₯)βˆ’ 𝑃𝑑(π‘₯+ 𝑦)|𝑑π‘₯ β‰€βˆšπ‘›β€–π‘¦β€–2𝑑

.

Proof. It’s again enough to prove this when 𝑑 = 1.∫R𝑛

|𝑃𝑑(π‘₯)βˆ’ 𝑃𝑑(π‘₯+ 𝑦)| 𝑑π‘₯ =∫R𝑛

∫ 1

0βŸ¨βˆ‡π‘ƒπ‘‘(π‘₯+ 𝑠𝑦), π‘¦βŸ©π‘‘π‘ 

𝑑π‘₯

≀ ‖𝑦‖2∫R𝑛

β€–βˆ‡π‘ƒπ‘‘(π‘₯)β€–2𝑑π‘₯ Cauchy-Schwarz

≀ ‖𝑦‖2(𝑛+ 1)𝐢𝑛

∫R𝑛

β€–π‘₯β€–2(1 + β€–π‘₯β€–22)

𝑛+32

𝑑π‘₯ computing gradient

= ‖𝑦‖2(𝑛+ 1)πΆπ‘›π‘†π‘›βˆ’1

∫ ∞

0

π‘Ÿπ‘›

(1 + π‘Ÿ2)𝑛+32

π‘‘π‘Ÿ polar coordinates

β‰€βˆšπ‘› ‖𝑦‖2 .

62

Page 63: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

The integral and multiplicative constants perfectly cancels out the 𝑛 + 1 and becomes 1(calculation omitted).

Proposition 3.3.8 (Fact 3): For all 0 < 𝑑 < 12and π‘₯ ∈ 𝐡𝑋 , we have

‖𝑃𝑑 * 𝐹 (π‘₯)βˆ’ 𝐹 (π‘₯)β€–π‘Œ β‰€βˆšπ‘›π‘‘ log(3/𝑑).

Proof. The LHS equals∫R𝑛

(𝐹 (π‘₯βˆ’ 𝑦)βˆ’ 𝐹 (π‘₯))𝑃𝑑(𝑦)π‘‘π‘¦π‘Œβ‰€βˆ«π‘₯+3𝐡𝑋

‖𝐹 (π‘₯βˆ’ 𝑦)βˆ’ 𝐹 (π‘₯)β€–π‘Œ 𝑃𝑑(𝑦)π‘‘π‘¦βŸ ⏞ (3.16)

+ π‘βˆ«Rπ‘›βˆ–(π‘₯+3𝐡𝑋)

𝑃𝑑(𝑦)π‘‘π‘¦βŸ ⏞ (3.17)

.

Using ‖𝐹 (π‘₯βˆ’ 𝑦)βˆ’ 𝐹 (π‘₯)β€–π‘Œ ≀ ‖𝐹‖Lip π‘πœŒ Β· ‖𝑦‖𝑋 and that ‖𝐹‖Lip is a constant,

(3.16) =∫π‘₯+3𝐡𝑋

‖𝑦‖𝑋𝑃𝑑(𝑦)𝑑𝑦

β‰€βˆ«4βˆšπ‘›π΅π‘™π‘›

2

‖𝑦‖2𝑃𝑑(𝑦)𝑑𝑦 using π‘₯ ∈ 𝐡𝑋 =β‡’ π‘₯+ 3𝐡𝑋 βŠ† 4βˆšπ‘›π΅π‘™2

= π‘‘πΆπ‘›π‘†π‘›βˆ’1

∫ 4βˆšπ‘›

𝑑

0

𝑠𝑛

(1 + 𝑠2)(𝑛+1)/2𝑑𝑠 polar coordinates

≀ π‘‘βˆšπ‘›

�∫ βˆšπ‘›

0

1βˆšπ‘›π‘‘π‘ +

∫ 4βˆšπ‘›/𝑑

βˆšπ‘›

1

𝑠𝑑𝑠

�≀ 𝑑

βˆšπ‘›(1 + ln

(4

𝑑

)),

where we used the fact that 𝑠𝑛

(1+𝑠2)𝑛+12

is maximized when 𝑠 =βˆšπ‘›, and is always ≀ 1

𝑠.

For the second term, note that 2𝐡𝑋 βŠ† π‘₯+ 3𝐡𝑋 , so

(3.17) = π‘βˆ«Rπ‘›βˆ–2𝐡𝑋

𝑃𝑑(𝑦) β‰€π‘π‘‘βˆšπ‘›

2

where we used the tail bound of Fact 1 (Pr. 3.3.7). Adding these two bounds gives theresult.

Now we have these three facts, we can finish the proof of the lemma.

Proof of Lemma 3.3.4.

Claim 3.3.9 (𝑃𝑑 * 𝐹 separates points). Let πœƒ = π‘π·βˆšπ‘›π‘‘ log(3/𝑑). Let 𝑀, 𝑦 ∈ 1

2𝐡𝑋 with

‖𝑀 βˆ’ 𝑦‖𝑋 β‰₯ πœƒ. Then

‖𝑃𝑑 * 𝐹 (𝑀)βˆ’ 𝑃𝑑 * 𝐹 (𝑦)β€–π‘Œ β‰₯ 1/𝐷

We do not claim at all that these numbers are sharp.

63

Page 64: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. We can find a point 𝑝 ∈ 𝒩𝛿 which is 𝛿 away from 𝑀, and π‘ž ∈ 𝒩𝛿 which is 𝛿 away from𝑦. We also know that 𝐹 is close to 𝑓 , and we can use Fact 3 (Pr. 3.3.8) to bound the errorof convolving.

By the triangle inequality

‖𝑃𝑑 * 𝐹 (𝑀)βˆ’ 𝑃𝑑 * 𝐹 (𝑦)β€–π‘Œ β‰₯ ‖𝑓(𝑝)βˆ’ 𝑓(π‘ž)β€– βˆ’ ‖𝐹 (𝑝)βˆ’ 𝑓(𝑝)β€– βˆ’ ‖𝐹 (π‘ž)βˆ’ 𝑓(π‘ž)β€–βˆ’ ‖𝐹 (𝑀)βˆ’ 𝐹 (𝑝)β€– βˆ’ ‖𝐹 (𝑦)βˆ’ 𝐹 (π‘ž)β€–βˆ’ ‖𝑃𝑑 * 𝐹 (𝑀)βˆ’ 𝐹 (𝑀)β€– βˆ’ ‖𝑃𝑑 * 𝐹 (𝑦)βˆ’ 𝐹 (𝑦)β€–

β‰₯ β€–π‘βˆ’ π‘žβ€–π·

βˆ’ 2𝑛𝛿 βˆ’ 𝑐𝛿 βˆ’βˆšπ‘›π‘‘ log(3/𝑑)

β‰₯ ‖𝑦 βˆ’ 𝑀‖ βˆ’ 2𝛿

π·βˆ’ 2𝑛𝛿 βˆ’ 𝑐𝛿 βˆ’

βˆšπ‘›π‘‘ log(3/𝑑).

Taking

πœƒ = π‘π·βˆšπ‘›π‘‘ log(3/𝑑),

we find this is at least a constant times β€–π‘¦βˆ’π‘€β€–/𝐷, we need πœƒ = π‘π·βˆšπ‘›π‘‘ log(3/𝑑). Note that√

𝑛𝑑 log(3/𝑑) is the largest error term because by the assumptions in the lemma, it is greaterthan 2𝑛𝛿.

Now consider

‖𝑃𝑑 * 𝐹 (𝑧 + πœƒπ‘Ž)βˆ’ 𝑃𝑑 * 𝐹 (𝑧)β€–

By the claim, if 𝑧 ∈ 14𝐡𝑋 , then we know

πœƒ/𝐷 ≀ ‖𝑃𝑑 * 𝐹 (𝑧 + πœƒπ‘Ž)βˆ’ 𝑃𝑑 * 𝐹 (𝑧)β€– (3.18)

=∫ πœƒ

0β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(𝑧 + π‘ π‘Ž)β€–π‘Œ (3.19)

64

Page 65: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Suppose β€–π‘₯βˆ’ 𝑦‖ ≀ 14(we will apply (3.19) to π‘₯βˆ’ 𝑦 = 𝑧), π‘₯ ∈ 1

2𝐡𝑋 , 𝑦 ∈ 1

8𝐡𝑋 . By throwing

away part of the integral,

1

πœƒ

∫ πœƒ

0

∫R𝑛

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯+ π‘ π‘Žβˆ’ 𝑦)β€–π‘Œ 𝑃𝑅𝑑(𝑦)𝑑𝑦 (3.20)

β‰₯ 1

πœƒ

∫ πœƒ

0

∫(1/8)𝐡𝑋

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯βˆ’ 𝑦 + π‘ π‘Ž)β€–π‘Œ 𝑃𝑅𝑑(𝑦) 𝑑𝑠 𝑑𝑦 (3.21)

=∫(1/8)𝐡𝑋

(1

πœƒ

∫ πœƒ

0β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯βˆ’ 𝑦 + π‘ π‘Ž)β€–π‘Œ 𝑑𝑠

)𝑃𝑅𝑑(𝑦) 𝑑𝑦 by Fubini (3.22)

β‰₯ 1

𝐷

∫(1/8)𝐡𝑋

𝑃𝑅𝑑(𝑦)𝑑𝑦 by (3.19) (3.23)

β‰₯ 𝑐

𝐷(3.24)

for some constant 𝑐. The calculation above says that the average length in every directionis big, and if you average the average length again, it is still big. How do we get rid of theadditional averaging? We care about

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )β€– * 𝑃𝑅𝑑(π‘₯) =∫R𝑛

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯βˆ’ 𝑦)β€–π‘Œ 𝑃𝑅𝑑(𝑦)𝑑𝑦

=1

πœƒ

∫ πœƒ

0

�∫R𝑛

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯βˆ’ 𝑦 + π‘ π‘Ž)β€–π‘Œ 𝑃𝑅𝑑(𝑦)𝑑𝑦�

βˆ’ 1

πœƒ

∫ πœƒ

0

∫R𝑛

β€–πœ•π‘Ž(𝑃𝑑 * 𝐹 )(π‘₯βˆ’ 𝑦)β€–π‘Œ (𝑃𝑅𝑑(𝑦 + π‘ π‘Ž)βˆ’ 𝑃𝑅𝑑(𝑦))𝑑𝑦

>𝑐

π·βˆ’ 𝑐′

πœƒ

∫ πœƒ

0

∫R𝑛

|𝑃𝑅𝑑(𝑦 + π‘ π‘Ž)βˆ’ 𝑃𝑅𝑑(𝑦)|𝑑𝑠 𝑑𝑦 by (3.24) and what?

β‰₯ 𝑐

π·βˆ’ 𝑐′

πœƒ

∫ πœƒ

0

βˆšπ‘›π‘ β€–π‘Žβ€–2𝑅𝑑

𝑑𝑠 by Fact 2 (Pr. 3.3.7)

This error term is

𝑐′

πœƒ

∫ πœƒ

0

βˆšπ‘›π‘ β€–π‘Žβ€–2𝑅𝑑

𝑑𝑠 =𝑐𝐷𝑛 log

(3𝑑

�𝑅

= 𝑂(1

𝐷

)by (3.2), so β€–πœ•π‘Ž(𝑃𝑑*𝐹 )β€–*𝑃𝑅𝑑(π‘₯) = Ξ©

(1𝐷

οΏ½(if we chose constants correctly), as desired. Factor

ofβˆšπ‘›?

To summarize: You compute the error such that you have distance πœƒ. Then you lookat the average derivative on each line like this. Then, the average derivative is roughlyβ€–πœ•π‘Ž(𝑃𝑑 *𝐹 )β€–*𝑃𝑅𝑑(π‘₯). So averaging the derivative in a bigger ball is the same up to constantsas taking a random starting point and averaging over the ball. What’s written in the lemmais that the derivative of a given point in this direction, averaged over a little bit bigger ball,it’s not unreasonable to see that at first take starting point and going in direction πœƒ andaveraging in the bigger ball is equivalent to randomizing over starting points.

3-21: A better bound in Bourgain Discretization

65

Page 66: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

4 Upper bound on π›Ώπ‘‹β†’π‘Œ

We give an example where we need a discretization bound 𝛿 going to 0.

Theorem 3.4.1. There exist Banach spaces 𝑋, π‘Œ with dim 𝑋 = 𝑛 such that if 𝛿 < 1 and𝒩𝛿 is a 𝛿-net in the unit ball and πΆπ‘Œ (𝒩𝛿) & πΆπ‘Œ (𝑋), then 𝛿 . 1

𝑛.

Together with the lower bound we proved, this gives

1

πœŒπ‘›πΆπ‘›< π›Ώπ‘‹β†’π‘Œ (1/2) <

1

𝑛

The lower bound would be most interesting to improve.

The example will be 𝑋 = ℓ𝑛1 , π‘Œ = β„“2.

We need two ingredients.

Firstly, we show that (𝐿1,Γˆβ€–π‘₯βˆ’ 𝑦‖1) is isometric to a subset of 𝐿2. More generally, we

prove the following.

Lemma 3.4.2. Given a measure space (Ξ©, πœ‡), (𝐿1(πœ‡),Γˆβ€–π‘₯βˆ’ 𝑦‖1) is isometric to a subset

of 𝐿2(Ω× R, πœ‡Γ— πœ†), where πœ† is the Lesbegue measure of R.

Note all separable Hilbert spaces are the same, 𝐿2(Ω×R, πœ‡Γ— πœ†) is the same whenever itis separable.

Proof. First we define 𝑇 : 𝐿1(πœ‡) β†’ 𝐿2(πœ‡Γ— πœ†) as follows. For 𝑓 ∈ 𝐿1(πœ‡), let

𝑇𝑓(𝑀, π‘₯) =

⎧βŽͺβŽͺ⎨βŽͺβŽͺ⎩1 0 ≀ 𝑓(𝑀) ≀ π‘₯

βˆ’1 π‘₯ ≀ 𝑓(π‘₯) ≀ 0

0 otherwise

Consider two functions 𝑓1, 𝑓2 ∈ 𝐿1(πœ‡). The function |𝑇𝑓1 βˆ’ 𝑇𝑓2| is the indicator of thearea between the graph of 𝑓1 and the graph of 𝑓2.

66

Page 67: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Thus the 𝐿2 norm is, letting 1 be the signed indicator,

‖𝑇𝑓1 βˆ’ 𝑇𝑓2‖𝐿2(πœ‡Γ—πœ†) =

溽

∫R

(1𝑓1(𝑀),𝑓2(𝑀) (π‘₯)

2�𝑑π‘₯ π‘‘πœ‡

=

溽|𝑓1(𝑀)βˆ’ 𝑓2(𝑀)| π‘‘πœ‡(𝑀)

More generally, if π‘ž ≀ 𝑝 < ∞, then (πΏπ‘ž, β€–π‘₯ βˆ’ 𝑦‖𝑝/π‘žπ‘ž ) is isometric to a subset of 𝐿𝑝. Wemay prove this later if we need it.

The second ingredient is due to Enflo .

Theorem 3.4.3 (Enflo, 1969). The best possible embedding of the hypercube into Hilbertspace has distortion

βˆšπ‘›:

𝐢2({0, 1}𝑛, β€– Β· β€–1) =βˆšπ‘›

Note that we don’t just calculate 𝐢2 up to a constants here; we know it exactly. This isvery rare.

Proof. We need to show an upper bound and a lower bound.

1. To show the upper bound, we show the identity mapping {0, 1}𝑛 β†’ ℓ𝑛2 has distortionβˆšπ‘›. We have

β€–π‘₯βˆ’ 𝑦‖2 =(

π‘›βˆ‘π‘–=1

|π‘₯𝑖 βˆ’ 𝑦𝑖|2)1/2

=

(π‘›βˆ‘π‘–=1

|π‘₯𝑖 βˆ’ 𝑦𝑖|)1/2

=Γˆβ€–π‘₯βˆ’ 𝑦‖1,

since the values of π‘₯, 𝑦 are only 0, 1. Then

1βˆšπ‘›β€–π‘₯βˆ’ 𝑦‖1 ≀ β€–π‘₯βˆ’ 𝑦‖2 =

Γˆβ€–π‘₯βˆ’ 𝑦‖1 ≀ β€–π‘₯βˆ’ 𝑦‖1

Thus 𝐢2({0, 1}𝑛) β‰€βˆšπ‘›.

(So this theorem tells us that if you want to represent the boolean hypercube as aEuclidean geometry, nothing does better than the identity mapping.)

67

Page 68: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

2. For the lower bound 𝐢2({0, 1}𝑛, β€–Β·β€–1) β‰₯βˆšπ‘›, we use the following.

Lemma 3.4.4 (Enflo’s inequality). We claim that for every 𝑓 : {0, 1}𝑛 β†’ β„“2,βˆ‘π‘₯∈F𝑛2

‖𝑓(π‘₯)βˆ’ 𝑓(π‘₯+ 𝑒)β€–22 β‰€π‘›βˆ‘π‘—=1

βˆ‘π‘₯∈F𝑛2

‖𝑓(π‘₯+ 𝑒𝑗)βˆ’ 𝑓(π‘₯)β€–22,

where 𝑒 = (1, . . . , 1) and 𝑒𝑖 are standard basis vectors of R𝑛.

Here, addition is over F𝑛2 .

This is specific to β„“2. See the next subsection for the proof.

In other words, Hilbert space is of Enflo-type 2 (Definition 1.1.13): the sum of thesquares of the lengths of the diagonals is at most the sum of the squares of the lengthsof the edges.

Suppose1

𝐷‖π‘₯βˆ’ 𝑦‖1 ≀ ‖𝑓(π‘₯)βˆ’ 𝑓(𝑦)β€–2 ≀ β€–π‘₯βˆ’ 𝑦‖1.

Our goal is to show that 𝐷 β‰₯βˆšπ‘›. Letting 𝑦 = π‘₯+ 𝑒,

‖𝑓(π‘₯+ 𝑒)βˆ’ 𝑓(π‘₯)β€–2 β‰₯1

𝐷‖(π‘₯+ 𝑒)βˆ’ π‘₯β€–1 =

𝑛

𝐷.

Plugging into Enflo’s inequality 3.4.4, we have, since there are 2𝑛 points in the space,

2𝑛(𝑛/𝐷)2 ≀ 𝑛 Β· 2𝑛 Β· 12 =β‡’ 𝐷2 β‰₯ 𝑛

as desired.

Proof of Theorem 3.4.1. Let 𝒩𝛿 be a 𝛿-net in 𝐡ℓ𝑛1. Let 𝑇 : ℓ𝑛1 β†’ 𝐿2 satisfy ‖𝑇 (π‘₯)βˆ’π‘‡ (𝑦)β€–2 =È

β€–π‘₯βˆ’ 𝑦‖1. 𝑇 restricted to 𝒩𝛿 has distortion β‰€Γˆ

2𝛿because for π‘₯, 𝑦 ∈ 𝒩𝛿 with π‘₯ = 𝑦, we

have1√2β€–π‘₯βˆ’ 𝑦‖1 ≀

Γˆβ€–π‘₯βˆ’ 𝑦‖1 ≀

1βˆšπ›Ώβ€–π‘₯βˆ’ 𝑦‖1.

However, the distortion of ℓ𝑛1 in β„“2 isβˆšπ‘›. The condition πΆπ‘Œ (𝒩𝛿) & πΆπ‘Œ (𝑋) means that for

some constant 𝐾, Ê2

𝛿β‰₯ 𝐾

βˆšπ‘› =β‡’ 𝛿 ≀ 2

𝐾2𝑛.

68

Page 69: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

4.1 Fourier analysis on the Boolean cube and Enflo’s inequality

We can prove this by induction on 𝑛 (exercise). Instead, I will show a proof that is Fourieranalytic and which generalizes to other situations.

Definition 3.4.5 (Walsh function): Let 𝐴 βŠ† {1, . . . , 𝑛}. Define the Walsh function

π‘Šπ΄ : F𝑛2 β†’ {Β±1}

byπ‘Šπ΄(π‘₯) = (βˆ’1)π‘—βˆˆπ΄.

Proposition 3.4.6 (Orthonormality): For 𝐴,𝐡 βŠ† {1, . . . , 𝑛}.

Eπ‘₯∈F𝑛2

π‘Šπ΄(π‘₯)π‘Šπ΅(π‘₯) = 𝛿𝐴𝐡.

Thus, {π‘Šπ΄}π΄βŠ†{1,...,𝑛} is an orthonormal basis of 𝐿2(F𝑛2 ).

Proof. Without loss of generality 𝑗 ∈ π΄βˆ–π΅. Then

E (βˆ’1)βˆ‘

π‘—βˆˆπ΄π‘₯𝑗+βˆ‘

π‘—βˆˆπ΅π‘₯𝑗 = 0.

Corollary 3.4.7. Let 𝑓 : F𝑛2 β†’ 𝑋 where 𝑋 is a Banach space. Then

𝑓 =βˆ‘

π΄βŠ†{1,Β·Β·Β· ,𝑛}𝑓(π‘Ž)π‘Šπ΄

where 𝑓(𝐴) = 12𝑛

βˆ‘π‘₯∈F𝑛2 𝑓(π‘₯)(βˆ’1)

βˆ‘π‘—βˆˆπ΄

π‘₯𝑗 .

Proof. How do we prove two vectors are the same in a Banach space? It is enough to showthat any composition with a linear functional gives the same value; thus it suffices to provethe claim for 𝑋 = R. The case 𝑋 = R holds by orthonormality.

Proof of Lemma 3.4.4. It is sufficient to prove the claim for 𝑋 = R; we get the inequalityfor β„“2 by summing coordinates. (Note this is a luxury specific to β„“2.)

The summand of the LHS of the inequality is

𝑓(π‘₯)βˆ’ 𝑓(π‘₯+ 𝑒) =βˆ‘

π΄βŠ†{1,...,𝑛}𝑓(𝐴) (π‘Šπ΄(π‘₯)βˆ’π‘Šπ΄(π‘₯+ 𝑒))

=βˆ‘

π΄βŠ†{1,...,𝑛}𝑓(𝐴)π‘Šπ΄(π‘₯)

(1βˆ’ (βˆ’1)|𝐴|

οΏ½=

βˆ‘π΄βŠ†{1,...,𝑛},|𝐴| odd

2𝑓(𝐴)π‘Šπ΄(π‘₯)

69

Page 70: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

The summand of the RHS of the inequality is

βˆ’π‘“(π‘₯+ 𝑒𝑗) + 𝑓(π‘₯) =βˆ‘

π΄βŠ†{1,...,𝑛}𝑓(𝐴)(π‘Šπ΄(π‘₯)βˆ’π‘Šπ΄(π‘₯+ 𝑒𝑗))

=βˆ‘

π΄βŠ†{1,...,𝑛},π‘—βˆˆπ΄2𝑓(𝐴)π‘Šπ΄(π‘₯).

Summing gives

βˆ‘π‘₯∈F𝑛2

(𝑓(π‘₯)βˆ’ 𝑓(π‘₯+ 𝑒))2 = 2𝑛

βˆ‘|𝐴| odd

2𝑓(𝐴)π‘Šπ΄

2

𝐿2(F𝑛2 )

= 2π‘›βˆ‘

|𝐴| odd4(𝑓(𝐴))2

π‘›βˆ‘π‘—=1

βˆ‘π‘₯∈F𝑛2

(𝑓(π‘₯+ 𝑒𝑗)βˆ’ 𝑓(π‘₯))2 =π‘›βˆ‘π‘—=1

2π‘›βˆ‘π΄:π‘—βˆˆπ΄

4(𝑓(𝐴)

)2= 2𝑛

βˆ‘π΄

βˆ‘π‘—βˆˆπ΄

4𝑓(𝐴)2

= 2π‘›βˆ‘π΄

4|𝐴|𝑓(𝐴)2

From this we see that the claim is equivalent toβˆ‘π΄βŠ†{1,...,𝑛},|𝐴| odd

𝑓(𝐴)2 β‰€βˆ‘

π΄βŠ†{1,...,𝑛}|𝐴|𝑓(𝐴)2

which is trivially true.

This inequality invites improvements, as this is a large gap. The Fourier analytic proofmade this gap clear.

5 Improvement for 𝐿𝑝

Theorem 3.5.1. thm:bourgain-lp Suppose 𝑝 β‰₯ 1 and 𝑋 is an 𝑛-dimensional normed space. Then

𝛿𝑋 →˓𝐿𝑝(πœ€) &πœ–2

𝑛5/2

In the world of embeddings into 𝐿𝑝, we are in the right ballpark for the value of 𝛿—weknow it is a power of 𝑛, if not exactly 1/𝑛. (It is an open question what the right power is.)This is a special case of a more general theorem, which I won’t state. This proof doesn’t usemany properties of 𝐿𝑝.

Proof. This follows from Theorem 3.5.2 and Theorem 3.5.3.

We will prove the theorem for πœ€ = 12. The proof needs to be modified for the πœ€-version.

70

Page 71: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Theorem 3.5.2. thm:bourgain-lp2 Let 𝑋, π‘Œ be Banach spaces, with dim𝑋 = 𝑛 < ∞. Supposethat there exists a universal constant 𝑐 such that 𝛿 ≀ π‘πœ€2

𝑛5/2 , and that 𝒩𝛿 is a 𝛿-net of 𝐡𝑋 .3

Then there exists a separable probability space (Ξ©, πœ‡)4, a finite dimensional subspace 𝑍 βŠ† π‘Œand a linear operator 𝑇 : 𝑋 β†’ 𝐿∞(πœ‡, 𝑍) such that for every π‘₯ ∈ 𝑋 with β€–π‘₯‖𝑋 = 1,

1βˆ’ πœ€

π·β‰€βˆ«Ξ©β€–π‘‡π‘₯(πœ”)β€–π‘Œ π‘‘πœ‡(πœ”) ≀ esssupπœ”βˆˆΞ©β€–(𝑇π‘₯)πœ”β€–π‘Œ ≀ 1 + πœ–

Note that the inequality can also be written as

1βˆ’ πœ€

𝐷‖𝑇π‘₯‖𝐿1(πœ‡,𝑍) ≀ ‖𝑇π‘₯β€–πΏβˆž(πœ‡,𝑍).

Now what happens when π‘Œ = 𝐿𝑝([0, 1])? We have

𝐿𝑝(πœ‡, 𝑍) βŠ† 𝐿𝑝(πœ‡, π‘Œ ) = 𝐿𝑝(πœ‡, 𝐿𝑝) = 𝐿𝑝(πœ‡Γ— πœ†)

A different way to think of this is that a function 𝑓 : 𝑋 β†’ 𝐿∞(πœ‡, 𝐿𝑝([0, 1])) can be thoughtof as a function of π‘₯ ∈ 𝑋 and πœ” ∈ [0, 1], 𝑓(πœ”, π‘₯) : Ω× [0, 1] β†’ R.

This is a very classical fact in measure theory:

Theorem 3.5.3 (Kolmogorov’s representation theorem). Any separable 𝐿𝑝(πœ‡) space is iso-metric to a subset of 𝐿𝑝.

This is an immediate corollary of the fact that if I give you a separable probability space,there is a separable transformation of the probability space (Ξ©, πœ‡) ∼= ([0, 1], πœ†) where πœ† isLebesgue measure. So there is a measure preserving isomorphism if the probability measureis atom-free. In general, the possible cases are

βˆ™ (Ξ©, πœ‡) ∼= ([0, 1], πœ†) if it is atom free.

βˆ™ (Ξ©, πœ‡) ∼= [0, 1]Γ— {1, . . . , 𝑛} if there are finitely many (𝑛) atoms

βˆ™ (Ξ©, πœ‡) ∼= [0, 1]Γ— N if there are countably many atoms,

βˆ™ (Ξ©, πœ‡) ∼= {1, . . . , 𝑛} or N if it’s completely atomic.

See Halmos’s book for deails. This is just called Kolmogorov’s representation theorem.Now for this to be an improvement of Bourgain’s discretization theorem for 𝐿𝑝, we need

that if 𝑍 βŠ‚ π‘Œ is a finite dimensional subspace and π‘Š βŠ‚ 𝐿∞(πœ‡, 𝑍) is a finite dimensionalsubspace such that on π‘Š , the 𝐿∞ and 𝐿1 norms are equivalent (a very strong restriction),then π‘Š embeds in π‘Œ . That’s the only property of 𝐿𝑝 we’re using.

However, not every π‘Œ has this property.

3Let me briefly say something about vector-valued 𝐿𝑝 spaces. Whenever you write 𝐿𝑝(πœ‡,𝑍), this equals

all functions 𝑓 : Ξ© β†’ 𝑍 such that (βˆ«β€–π‘“(πœ”)β€–π‘π‘π‘‘πœ‡(πœ”))

1/𝑝< ∞ (our norm is bounded).

4it will in the end be a uniform measure on half the ball

71

Page 72: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Theorem 3.5.4 (Ostrovskii and Randrianantoanina). Not every π‘Œ has this property.

You can construct spaces of functions taking values in a finite dimensional subspace ofit which cannot be embedded back into π‘Œ . Nevertheless, you have to work to find badcounterexamples. This is not published yet.

Next class we’ll prove Theorem 3.5.2, which implies our first theorem. What is theadvantage of formulating the theorem in this way? We can use Bourgain’s almost-extentiontheorem along the ball. The function is already Lipschitz, so we can already differentitateit. The measure itself will be the unit ball. We’ll look at the derivative of the extendedfunction in direction πœ” at point π‘₯. That is what our 𝑇 will be. We will look in all possibledirections. Then the lower bound we have here says that the derivative of the function isonly invertible on average, not at every point in the sphere. That tends to be big on average.We will work for that, but you see the advantage: In this formulation, we’re going to get anaverage lower-bound, not an always lower-bound, and in this way we will be able to shaveoff two exponents. But this is not for Banach spaces in general. The advantage we have hereis that 𝐿𝑝 of 𝐿𝑝 is 𝐿𝑝.

3/23: We will prove the improvement of Bourgain’s Discretization Theorem.

Proof of Theorem 3.5.2. There exists 𝑓 : 𝒩𝛿 β†’ π‘Œ such that

1

𝐷‖π‘₯βˆ’ π‘¦β€–π‘Œ ≀ ‖𝑓(π‘₯)βˆ’ 𝑓(𝑦)β€–π‘Œ ≀ β€–π‘₯βˆ’ 𝑦‖𝑋

for π‘₯, 𝑦 ∈ 𝒩𝛿. By Bourgain’s almost extension theorem, letting 𝑍 = span(𝑓(𝒩𝛿)), thereexists 𝐹 : 𝑋 β†’ 𝑍 such that

1. ‖𝐹‖𝐿𝑝. 1.

2. For every π‘₯ ∈ 𝒩𝛿, ‖𝐹 (π‘₯)βˆ’ 𝑓(π‘₯)β€– ≀ 𝑛𝛿.

3. 𝐹 is smooth.

Let Ξ© = 12𝐡𝑋 , with πœ‡ the normalized Lebesgue measure on Ξ©:

πœ‡(𝐴) =vol(𝐴 ∩

(12𝐡𝑋

οΏ½)

vol(12𝐡𝑋)

.

Define 𝑇 : 𝑋 β†’ 𝐿∞(πœ‡, 𝑍) as follows. For all 𝑦 ∈ 𝑋,

(𝑇𝑦)(π‘₯) = 𝐹 β€²(π‘₯)(𝑦) = πœ•π‘¦πΉ (π‘₯).

This function is in 𝐿∞ because

β€–(𝑇𝑦)(π‘₯)β€– = ‖𝐹 β€²(π‘₯)(𝑦)β€– ≀ ‖𝐹 β€²(π‘₯)β€– ‖𝑦‖ . ‖𝑦‖ .

This proves the upper bound.

72

Page 73: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

We need

1

𝐷‖𝑦‖ . ‖𝑇𝑦‖𝐿1(πœ‡,𝑍)

=∫

12𝐡𝑋

‖𝐹 β€²(π‘₯)𝑦‖ π‘‘πœ‡(π‘₯)

=1

vol(12𝐡𝑋)

∫12𝐡𝑋

‖𝐹 β€²(π‘₯)𝑦‖ 𝑑π‘₯.

We need two simple and general lemmas.

Lemma 3.5.5. There exists an isometric embedding 𝐽 : 𝑋 β†’ β„“βˆž.

This is a restatement of the Hahn-Banach theorem.

Proof. 𝑋* is separable, so let {π‘₯*π‘˜}βˆžπ‘˜=1 be a set of linear functionals that is dense in 𝑆𝑋* =πœ•π΅π‘‹* . Let

𝐽𝑋 = (π‘₯*π‘˜(π‘₯))βˆžπ‘˜=1 ∈ β„“βˆž.

For all π‘₯, |π‘₯*π‘˜(π‘₯)| ≀ β€–π‘₯*π‘˜β€– ≀ β€–π‘₯*π‘˜β€– β€–π‘₯β€– = β€–π‘₯β€–. By Hahn Banach, since every π‘₯ admits anormalizing functional,

‖𝐽π‘₯β€–β„“βˆž = supπ‘˜

|π‘₯*π‘˜(π‘₯)| = supπ‘₯*βˆˆπ‘†π‘‹

|π‘₯*(π‘₯)| = β€–π‘₯β€– .

This may see a triviality, but there are 4 situations where this theorem will make itsappearance.

Remark 3.5.6: Every separable metric space (𝑋, 𝑑) admits an isometric embedding intoβ„“βˆž, given as follows. Take {π‘₯π‘˜}βˆžπ‘˜=1 dense in 𝑋, and let

𝑓(π‘₯) = (𝑑(π‘₯, π‘₯π‘˜)βˆ’ 𝑑(π‘₯, π‘₯0))βˆžπ‘˜=1.

Proof is left as an exercise.

Lemma 3.5.7 (Nonlinear Hahn-Banach Theorem). Let (𝑋, 𝑑) be any metric space and 𝐴 βŠ†π‘‹ any nonempty subset. If 𝑓 : 𝐴 β†’ R is 𝐿-Lipschitz then there exists 𝐹 : 𝑋 β†’ R thatextends 𝑓 and

‖𝐹‖Lip = 𝐿 = ‖𝑓‖Lip .

The lemma says we can always extend the function and not lose anything in the Lips-chitz constant. Recall the Hahn-Banach Theorem says that we can extend functionals fromsubspace and preserve the norm. Extension in vector spaces and extension in R are different.

This lemma seems general, but it is very useful.

We mimic the proof of the Hahn-Banach Theorem.

73

Page 74: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. We will prove that one can extend 𝑓 to one additional point while preserving theLipschitz constant.

Then use Zorn’s Lemma on the poset of all extensions to supersets ordered by inclusion(with consistency) to finish.

Let 𝐴 βŠ† 𝑋, π‘₯ ∈ π‘‹βˆ–π΄. We define 𝑑 = 𝐹 (π‘₯) ∈ R. To make 𝐹 is Lipschitz, for all π‘Ž ∈ 𝐴,we need

|π‘‘βˆ’ 𝑓(π‘Ž)| ≀ 𝑑𝑋(π‘₯, π‘Ž).

Then we need𝑑 ∈

β‹‚π‘Žβˆˆπ΄

[𝑓(π‘Ž)βˆ’ 𝑑𝑋(π‘₯, π‘Ž), 𝑓(π‘Ž) + 𝑑𝑋(π‘₯, π‘Ž)].

The existence of 𝑑 is equivalent toβ‹‚π‘Žβˆˆπ΄

[𝑓(π‘Ž)βˆ’ 𝑑𝑋(π‘₯, π‘Ž), 𝑓(π‘Ž) + 𝑑𝑋(π‘₯, π‘Ž)].

By compactness it is enough to check this is true for finite intersections. Because we are onthe real line (which is a total order), it is in fact enough to check this is true for pairwiseintersections.

[𝑓(π‘Ž)βˆ’ 𝑑𝑋(π‘₯, π‘Ž), 𝑓(π‘Ž) + 𝑑𝑋(π‘₯, π‘Ž)] ∩ [𝑓(𝑏)βˆ’ 𝑑𝑋(π‘₯, 𝑏), 𝑓(𝑏) + 𝑑𝑋(π‘₯, 𝑏)].

We need to check

𝑓(π‘Ž) + 𝑑(π‘₯, π‘Ž) β‰₯ 𝑓(𝑏)βˆ’ 𝑑(π‘₯, 𝑏)

⇐⇒ |𝑓(𝑏)βˆ’ 𝑓(π‘Ž)| ≀ 𝑑(π‘₯, π‘Ž) + 𝑑(π‘₯, 𝑏).

This is true by the triangle inequality:

|𝑓(𝑏)βˆ’ 𝑓(π‘Ž)| ≀ 𝑑(π‘Ž, 𝑏) ≀ 𝑑(π‘₯, π‘Ž) + 𝑑(π‘₯, 𝑏).

This theorem is used all over the place. Most of the time people just give a closed formula:define 𝐹 on all the points at once by

𝐹 (π‘₯) = inf {𝑓(π‘Ž) + 𝑑(π‘₯, π‘Ž) : π‘Ž ∈ 𝐴} .

Now just check that this satisfies Lipschitz. From our proof you ses exactly why they define𝐹 this way: it comes from taking the inf of the upper intervals (we can also take the sup ofthe lower intervals). The extension is not unique; there are many formulas for the extension.

There is a whole isometric theory about exact extensions: look at all the possible ex-tensions, what is the best one? 𝐹 is the pointwise smallest extension; the other one is thepointwise largest. The theory of absolutely minimizing extensions is related to an interestingnonlinear PDE called the infinite Laplacian. This is different from what we’re doing becausewe lose constants in many places.

74

Page 75: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Corollary 3.5.8. Let (𝑋, 𝑑) be any metric space. Let 𝐴 βŠ† 𝑋 be a nonempty subset 𝑓 : π΄β†’β„“βˆž Lipschitz. Then there exists 𝐹 : 𝑋 β†’ β„“βˆž that extends 𝑓 and ‖𝐹‖Lip = ‖𝑓‖Lip.

Proof. Note that saying 𝑓 : 𝐴 β†’ β„“βˆž, 𝑓(π‘Ž) = (𝑓1(π‘Ž), 𝑓2(π‘Ž), . . .) has ‖𝑓‖Lip = 1 means that 𝑓𝑖is 1-Lipschitz for every 𝑖.

𝑓 takes 𝒩𝛿 βŠ† 𝐡𝑋 to 𝑓(𝒩𝛿) βŠ† 𝑍. π‘“βˆ’1 takes it back to 𝑋; 𝐽 is the embedding to β„“βˆž.

𝐡𝑋 𝑍 𝐺

οΏ½οΏ½

𝒩𝛿

?οΏ½

OO

𝑓// 𝑓(𝒩𝛿)?οΏ½

OO

π‘“βˆ’1|𝑓(𝒩𝛿)//

π½βˆ˜π‘“βˆ’1|𝑓(𝒩𝛿)

<<𝑋

𝐽 // β„“βˆž

Note 𝐽 ∘ π‘“βˆ’1|𝑓(𝒩𝛿) is 𝐷-Lipschitz. By nonlinear Hahn-Banach, there exists 𝐺 : 𝑍 β†’ β„“βˆž with‖𝐺‖Lip ≀ 𝐷 such that

𝐺(𝑓(π‘₯)) = 𝐽(π‘₯)

for all π‘₯ ∈ 𝒩𝛿.

We want 𝐺 to be differentiable. We do this by convolving with a smooth bump functionwith small support to get 𝐻 : 𝑍 β†’ β„“βˆž such that

1. 𝐻 is smooth

2. ‖𝐻‖Lip ≀ 𝐷,

3. for all π‘₯ ∈ 𝐹 (𝐡𝑋), ‖𝐺(π‘₯)βˆ’π»(π‘₯)β€– ≀ 𝑛𝐷𝛿.

Define a linear operator

𝑆 : 𝐿1(πœ‡, 𝑍) β†’ β„“βˆž

by for all β„Ž ∈ 𝐿1(πœ‡, 𝑍), β„Ž : 12𝐡𝑋 β†’ 𝑍,

π‘†β„Ž =∫

12𝐡𝑋

𝐻 β€²(𝐹 (π‘₯))(β„Ž(π‘₯)) π‘‘πœ‡(π‘₯).

(Type checking: F is a point in 𝑋, we can differentiate at this point and get linear operator𝑍 β†’ β„“βˆž. β„Ž(π‘₯) is a point in 𝑍, so we can plug into linear operator and get a point in β„“βˆž.This is a vector-valued integration; do the integration coordinatewise. So π‘†β„Ž is in β„“βˆž.)

75

Page 76: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Now

β€–π‘†β„Žβ€–β„“βˆž β‰€βˆ«

12𝐡𝑋

‖𝐻 β€²(𝐹 (π‘₯))(β„Ž(π‘₯))β€–β„“βˆž π‘‘πœ‡(π‘₯)

β‰€βˆ«

12𝐡𝑋

‖𝐻 β€²(𝐹 (π‘₯))β€–π‘β†’β„“βˆžβŸ ⏞ =:𝐷

β€–β„Ž(π‘₯)β€– π‘‘πœ‡(π‘₯)

≀ 𝐷∫

12𝐡𝑋

β€–β„Ž(π‘₯)β€–

≀ 𝐷 β€–β„Žβ€–πΏ1(πœ‡,𝑍)

=β‡’ ‖𝑆‖𝐿1(πœ‡,𝑍)β†’β„“βˆžβ‰€ 𝐷

𝑆 π‘‡π‘¦βŸ ⏞ β„Ž

=∫

12𝐡𝑋

𝐻 β€²(𝐹 (π‘₯))((𝑇𝑦)(π‘₯)) π‘‘πœ‡(π‘₯)

=∫

12𝐡𝑋

𝐻 β€²(𝐹 (π‘₯))(𝐹 β€²(π‘₯)(𝑦)) π‘‘πœ‡(π‘₯)

=∫

12𝐡𝑋

(𝐻 ∘ 𝐹 )β€²(π‘₯)(𝑦) π‘‘πœ‡(π‘₯) chain rule

Now we show 𝐻 ∘ 𝐹 is very close to 𝐽 ; it’s close to being invertible. (Recall 𝐺(𝑓(π‘₯)) = 𝐽π‘₯.)This does not a priori mean the derivative is close. We use a geometric argument to say

that if a function is sufficiently close to being an isometry, then the integral of its derivativeon a finite dimensional space is close to being invertible. 𝐻 ∘ 𝐹 was a proxy to 𝐽 .

Check that 𝐻 ∘ 𝐹 is close to 𝐽 . For 𝑦 ∈ 𝒩𝛿, by choice of 𝐻,

‖𝐻(𝐹 (𝑦))βˆ’ π½π‘¦β€–β„“βˆž ≀ ‖𝐻(𝐹 (𝑦))βˆ’πΊ(𝐹 (𝑦))β€–β„“βˆž + ‖𝐺(𝐹 (𝑦))βˆ’πΊ(𝑓(𝑦))β€–β„“βˆžβ‰€ 𝑛𝐷𝛿 +𝐷 ‖𝐹 (𝑦)βˆ’ 𝑓(𝑦)‖𝑍. 𝑛𝐷𝛿.

For general π‘₯ ∈ 12𝐡𝑋 , there exists 𝑦 ∈ 𝒩𝛿 such that β€–π‘₯βˆ’ 𝑦‖𝑋 ≀ 2𝛿. Then

‖𝐻(𝐹 (π‘₯))βˆ’ 𝐽π‘₯β€–β„“βˆž ≀ 𝑛𝐷𝛿 +𝐷𝛿 + 𝛿

. 𝑛𝐷𝛿.

For all π‘₯ ∈ 12𝐡𝑋 , ‖𝐻 ∘ 𝐹 (π‘₯)βˆ’ 𝐽π‘₯β€– ≀ 𝐢𝑛𝐷𝛿. Define 𝑔(π‘₯) = 𝐻 ∘ 𝐹 (π‘₯) βˆ’ 𝐽π‘₯. Then

β€–π‘”β€–πΏβˆž( 12𝐡𝑋) ≀ 𝐢𝑛𝐷𝛿.

We need a geometric lemma.

Lemma 3.5.9. Suppose (𝑉, ‖·‖𝑉 ) is any Banach space, π‘ˆ = (R𝑛, β€–Β·β€–π‘ˆ) is a 𝑛-dimensionalBanach space. Let 𝑔 : π΅π‘ˆ β†’ 𝑉 be continuous on π΅π‘ˆ and differentiable on int(π΅π‘ˆ). Then

1

vol(π΅π‘ˆ)

βˆ«π΅π‘ˆ

𝑔′(𝑒) 𝑑𝑒

π‘ˆβ†’π‘‰

≀ 𝑛 β€–π‘”β€–πΏβˆž(π΅π‘ˆ ) .

76

Page 77: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

3/28: Finishing up Bourgain’s TheoremPreviously we had 𝑋, π‘Œ Banach spaces with dim(𝑋) = 𝑛. Let 𝑁𝛿 βŠ† 𝐡𝑋 be a 𝛿-net,

𝛿 ≀ 𝑐/𝑛2𝐷, 𝑓 : 𝑁𝛿 β†’ π‘Œ . We have

1

𝐷‖π‘₯βˆ’ 𝑦‖ ≀ ‖𝑓(π‘₯)βˆ’ 𝑓(𝑦)β€– ≀ β€–π‘₯βˆ’ 𝑦‖

WLOG π‘Œ = span(𝑓(𝑁𝛿)). The almost extension theorem gave us a smooth map from𝐹 : 𝑋 β†’ π‘Œ , ‖𝐹‖𝐿𝑖𝑝 . 1 and for all π‘₯ ∈ 𝑁𝛿, ‖𝐹 (π‘₯) βˆ’ 𝑓(π‘₯)β€– . 𝑛𝛿. Letting πœ‡ be thenormalized Lebesgue measure on 1/2𝐡𝑋 , define 𝑇 : 𝑋 β†’ 𝐿∞(πœ‡, π‘Œ ) by

(𝑇𝑦)(π‘₯) = 𝐹 β€²(π‘₯)𝑦

for all 𝑦 ∈ 𝑋.The goal was to prove

‖𝑇𝑦‖𝐿1(πœ‡,π‘Œ ) &1

𝐷‖𝑦‖

and the approach was to show for 𝑦 ∈ 𝑋 the average 1vol(1/2𝐡𝑋)

∫1/2𝐡𝑋

‖𝐹 β€²(π‘₯)𝑦‖𝑑𝑦 is big.Fix a linear isometric embedding 𝐽 : 𝑋 β†’ π‘™βˆž., we proved that there exists 𝐺 : π‘Œ β†’ π‘™βˆž

such that βˆ€π‘₯ ∈ 𝑁𝛿, 𝐺(𝑓(π‘₯)) = 𝐽(π‘₯). We have ‖𝐺‖𝐿𝑖𝑝 ≀ 𝐷. Fix 𝐻 : π‘Œ β†’ π‘™βˆž where 𝐻 issmooth, ‖𝐻‖𝐿𝑖𝑝 ≀ 𝐷, and βˆ€π‘₯ ∈ 𝐹 (𝐡𝑋), ‖𝐻(π‘₯)βˆ’πΊ(π‘₯)β€– ≀ 𝑛𝐷𝛿.

Define 𝑆 : 𝐿1(πœ‡, π‘Œ ) β†’ π‘™βˆž by

π‘†β„Ž =∫1/2𝐡𝑋

𝐻 β€²(𝐹 (π‘₯))(β„Ž(π‘₯))π‘‘πœ‡(π‘₯)

for all β„Ž ∈ 𝐿1(πœ‡, π‘Œ ). Note that ‖𝑆‖𝐿1(πœ‡,π‘Œ )β†’π‘™βˆž ≀ 𝐷. By the chain rule, 𝑆𝑇𝑦 =∫1/2𝐡𝑋

(𝐻 ∘𝐹 )β€²(π‘₯)π‘¦π‘‘πœ‡(π‘₯). We checked for every π‘₯ ∈ 𝐡𝑋 , ‖𝐻(𝐹 (π‘₯))βˆ’ 𝐽π‘₯β€–π‘™βˆž] . 𝑛𝐷𝛿.

Suppose you are on a point on the net. Then you know 𝐻 is very close to 𝐺. 𝐹 was 𝑛𝛿close to 𝑓 , and 𝐻 is 𝐷-Lipschitz, which is how we get 𝑛𝛿𝐷 (take a net point, find the closestpoint on the net, and use the fact that the functions are 𝐷-Lipschitz).

This is where we stopped. Now how do we finish?We need a small geometric lemma:

Lemma 3.5.10. Geometric Lemma.π‘ˆ, 𝑉 are Banach spaces, dim(π‘ˆ) = 𝑛. Let π‘ˆ = (R𝑛, β€– Β· β€–π‘ˆ). Suppose that 𝑔 : π΅π‘ˆ β†’ 𝑉 iscontinuous on π΅π‘ˆ and smooth on the interior. Then

1

vol(π΅π‘ˆ)

βˆ«π΅π‘ˆ

𝑔′(𝑒)𝑑𝑒

π‘ˆβ†’π‘‰

≀ π‘›β€–π‘”β€–πΏβˆž(𝑆𝑋)

where the inside of the LHS norm should be thought of as an operator. Let 𝑅 be thenormalized integral operator. Then

𝑅π‘₯ =1

vol(π΅π‘ˆ)

βˆ«π΅π‘ˆ

𝑔′(𝑒)π‘₯𝑑𝑒

77

Page 78: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

The way to think of using this is if you have an 𝐿∞ bound, you get an 𝐿1 bound.Assuming this lemma, let’s apply it to 𝑔 = 𝐻 ∘ 𝐹 βˆ’ 𝐽 . We get∫

1/2𝐡𝑋

((𝐻 ∘ 𝐹 )β€² βˆ’ 𝐽) π‘‘πœ‡(π‘₯)π‘‹β†’π‘™βˆž

. 𝑛2π·π›Ώβˆ«1/2𝐡𝑋

(𝐻 ∘ 𝐹 )β€²(π‘₯)π‘‘πœ‡(π‘₯)βˆ’ 𝐽

π‘‹β†’π‘™βˆž

= ‖𝑆𝑇 βˆ’ π½β€–π‘‹β†’π‘™βˆž . 𝑛2𝐷𝛿

Now we want to bound from below ‖𝑇𝑦‖𝐿1(πœ‡,π‘Œ ). We have

‖𝑇𝑦‖𝐿1(πœ‡,π‘Œ ) β‰₯β€–π‘†π‘‡π‘¦β€–π‘™βˆž

‖𝑆‖𝐿1(πœ‡,π‘Œ )β†’π‘™βˆž

β‰₯ 1

π·β€–π‘†π‘‡π‘¦β€–π‘™βˆž β‰₯ 1

𝐷(‖𝐽𝑦‖ βˆ’ ‖𝑆𝑇 βˆ’ π½β€–π‘‹β†’π‘™βˆžβ€–π‘¦β€–)

=1

𝐷

(‖𝑦‖ βˆ’ (𝑛2𝐷𝛿‖𝑦‖)

οΏ½There is a bit of magic in the punchline. We want to bound the operator below, and weunderstand it as an average of derivatives. We succeeded to show the function itself is small,and there is our geometric lemma which gives a bound on the derivative if you know a boundon the function.

Now let’s prove the lemma.

Proof of Lemma 3.5.10. Fix a direction 𝑦 ∈ R𝑛 and normalize so that ‖𝑦‖2 = 1. For every𝑒 ∈ Proj𝑦βŠ₯(π΅π‘ˆ), let π‘Žπ‘ˆ ≀ π‘π‘ˆ ∈ R be such that 𝑒 + R𝑦 ∩ π΅π‘ˆ = 𝑒 + [π‘Žπ‘ˆ , π‘π‘ˆ ]𝑦 (basically, thisis the intersection of the projection line with the ball).

Using Fubini,1

vol(π΅π‘ˆ)

βˆ«π΅π‘ˆ

𝑔′(𝑒)𝑑𝑒

𝑉

=

1

vol(π΅π‘ˆ)

∫Proj

𝑦βŠ₯ (π΅π‘ˆ )

∫ π‘π‘ˆ

π‘Žπ‘ˆ

𝑑

𝑑𝑠𝑔(𝑒+ 𝑠𝑦)𝑑𝑠𝑑𝑒

𝑉

=

1

vol(π΅π‘ˆ)

∫Proj

𝑦βŠ₯ (π΅π‘ˆ )(𝑔(𝑒+ π‘π‘ˆπ‘¦)βˆ’ 𝑔(𝑒+ π‘Žπ‘ˆπ‘¦)) 𝑑𝑒

𝑉

≀ 1

vol𝑛(π΅π‘ˆ)Β· 2volπ‘›βˆ’1(Proj𝑦βŠ₯(π΅π‘ˆ))β€–π‘”β€–πΏβˆž(𝑆𝑋 ,𝑉 )

We need to show that2volπ‘›βˆ’1

(Proj𝑦βŠ₯(π΅π‘ˆ)

οΏ½vol𝑛(π΅π‘ˆ)

≀ π‘›β€–π‘¦β€–π‘ˆ

The convex hull of 𝑦 over the projection is the cone over the projection.

vol𝑛(conv

�𝑦

β€–π‘¦β€–π‘ˆβˆͺ Proj𝑦βŠ₯(π΅π‘ˆ)

οΏ½) =

1

π‘›β€–π‘¦β€–π‘ˆΒ· volπ‘›βˆ’1

(Proj𝑦βŠ₯(π΅π‘ˆ)

�Letting 𝐾 = conv

({Β± 𝑦

β€–π‘¦β€–π‘ˆ

}∩ Proj𝑦βŠ₯(π΅π‘ˆ)

), this is the same as saying that vol𝑛(𝐾) ≀

vol𝑛(π΅π‘ˆ). Note that 𝐾 is the double cone, that is where the factor of 2 got absorbed. Thisis an application of Fubini’s theorem.

78

Page 79: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Geometrically, the idea is that when we look on the projective lines for whatever part ofthe cone is not in our ball set, we will be able to fit the outside inside the set. In formulas,π‘π‘ˆ is the largest multiple of 𝑒 which is inside the boundary of the cone. π‘π‘ˆ β‰₯ 1.

𝐾 =⋃

π‘’βˆˆProj𝑦βŠ₯ (π΅π‘ˆ )

�𝑒+

οΏ½βˆ’ π‘π‘ˆ βˆ’ 1

π‘π‘ˆβ€–π‘¦β€–π‘ˆ,+

π‘π‘ˆ βˆ’ 1

π‘π‘ˆβ€–π‘¦β€–π‘ˆ

οΏ½οΏ½We also have

1

π‘π‘ˆ(π‘π‘ˆπ‘’+ π‘Žπ‘π‘ˆπ‘’π‘¦)Β± (1βˆ’ 1

π‘π‘ˆ)

𝑦

β€–π‘¦β€–π‘ˆβˆˆ π΅π‘ˆ

and we get that 𝐾 βŠ‚ π΅π‘ˆ by Fubini.

Thus, we’ve completed the proof of Bourgain’s theorem for the semester. The big re-maining question is can we do something like this in the general case?

6 Kirszbraun’s Extension Theorem

Last time we proved nonlinear Hahn-Banach theorem. I want to prove one more LipschitzExtension theorem, which I can do in ten minutes which we will need later in the semester.

Theorem 3.6.1 (Kirszbraun’s extension theorem (1934)). Let 𝐻1, 𝐻2 be Hilbert spaces and𝐴 βŠ† 𝐻1. Let 𝑓 : 𝐴→ 𝐻2 Lipschitz. Then, there exists a function 𝐹 : 𝐻1 β†’ 𝐻2 that extends𝑓 and has the same Lipschitz constant (‖𝐹‖Lip = ‖𝑓‖Lip).

We did this for real valued functions and π‘™βˆž functions. This version is non-trivial, andrelates to many open problems.

Proof. There is an equivalent geometric formulation. Let 𝐻1, 𝐻2 be Hilbert spaces {π‘₯𝑖}π‘–βˆˆπΌ βŠ†π»1 and {𝑦𝑖}π‘–βˆˆπΌ βŠ† 𝐻2, {π‘Ÿπ‘–}π‘–βˆˆπΌ βŠ† (0,∞). Suppose that βˆ€π‘–, 𝑗 ∈ 𝐼, β€–π‘¦π‘–βˆ’ 𝑦𝑗‖𝐻2 ≀ β€–π‘₯π‘–βˆ’ 𝑦𝑖‖𝐻1 . Ifβ‹‚

π‘–βˆˆπΌπ΅π»1(π‘₯𝑖, π‘Ÿπ‘–) = βˆ…

then β‹‚π‘–βˆˆπΌπ΅π»1(𝑦𝑖, π‘Ÿπ‘–) = βˆ…

as well.Intuitively, this says the configuration of points in 𝐻2 are a squeezed version of the 𝐻1

points. Then, we’re just saying something obvious. If there’s some point that intersects allballs in 𝐻1, then using the same radii in the squeezed version will also be nonempty. Ourgeometric formulation will imply extension. We have 𝑓 : 𝐴 β†’ 𝐻2, WLOG ‖𝑓‖𝐿𝑖𝑝 = 1. Forall π‘Ž, 𝑏 ∈ 𝐴, ‖𝑓(π‘Ž)βˆ’ 𝑓(𝑏)‖𝐻2 ≀ β€–π‘Žβˆ’ 𝑏‖𝐻1 . Fix any π‘₯ ∈ 𝐻1 βˆ– 𝐴. What can we say about theintersection of the following balls?:

β‹‚π‘Žβˆˆπ΄π΅π»1(π‘Ž, β€–π‘Žβˆ’ π‘₯‖𝐻1). Well by design it is not empty

since π‘₯ is in this set. So the conclusion from the geometric formulation saysβ‹‚π‘Žβˆˆπ΄

𝐡𝐻2 (𝑓(π‘Ž), β€–π‘Žβˆ’ π‘₯‖𝐻1) = βˆ…

79

Page 80: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

So take some 𝑦 in this set. Then β€–π‘¦βˆ’ 𝑓(π‘Ž)‖𝐻2 ≀ β€–π‘₯βˆ’ π‘Žβ€–π»2 βˆ€π‘Ž ∈ 𝐴. Define 𝐹 (π‘₯) = 𝑦. Thenwe can just do the one-more-point argument with Zorn’s lemma to finish.

Let us now prove the geometric formulation. It is enough to prove the geometric formu-lation when |𝐼| < ∞ and 𝐻1, 𝐻2 are finite dimensional. To show all the balls intersect, itis enough to show that finitely many of them intersect (this is the finite intersection prop-erty: balls are weakly compact). Now the minute 𝐼 is finite, we can write 𝐼 = {1, Β· Β· Β· , 𝑛},𝐻 β€²

1 = span{π‘₯1, Β· Β· Β· , π‘₯𝑛}, 𝐻 β€²2 = span{𝑦1, Β· Β· Β· , 𝑦𝑛}. This reduces everything to finite dimen-

sions. We have a nice argument using Hahn-Banach. FINISH THIS

Remark 3.6.2: In other norms, the geometric formulation in the previous proof is just nottrue. This effectively characterizes Hilbert spaces. Related is the Kneser-Poulsen conjecture,which effectively says the same thing in terms of volumes:

Conjecture 3.6.3 (Kneser-Poulsen). Let π‘₯1, . . . , π‘₯π‘˜; 𝑦1, . . . , π‘¦π‘˜ ∈ R𝑛 and β€–π‘¦π‘–βˆ’π‘¦π‘—β€– ≀ β€–π‘₯π‘–βˆ’π‘₯𝑗‖for all 𝑖, 𝑗. Then βˆ€π‘Ÿπ‘– β‰₯ 0

vol

(π‘˜β‹‚π‘–=1

𝐡(𝑦𝑖, π‘Ÿπ‘–)

)β‰₯ vol

(π‘˜β‹‚π‘–=1

𝐡(π‘₯𝑖, π‘Ÿπ‘–)

)This is known for 𝑛 = 2, where volume is area using trigonometry and all kinds of ad-hoc

arguments. It’s also known for π‘˜ = 𝑛+ 2.

3/30/16Here is an equivalent formulation of the Kirszbraun extension theorem.

Theorem 3.6.4. Let 𝐻1, 𝐻2 be Hilbert spaces, {π‘₯𝑖}π‘–βˆˆπΌ βŠ† 𝐻1, {𝑦𝑖}π‘–βˆˆπΌ βŠ† 𝐻2, {π‘Ÿπ‘–}βˆžπ‘–=1 βŠ† (0,∞).Suppose that ‖𝑦𝑖 βˆ’ 𝑦𝑗‖𝐻2

≀ β€–π‘₯𝑖 βˆ’ π‘₯𝑗‖𝐻2for all 𝑖, 𝑗 ∈ 𝐼. Then

β‹‚π‘–βˆˆπΌ 𝐡𝐻1(π‘₯𝑖, π‘Ÿπ‘–) = πœ‘ impliesβ‹‚

π‘–βˆˆπΌ 𝐡𝐻2(𝑦𝑖, π‘Ÿπ‘–) = πœ‘.

Proof. By weak compactness and orthogonal projection, it is enough to prove this when𝐼 = {1, . . . , 𝑛} and 𝐻1, 𝐻2 are both finite dimensional. Fix any π‘₯ ∈ ⋂𝑛

𝑖=1𝐡𝐻1(π‘₯𝑖, π‘Ÿπ‘–). Ifπ‘₯ = π‘₯𝑖0 for some 𝑖0 ∈ {1, . . . , 𝑛}, ‖𝑦𝑖0 βˆ’ 𝑦𝑖‖𝐻2

≀ β€–π‘₯𝑖0 βˆ’ π‘₯𝑖‖𝐻1≀ π‘Ÿπ‘–, and 𝑦𝑖0 ∈

⋂𝑛𝑖=1𝐡𝐻2(𝑦𝑖, π‘Ÿπ‘–).

Assume π‘₯ ∈ {π‘₯1, . . . , π‘₯𝑛}.Define 𝑓 : 𝐻 β†’ R by

𝑓(𝑦) = max

{‖𝑦 βˆ’ 𝑦1‖𝐻2

β€–π‘₯βˆ’ π‘₯1‖𝐻1

, . . . ,‖𝑦 βˆ’ 𝑦𝑛‖𝐻2

β€–π‘₯βˆ’ π‘₯𝑛‖𝐻2

}.

Note that 𝑓 is continuous and lim‖𝑦‖𝐻2β†’βˆž 𝑓(𝑦) = ∞.

So, the minimum of 𝑓(𝑦) over R𝑛 is attained. Let

π‘š = minπ‘¦βˆˆR𝑛

𝑓(𝑦).

Fix 𝑦 ∈ R𝑛, 𝑓(𝑦) = π‘š.

80

Page 81: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Observe that we are done if π‘š ≀ 1. Suppose by way of contradiction π‘š > 1. Let 𝐽 bethe indices where the ratio is the minimum.

𝐽 =

{𝑖 ∈ {1, . . . , 𝑛} :

‖𝑦 βˆ’ 𝑦𝑖‖𝐻2

β€–π‘₯βˆ’ π‘₯𝑖‖𝐻1

= π‘š

}.

By definition, if 𝑗 ∈ 𝐽 , then ‖𝑦 βˆ’ 𝑦𝑗‖𝐻2= π‘š β€–π‘₯βˆ’ π‘₯𝑗‖𝐻1

, and for 𝑗 ∈ 𝐽 , ‖𝑦 βˆ’ 𝑦𝑗‖𝐻2<

π‘š β€–π‘₯βˆ’ π‘₯𝑗‖𝐻1.

Claim 3.6.5. 𝑦 ∈ conv({𝑦𝑗}π‘—βˆˆπ½).

Proof. If not, find a separating hyperplane. If we move 𝑦 a small enough distance towardsconv({𝑦𝑗}π‘—βˆˆπ½) perpendicular to the separating hyperplane to 𝑦′, then for 𝑗 ∈ 𝐽 ,

‖𝑦′ βˆ’ 𝑦𝑗‖𝐻2< ‖𝑦 βˆ’ 𝑦𝑗‖𝐻2

= π‘š β€–π‘₯βˆ’ π‘₯𝑗‖𝐻1.

For 𝑗 ∈ 𝐽 , it is still true that ‖𝑦′ βˆ’ 𝑦𝑗‖𝐻2< π‘š β€–π‘₯βˆ’ π‘₯𝑗‖𝐻1

. Then 𝑓(𝑦′) < π‘š, contradictingthat 𝑦 is a minimizer.

By the claim, there exists {πœ†π‘—}π‘—βˆˆπ½ , πœ†π‘— β‰₯ 0,βˆ‘π‘—βˆˆπ½ πœ†π‘— = 1, 𝑦 =

βˆ‘π‘—βˆˆπ½ πœ†π‘—π‘¦π‘—.

We use this the coefficients of this convex combination to define a probability distribution.Define a random vector in 𝐻1 by 𝑋, with

P(𝑋 = π‘₯𝑗) = πœ†π‘—, 𝑗 ∈ 𝐽.

For all 𝑗 ∈ {1, . . . , 𝑛}, let 𝑦𝑗 = β„Ž(π‘₯𝑗). Let E[β„Ž(𝑋)] =βˆ‘π‘—βˆˆπ½ πœ†π‘—π‘¦π‘— = 𝑦. Let 𝑋 β€² be an

independent copy of 𝑋.Using β€–β„Ž(𝑋)βˆ’ 𝑦‖𝐻2

= π‘š ‖𝑋 βˆ’ π‘₯‖𝐻1, we have

E β€–β„Ž(𝑋)βˆ’ Eβ„Ž(𝑋)β€–2𝐻2= E β€–β„Ž(𝑋)βˆ’ 𝑦‖2𝐻2

(3.25)

= π‘š2E ‖𝑋 βˆ’ π‘₯β€–2𝐻1(3.26)

> E ‖𝑋 βˆ’ π‘₯β€–2𝐻1(3.27)

β‰₯ E ‖𝑋 βˆ’ 𝐸𝑋‖2𝐻1(3.28)

In Hilbert space, it is always true that

E ‖𝑋 βˆ’ 𝐸𝑋‖2𝐻1=

1

2E ‖𝑋 βˆ’π‘‹ β€²β€–2𝐻2

.

Using this on both sides of the above,

1

2E β€–β„Ž(𝑋)βˆ’ β„Ž(𝑋 β€²)β€–2𝐻2

>1

2‖𝑋 βˆ’π‘‹ β€²β€–2𝐻1

.

So far we haven’t used the only assumption on the points. By the assumption ‖𝑋 βˆ’π‘‹ ′‖𝐻1β‰₯

β€–β„Ž(𝑋)βˆ’ β„Ž(𝑋 β€²)‖𝐻2. This is a contradiction.

81

Page 82: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

This is not true in 𝐿𝑝 spaces when 𝑝 = 2, but there are other theorems one can formulate(ex. with different radii). We have to ask what happens to each of the inequalities in theproof. As stated the theorem is a very β€œHilbertian” phenomenon.

Definition 3.6.6: Let (𝑋, ‖·‖𝑋) be a Banach space and 𝑝 β‰₯ 1. We say that 𝑋 hasRademacher type 𝑝 if for every 𝑛, for every π‘₯1, . . . , π‘₯𝑛 ∈ 𝑋,(

Eπœ€=(πœ€1,...,πœ€π‘›)∈{Β±1}𝑛

[π‘›βˆ‘π‘–=1

πœ€π‘–π‘‹π‘–

𝑝𝑋

]) 1𝑝

≀ 𝑇

(π‘›βˆ‘π‘–=1

‖𝑋𝑖‖𝑝𝑋

) 1𝑝

.

The smallest 𝑇 is denoted 𝑇𝑝(𝑋).

Definition (Definition 1.1.13): A metric space (𝑋, 𝑑𝑋) is said to have Enflo type 𝑝 if forall 𝑓 : {Β±1}𝑛 β†’ 𝑋,

οΏ½Eπœ€[𝑑𝑋(𝑓(πœ€), 𝑓(βˆ’πœ€))𝑝]

οΏ½ 1𝑝

.𝑋

οΏ½π‘›βˆ‘π‘—=1

Eπœ€[𝑑(𝑓(πœ€), 𝑓(πœ€1, . . . , πœ€π‘—βˆ’1,βˆ’πœ€π‘—, πœ€π‘—+1, Β· Β· Β· ))𝑝]

οΏ½ 1𝑝

.

If 𝑋 is also a Banach space of Enflo type 𝑝, then it is also of Rademacher type 𝑝.

Question 3.6.7: Let 𝑋 be a Banach space of Rademacher type 𝑝. Does 𝑋 also have Enflotype 𝑝?

We know special Banach spaces for which this is true, like 𝐿𝑝 spaces, but we don’t knowthe anser in full generality.

Theorem 3.6.8 (Pisier). thm:pisier Let 𝑋 be a Banach space of Rademacher type 𝑝. Then 𝑋has Enflo type π‘ž for every 1 < π‘ž < 𝑝.

We first need the following.

Theorem 3.6.9 (Kahane’s Inequality). For every ∞ > 𝑝, π‘ž β‰₯ 1, there exists 𝐾𝑝,π‘ž such thatfor every Banach space, (𝑋, ‖·‖𝑋) for every π‘₯1, . . . , π‘₯𝑛 ∈ 𝑋,(

Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘₯𝑖

𝑝𝑋

) 1𝑝

≀ 𝐾𝑝,π‘ž

(Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘₯𝑖

π‘žπ‘‹

) 1π‘ž

.

In the real case this is Khintchine’s inequality.By Kahane’s inequality, 𝑋 has type 𝑝 iff(

Eπ‘›βˆ‘π‘–=1

πœ€π‘–π‘₯𝑖

π‘žπ‘‹

) 1π‘ž

.𝑋,𝑝,π‘ž

(π‘›βˆ‘π‘–=1

β€–π‘₯𝑖‖𝑝𝑋

) 1𝑝

.

Define 𝑇 : ℓ𝑛𝑝 (𝑋) β†’ 𝐿𝑝({Β±1}𝑛, 𝑋) by 𝑇 (π‘₯1, . . . , π‘₯𝑛)(πœ€) =βˆ‘π‘›π‘–=1 πœ€π‘–π‘‹π‘–.

82

Page 83: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

(Here ‖𝑓‖𝐿𝑝({Β±1}𝑛,𝑋) =(

12𝑛

βˆ‘πœ€βˆˆ{Β±1}𝑛 ‖𝑓(πœ€)β€–

𝑝𝑋

οΏ½ 1𝑝 .)

Rademacher type 𝑝 mens

‖𝑇‖ℓ𝑛𝑝 (𝑋)β†’πΏπ‘ž({Β±1}𝑛,𝑋) .𝑋,𝑝,π‘ž 1.

For an operator 𝑇 : π‘Œ β†’ 𝑍, the adjoint 𝑇 * : 𝑍* β†’ π‘Œ * satisfies

β€–π‘‡β€–π‘Œβ†’π‘ = ‖𝑇 *‖𝑍*β†’π‘Œ * .

Let 𝑝* = π‘π‘βˆ’1

be the dual of 𝑝 (1𝑝+ 1

𝑝*= 1). Now

πΏπ‘ž({Β±1}𝑛, 𝑋)* = πΏπ‘ž*({Β±1}𝑛, 𝑋*).

Here, 𝑔* : {Β±1}𝑛 β†’ 𝑋*, 𝑓 : {Β±1}𝑛 β†’ 𝑋, 𝑔*(𝑓) = E𝑔*(πœ€)(𝑓(πœ€)), ℓ𝑛𝑝 (𝑋) = ℓ𝑛𝑝*(𝑋*).

Note‖𝑇 *β€–πΏπ‘ž* ({Β±1}𝑛,𝑋*)→ℓ𝑛

𝑝* (𝑋*) . 1,

For 𝑇 : π‘Œ β†’ 𝑍, 𝑇 * : 𝑍* β†’ π‘Œ * is defined by 𝑇 *(𝑧*)(𝑦) = 𝑧*(𝑇𝑦).For 𝑔* : {Β±1}𝑛 β†’ 𝑋*, 𝑔*

βˆ‘π΄βŠ†{1,...,𝑛} 𝑔*(𝐴)π‘Šπ΄. We claim

𝑇 *𝑔* = (𝑔*({1}), . . . , 𝑔({𝑛})).We check

𝑇 *𝑔*(π‘₯1, . . . , π‘₯𝑛) = 𝑔*(𝑇 (π‘₯1, . . . , π‘₯𝑛))

≀ E𝑔*(πœ€)(

π‘›βˆ‘π‘–=1

πœ€π‘–π‘‹π‘–

)=

π‘›βˆ‘π‘–=1

(E𝑔*(πœ€)πœ€π‘–)(𝑋𝑖)

=π‘›βˆ‘π‘–=1

οΏ½ βˆ‘π΄βŠ†{1,...,𝑛}

𝑔*(𝐴)(Eπ‘Šπ΄(πœ€)πœ€π‘–)

οΏ½(π‘₯𝑖)

=π‘›βˆ‘π‘–=1

𝑔*({𝑖})(π‘₯)= (𝑔*({1}), . . . , 𝑔*({𝑛}))(π‘₯1, . . . , π‘₯𝑛).

Rademacher type 𝑝 means for all 𝑔*𝑖 : {Β±1}𝑛 β†’ 𝑋*, for all π‘ž ∈ [1,∞),(π‘›βˆ‘π‘–=1

‖𝑔*({𝑖})‖𝑝*𝑋) . ‖𝑔*β€–πΏπ‘ž({Β±1}𝑛,𝑋*)

Theorem 3.6.10 (Pisier). If 𝑋 has Rademacher type 𝑝 and 1 ≀ π‘Ž < 𝑝, then for every 𝑏 > 1,for all 𝑓 : {Β±1}𝑛 β†’ 𝑋 with E𝑓 = 0,

‖𝑓‖𝑏 -

οΏ½

π‘›βˆ‘π‘—=1

β€–πœ•π‘—π‘“β€–π‘Žπ‘‹

οΏ½ 1π‘Ž

𝑏

.

83

Page 84: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Here, for 𝑓 : {Β±1}𝑛 β†’ 𝑋, πœ•π‘—π‘“ : {Β±1}𝑛 β†’ 𝑋 is defined by

πœ•π‘—π‘“(πœ€) =𝑓(πœ€)βˆ’ 𝑓(πœ€1, . . . , πœ€π‘—βˆ’1,βˆ’πœ€π‘—, πœ€π‘—+1, . . . πœ€π‘›)

2=

βˆ‘π΄ βŠ† {1, . . . , 𝑛}

𝑗 ∈ 𝐴

𝑓(𝐴)π‘Šπ΄.

Note πœ•2𝑗 = πœ•π‘—.The Laplacian is Ξ” =

βˆ‘π‘›π‘—=1 πœ•π‘— =

βˆ‘π‘›π‘—=1 πœ•

2𝑗 . We’ve proved that

Δ𝑓 =βˆ‘

π΄βŠ†{1,...,𝑛}|𝐴| 𝑓(𝐴)π‘Šπ΄.

The heat semigroup on the cube5 is for 𝑑 > 0,

π‘’βˆ’π‘‘Ξ”π‘“ =βˆ‘

π΄βŠ†{1,...,𝑛}π‘’βˆ’π‘‘|𝐴| 𝑓(𝐴)π‘Šπ΄.

This is a function on the cube

πœ€ ↦→

οΏ½π‘›βˆ‘π‘—=1

β€–πœ•π‘—π‘“(πœ€)β€–π‘Žπ‘‹

οΏ½ 1π‘Ž

∈ [0,∞).

For 𝑏 = π‘Ž,

β€–π‘“β€–π‘Ž .

οΏ½π‘›βˆ‘π‘—=1

β€–πœ•π‘—π‘“β€–π‘žπΏπ‘Ž({Β±1}𝑛,𝑋)

οΏ½ 1π‘Ž

.

Let 𝑓 : {Β±1}𝑛 β†’ 𝑋 be

‖𝑓(πœ€)βˆ’ 𝑓(βˆ’πœ€)β€–π‘Ž ≀ ‖𝑓(πœ€)βˆ’ Eπ‘“β€–π‘Ž + E β€–E𝑓 βˆ’ E𝑓(βˆ’πœ€)β€–π‘Ž

.

οΏ½π‘›βˆ‘π‘—=1

‖𝑓(πœ€)βˆ’ 𝑓(πœ€1, . . . ,βˆ’πœ€π‘—, . . . , πœ€π‘›)β€–π‘Žπ‘‹

οΏ½ 1π‘Ž

.

We need some facts about the heat semigroup.

Fact 3.6.11 (Fact 1): Heat flow is a contraction:π‘’βˆ’π‘‘Ξ”π‘“

πΏπ‘ž({Β±1}𝑛,𝑋)

≀ β€–π‘“β€–πΏπ‘ž({Β±1}𝑛,𝑋) .

Proof. We have

π‘’βˆ’π‘‘Ξ”π‘“ =βˆ‘

π΄βŠ†{1,...,𝑛}π‘’βˆ’π‘‘|𝐴| 𝑓(𝐴)π‘Šπ΄ =

βˆ‘π‘…π‘‘ * 𝑓

5also known as the noise operator

84

Page 85: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Here the convolution is defined as 12𝑛

βˆ‘π›Ώβˆˆ{Β±1}𝑛 𝑅𝑑(πœ€π›Ώ)𝑓(𝛿). This is a vector valued function.

In the real case it’s Parseval, multiplying the Fourier coefficients. It’s enough to prove forthe real line. Identities on the level of vectors. Here

𝑅𝑑(πœ€) =βˆ‘

π΄βŠ†{1,...,𝑛}π‘’βˆ’π‘‘|𝐴|π‘Šπ΄(πœ€);

this is the Riesz product.

The function is

=π‘›βˆπ‘–=1

(1 + π‘’βˆ’π‘‘πœ€π‘–) = 0, 𝑑 > 0.

E𝑅𝑑 = 1 so 𝑅𝑑 is a probability measure and heat flow is an averaging operator.

4/11: Continue Theorem 3.6.8. The dual to having Rademacher type 𝑝: for all 𝑔* :

{Β±1}𝑛 β†’ 𝑋*,(βˆ‘π‘›

𝑖=1 ‖𝑔({𝑖})‖𝑝*𝑋*

) 1𝑝* . ‖𝑔*β€–πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*) for every 1 < π‘Ÿ <∞.

Note that for the purpose of proving Theorem 3.6.8 we only need π‘Ÿ = 𝑝 in Kahane’sinequality. However, I recommend that you read up on the full inequality.

For 𝑓 : {Β±1}𝑛 β†’ 𝑋, define πœ•π‘—π‘“ : {Β±1}𝑛 β†’ 𝑋 by πœ•π‘—π‘“ = 𝑓(πœ€)βˆ’π‘“(πœ€1,...,βˆ’πœ€π‘— ,...,πœ€π‘›)2

. Then

πœ•2𝑗 = πœ•π‘— = πœ•*𝑗 and πœ•π‘—π‘Šπ΄ =

βŽ§βŽ¨βŽ©π‘Šπ΄, 𝑗 ∈ 𝐴

0, 𝑗 ∈ 𝐴.

The Laplacian is Ξ” =βˆ‘π‘›π‘—=1 πœ•π‘— =

βˆ‘π‘›π‘—=1 πœ•

2𝑗 ,

Δ𝑓 =βˆ‘π΄βŠ†[𝑛]

|𝐴| 𝑓(𝐴)π‘Šπ΄.

If 𝑑 > 0 then

π‘’βˆ’π‘‘Ξ” =βˆ‘π΄βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐴|π‘Šπ΄

is the heat semigroup. It is a contraction. π‘’βˆ’π‘‘Ξ” = (βˆπ‘›π‘–=1(1 + π‘’βˆ’π‘‘πœ€π‘–)) * 𝑓 .

We have π‘’βˆ’π‘‘Ξ”π‘“

𝐿𝑝({Β±1}𝑛,𝑋)

β‰₯ π‘’βˆ’π‘›π‘‘ ‖𝑓‖𝐿𝑝({Β±1}𝑛,𝑋) (3.29)

π‘’βˆ’π‘‘Ξ”(𝑉[𝑛]π‘’βˆ’π‘‘Ξ”π‘“) = π‘’βˆ’π‘‘π‘›π‘Š[𝑛]𝑓 (3.30)

85

Page 86: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

where π‘Š[𝑛](πœ€) =βˆπ‘›π‘–=1 πœ€π‘–. For 𝑓 =

βˆ‘π΄βŠ†[𝑛]

𝑓(𝐴)π‘Šπ΄, we have

π‘Š[𝑛]π‘’βˆ’π‘‘Ξ”π‘“ =

βˆ‘π΄βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐴| 𝑓(𝐴)π‘Š[𝑛]βˆ–π΄ (3.31)

π‘’βˆ’π‘‘Ξ”(π‘Š[𝑛]π‘’βˆ’π‘‘Ξ”π‘“) =

βˆ‘π΄βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐴| 𝑓(𝐴)π‘’βˆ’π‘‘(π‘›βˆ’|𝐴|) (3.32)

= π‘’βˆ’π‘‘π‘›βˆ‘π΄βŠ†[𝑛]

𝑓(𝐴)π‘Š[𝑛]βˆ–π΄ (3.33)

= π‘’βˆ’π‘‘π‘›π‘Š[𝑛]𝑓 (3.34)

π‘’βˆ’π‘‘π‘›π‘Š[𝑛]𝑓

𝐿𝑝({Β±1}𝑛,𝑋)

=π‘’βˆ’π‘‘Ξ”(π‘Š[𝑛](𝑒

βˆ’π‘‘Ξ”π‘“))𝑝

(3.35)

β‰€οΏ½οΏ½οΏ½π‘Š[𝑛]𝑒

βˆ’π‘‘Ξ”π‘“π‘. (3.36)

The key claim is the following.

Claim 3.6.12. clm:pis1 Suppose 𝑋 has Rademacher type 𝑝. For all 1 < π‘Ÿ < ∞, for all 𝑑 > 0,for all 𝑔* : {Β±1}𝑛 β†’ 𝑋*,οΏ½

π‘’βˆ’π‘‘Ξ”πœ•π‘—π‘”*(πœ€)

π‘ž*𝑋*

οΏ½ 1π‘ž*πΏπ‘Ÿ* ({Β±1}𝑛)

.𝑋,π‘Ÿ,𝑝

‖𝑔*β€–πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*)

(𝑒𝑑 βˆ’ 1)𝑝*π‘ž*

.

Assume the claim. We prove the following.

Claim 3.6.13. clm:pis2 If 𝑋 has Rademacher type 𝑝, 1 < π‘ž < 𝑝, and 1 < π‘Ÿ < ∞, then forevery function 𝑓 : {Β±1}𝑛 β†’ 𝑋 with E𝑓 = 0 = 𝑓(πœ‘) we have

β€–π‘“β€–πΏπ‘Ÿ({Β±1}𝑛,𝑋) .1

π‘βˆ’ π‘ž

οΏ½

π‘›βˆ‘π‘—=1

β€–πœ•π‘—π‘“(πœ€)β€–π‘žπ‘‹

οΏ½ 1π‘ž

πΏπ‘Ÿ({Β±1}𝑛,𝑋)

.

When π‘Ÿ = π‘ž,

‖𝑓 βˆ’ Eπ‘“β€–π‘ž .οΏ½

1

π‘βˆ’ π‘ž

οΏ½π‘žEπœ€

π‘›βˆ‘π‘—=1

β€–πœ•π‘—π‘“(πœ€)β€–π‘žπ‘‹ (3.37)

=

οΏ½1

π‘βˆ’ π‘ž

οΏ½π‘ž 1

2π‘ž

π‘›βˆ‘π‘—=1

‖𝑓(πœ€)βˆ’ 𝑓(πœ€1, . . . ,βˆ’πœ€π‘—, πœ€π‘›)β€–π‘žπ‘‹ (3.38)

‖𝑓(πœ€)βˆ’ 𝑓(βˆ’πœ€)β€–π‘ž ≀ (3.39)

= 2 ‖𝑓(πœ€)βˆ’ Eπ‘“β€–π‘ž (3.40)

.1

π‘βˆ’ π‘ž

οΏ½π‘›βˆ‘π‘—=1

E ‖𝑓(πœ€)βˆ’ 𝑓(πœ€1, . . . ,βˆ’πœ€π‘—, πœ€π‘›)β€–2οΏ½

(3.41)

by definition of Enflo type.We show that the Key Claim 3.6.12 implies Claim 3.6.13.

86

Page 87: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. Normalize so that οΏ½

π‘›βˆ‘π‘—=1

β€–πœ•π‘—π‘“(πœ€)β€–π‘žπ‘‹

οΏ½ 1π‘ž

πΏπ‘Ÿ({Β±1}𝑛,𝑋)

= 1.

For every 𝑠 > 0, by Hahn-Banach, take 𝑔*𝑠 : {Β±1}𝑛 β†’ 𝑋* such that ‖𝑔*π‘ β€–πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*) = 1and

E�𝑔*𝑠(πœ€)(Δ𝑒

βˆ’π‘ Ξ”π‘“(πœ€))οΏ½=Ξ”π‘’βˆ’π‘ Ξ”π‘“

πΏπ‘Ÿ({Β±1}𝑛,𝑋)

.

Write this as¬𝑔*𝑠 ,Δ𝑒

βˆ’π‘ Ξ”π‘“).

Recall that πΏπ‘Ÿ({Β±1}𝑛, 𝑋)* = πΏπ‘Ÿ*({Β±1}𝑛, 𝑋*). Taking β„Ž ∈ πΏπ‘Ÿ({Β±1}𝑛, 𝑋)* and 𝑔* βˆˆπΏπ‘Ÿ*({Β±1}𝑛, 𝑋*), we have

𝑔*(β„Ž) = βŸ¨π‘”*, β„ŽβŸ© = E[𝑔*(πœ€)β„Ž(πœ€)] = E[βŸ¨π‘”*, β„ŽβŸ© (πœ€)].

We have

Ξ”π‘’βˆ’π‘ Ξ”π‘“ =βˆ‘π΄βŠ†[𝑛]

|𝐴|π‘’βˆ’π‘ |𝐴| 𝑓(𝐴)π‘Šπ΄ (3.42)Ξ”π‘’βˆ’π‘ Ξ”π‘“

πΏπ‘Ÿ(𝑋)

=¬𝑔*𝑠 ,

βˆ‘πœ•2𝑗 𝑒

βˆ’π‘ Ξ”π‘“)

(3.43)

=π‘›βˆ‘π‘—=1

¬𝑔*𝑠 , πœ•π‘—π‘’

βˆ’π‘ Ξ”πœ•π‘—π‘“)

(3.44)

=π‘›βˆ‘π‘—=1

Β¬π‘’βˆ’π‘ Ξ”πœ•π‘—π‘”

*𝑠 , πœ•π‘—π‘“

)(3.45)

=(π‘’βˆ’π‘ Ξ”πœ•π‘—π‘”

*𝑠

�𝑛𝑗=1⏟ ⏞

βˆˆπΏπ‘Ÿ* (β„“π‘›π‘ž* (𝑋

*))

(πœ•π‘—π‘“)𝑛𝑗=1⏟ ⏞

βˆˆπΏπ‘Ÿ(β„“π‘›π‘ž (𝑋))

. (3.46)

Note that (πΏπ‘Ÿ(β„“π‘›π‘ž (𝑋)))* = πΏπ‘Ÿ*(β„“

π‘›π‘ž*(𝑋

*)), so we have a pairing here.

≀

οΏ½

π‘›βˆ‘π‘—=1

π‘’βˆ’π‘ Ξ”πœ•π‘—π‘”

*𝑠(πœ€)

π‘ž*𝑋*

οΏ½ 1π‘ž

πΏπ‘Ÿ* (𝑋𝑛)

οΏ½

π‘›βˆ‘π‘—=1

β€–πœ•π‘—π‘“(πœ€)β€–π‘žπ‘‹

οΏ½ 1π‘ž

πΏπ‘Ÿ(𝑋)

(3.47)

≀‖𝑔*π‘ β€–πΏπ‘Ÿ* (𝑋*)

(𝑒𝑑 βˆ’ 1)𝑝*π‘ž*

by Key Claim 3.6.12

(3.48)

.1

(𝑒𝑑 βˆ’ 1)𝑝*π‘ž*.Ξ”π‘’βˆ’π‘ Ξ”π‘“

πΏπ‘Ÿ(𝑋)

.𝑋,𝑝,π‘Ÿ1

(𝑒𝑑 βˆ’ 1)𝑝*π‘ž*. (3.49)

Integrating, ∫ ∞

0Ξ”π‘’βˆ’π‘ Ξ”π‘“ =

βˆ‘π΄βŠ†[𝑛]

∫ ∞

0|𝐴|π‘’βˆ’π‘ |𝐴| 𝑓(𝐴)π‘Šπ΄. (3.50)

87

Page 88: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Then

β€–π‘“β€–πΏπ‘Ÿ(𝑋) =∫ ∞

0Ξ”π‘’βˆ’π‘ Ξ”π‘“ 𝑑𝑠

πΏπ‘Ÿ(𝑋)

(3.51)

β‰€βˆ« ∞

0

Ξ”π‘’βˆ’π‘ Ξ”π‘“

πΏπ‘Ÿ(𝑋)

𝑑𝑠 (3.52)

β‰€βˆ« ∞

0

𝑑𝑠

(𝑒𝑑 βˆ’ 1)𝑝*π‘ž*<∞. (3.53)

We now prove the key claim 3.6.12.

Proof of Key Claim 3.6.12. This is an interpolation argument. We first introduce a naturaloperation.

Given 𝑔* : {Β±1}𝑛 β†’ 𝑋*, for every 𝑑 > 0 define a mapping 𝑔*𝑑 : {Β±1}𝑛 Γ— {Β±1}𝑛 β†’ 𝑋* asfollows. For 𝛿 ∈ {Β±1}𝑛,

𝑔*(πœ€, 𝛿) =βˆ‘π΄βŠ†[𝑛]

𝑔*(𝐴)βˆπ‘–βˆˆπ΄

(π‘’βˆ’π‘‘πœ€π‘– + (1βˆ’ π‘’βˆ’π‘‘)𝛿𝑖) (3.54)

= 𝑔*(π‘’βˆ’π‘‘πœ€+ (1βˆ’ π‘’βˆ’π‘‘)𝛿) (3.55)

Think of 𝑔*(πœ€) =βˆ‘π΄βŠ†[𝑛] 𝑔*(𝐴)βˆπ‘–βˆˆπ΄ πœ€π‘– (extended outside {Β±1}𝑛 by interpolation).

Observe

𝑔*(πœ€, 𝛿) =βˆ‘π΅βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐡|(1βˆ’ π‘’βˆ’π‘‘)π‘›βˆ’|𝐡|𝑔*

οΏ½βˆ‘π‘–βˆˆπ΅

πœ€π‘–π‘’π‘– +βˆ‘π‘– ∈𝐡

𝛿𝑖𝑒𝑖

οΏ½(3.56)

𝑔*

οΏ½βˆ‘π‘–βˆˆπ΅

πœ€π‘–π‘’π‘– +βˆ‘π‘– ∈𝐡

𝛿𝑖𝑒𝑖

οΏ½=

βˆ‘π΄βŠ†[𝑛]

𝑔*(𝐴)π‘Š (𝐴 ∩𝐡)(πœ€)π‘Šπ΄βˆ–π΅(𝛿) (3.57)βˆ‘π΅βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐡|(1βˆ’ π‘’βˆ’π‘‘)π‘›βˆ’|𝐡|𝑔*(βˆ‘π‘–βˆˆπ΅

πœ€π‘–π‘’π‘– +βˆ‘π‘– ∈𝐡

𝛿𝑖𝑒𝑖) (3.58)

=βˆ‘π΅βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐡|(1βˆ’ π‘’βˆ’π‘‘)π‘›βˆ’|𝐡| βˆ‘π΄βŠ†[𝑛]

𝑔*(𝐴)π‘Šπ΄βˆ©π΅(πœ€)π‘Šπ΄βˆ–π΅(𝛿) (3.59)

=βˆ‘π΄βŠ†[𝑛]

𝑔*(𝐴) βˆ‘π΅βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐡|(1βˆ’ π‘’βˆ’π‘‘)π‘›βˆ’|𝐡|π‘Šπ΄βˆ©π΅(πœ€)π‘Šπ΄βˆ–π΅(𝛿). (3.60)

Let 𝛾 =

βŽ§βŽ¨βŽ©πœ€π‘–, w.p. π‘’βˆ’π‘‘

𝛿𝑖, w.p. 1βˆ’ π‘’βˆ’π‘‘.. Then

𝑑

(βˆπ‘–βˆˆπ΄

𝛾𝑖

)=βˆπ‘–βˆˆπ΄

(π‘’βˆ’π‘‘ + (1βˆ’ π‘’βˆ’π‘‘)𝛿𝑖 (3.61)

βˆ‘π΄

𝑔*(𝐴)E𝑀𝐴(𝛾) = E(βˆ‘π΄

𝑔*(𝐴)π‘Šπ΄(𝛾)

)(3.62)

=βˆ‘π΅βŠ†[𝑛]

π‘’βˆ’π‘‘|𝐡|(1βˆ’ π‘’βˆ’π‘‘)π‘›βˆ’|𝐡|𝑔*(𝛾) (3.63)

88

Page 89: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Instead of a linear interpolation, we did a random interpolation.What is ‖𝑔*𝑑 β€–πΏπ‘Ÿ* ({Β±1}𝑛×{Β±1}𝑛,𝑋*) ≀

βˆ‘π΅βŠ†[𝑛] 𝑒

βˆ’π‘‘|𝐡|(1βˆ’π‘’βˆ’π‘‘)π‘›βˆ’|𝐡|?

4/13: Continue Theorem 3.6.8.We’re proving a famous and nontrivial theorem, so there are more computations (this is

the only proof that’s known).We previously reduced everything to the following β€œKey Claim” 3.6.12.Fix 𝑑 > 0. Define 𝑔*𝑑 : {Β±1}𝑛 Γ— {Β±1}𝑛 β†’ 𝑋*. Then

𝑔*𝑑 (πœ€, 𝛿) = 𝑔*(π‘’βˆ’π‘‘πœ€+ (1βˆ’ π‘’βˆ’π‘‘)𝛿)

=βˆ‘

π΄βŠ†{1,Β·Β·Β· ,𝑛}𝑔*(𝐴)

βˆπ‘–βˆˆπ΄

(π‘’βˆ’π‘‘πœ€π‘– + (1βˆ’ π‘’βˆ’π‘‘)𝛿𝑖

οΏ½Then it’s a fact that ‖𝑔*𝑑 β€–πΏπ‘Ÿ* ({Β±1}𝑛×{Β±1}𝑛,𝑋*)≀‖𝑔*β€–πΏπ‘Ÿ({Β±1}𝑛,𝑋*) which is true for any Banach

space from a Bernoulli interpretation.Let us examine the Fourier expansion of the original definition of 𝑔* in the variable 𝛿𝑖.

What is the linear part in 𝛿? We have

𝑔*𝑑 (πœ€, 𝛿) =π‘›βˆ‘π‘–=1

𝛿𝑖

οΏ½ βˆ‘π΄βŠ†{1,Β·Β·Β· ,𝑛},π‘–βˆˆπ΄

(1βˆ’ π‘’βˆ’π‘‘(|𝐴|βˆ’1)πœ€π‘—)

οΏ½+ Ξ¦(πœ€, 𝛿)

where Ξ¦ is the remaining part. The way to write this is to write EΞ¦(πœ€, 𝛿)𝛿𝑖 = 0, for all𝑖 = 1, Β· Β· Β· , 𝑛 since we know the extra part is orthogonal to all the linear parts, since theWalsh function determines an orthogonal basis.

So βˆ‘π΄βŠ†{1,Β·Β·Β· ,𝑛},π‘–βˆˆπ΄

(1βˆ’ π‘’βˆ’π‘‘)π‘’βˆ’π‘‘|𝐴|π‘’π‘‘βˆ

π‘—βˆˆπ΄βˆ–{𝑖}𝑔*(𝐴) = (𝑒𝑑 βˆ’ 1)πœ€π‘–π‘’

βˆ’π‘‘Ξ”πœ•π‘–π‘”*

What does πœ•π‘– do to 𝑔*? It only keeps the 𝑖s which belong to it, and it keeps it with thesame coefficient. Then you hit it with π‘’βˆ’π‘‘Ξ”, which corresponds to π‘’βˆ’π‘‘|𝐴|. Then the last partis just the Walsh function.

All in all, you get that 𝑔*𝑑 (πœ€, 𝛿) = (𝑒𝑑 βˆ’ 1)βˆ‘π‘›π‘–=1 π›Ώπ‘–πœ€π‘–π‘’

βˆ’π‘‘Ξ”πœ•π‘–π‘”*(πœ€) + Ξ¦(πœ€, 𝛿).

Now fix πœ€ ∈ {Β±1}𝑛. Recall the dual formulation of Rademacher type 𝑝: For every functionβ„Ž* : {Β±1}𝑛 β†’ 𝑋*, compute β€–(βˆ‘π‘›

𝑖=1 β€–β„Ž*(𝑖)‖𝑝*

𝑋 )1/𝑝*β€–πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*) ≀ β€–β„Ž*β€–πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*).

Then, applying this inequality to our situation with 𝑔*𝑑 (πœ€, 𝛿) where we have fixed πœ€, weget

(

π‘›βˆ‘π‘–=1

β€–π‘’βˆ’π‘‘Ξ”πœ•π‘–π‘”*(πœ€)‖𝑝*

𝑋*

)1/𝑝*πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*)

. (𝑒𝑑 βˆ’ 1)βˆ’1(E𝛿‖𝑔*𝑑 (πœ€, 𝛿)β€–π‘Ÿ

*

𝑋*

οΏ½Then, we get

(

π‘›βˆ‘π‘–=1

β€–π‘’βˆ’π‘‘Ξ”πœ•π‘–π‘”*‖𝑝*

𝑋*

)1/𝑝*πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*)

. (𝑒𝑑 βˆ’ 1)βˆ’1‖𝑔*𝑑 β€–πΏπ‘Ÿ* ({Β±1}𝑛×{Β±1}𝑛,𝑋*)

. (𝑒𝑑 βˆ’ 1)βˆ’1β€–π‘”β€–πΏπ‘Ÿ* ({Β±1}𝑛,𝑋*)

89

Page 90: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Now we want to take 𝑑→ ∞. Let us look at β€–π‘’βˆ’π‘‘Ξ”πœ•π‘–π‘”*β€–πΏβˆž({Β±1}𝑛,𝑋*). π‘’βˆ’π‘‘Ξ” is a contraction,

and so does πœ•π‘–. We proved the first term is an averaging operator, which does not decreasenorms. We also have that πœ•π‘– is averaging a difference of two values, so it also does notdecrease norm. Therefore, we can put a max on it and still get our inequality:

max1≀𝑖≀𝑛

β€–π‘’βˆ’π‘‘Ξ”πœ•π‘–π‘”*β€–πΏβˆž({Β±1}𝑛,𝑋*) ≀ ‖𝑔*β€–πΏβˆž({Β±1}𝑛,𝑋*)

Then, (

π‘›βˆ‘π‘–=1

β€–π‘’βˆ’π‘‘Ξ”πœ•π‘–π‘”*β€–π‘Žπ‘‹*

)1/π‘ŽπΏπ‘({Β±1}𝑛,𝑋*)

= β€–(β€–π‘’βˆ’π‘‘Ξ”πœ•π‘–π‘”*‖𝑋*)𝑛𝑖=1‖𝐿𝑏(π‘™π‘›π‘Ž (𝑋*))

Define 𝑆 : πΏπ‘Ÿ*(𝑋*) β†’ πΏπ‘Ÿ*(𝑙

𝑛𝑝*(𝑋

*)), and it maps

𝑆(𝑔*) = (π‘’βˆ’π‘‘Ξ”πœ•π‘–π‘”*(πœ€))𝑛𝑖=1

The first inequality is the same thing as saying the operator norm of 𝑆

β€–π‘†β€–πΏπ‘Ÿ* (𝑋*)β†’πΏπ‘Ÿ* (𝑙𝑛𝑝* (𝑋

*)) .1

𝑒𝑑 βˆ’ 1

We also know that

β€–π‘†β€–πΏβˆž(𝑋*)β†’πΏβˆž(π‘™π‘›βˆž(𝑋*)) ≀ 1

We now want to interpolate these two statements (π‘Ÿ* and ∞) to arrive at π‘ž*, as in thelast remaining portion of the proof which is left.

Define πœƒ ∈ [0, 1] by 1π‘ž*

= πœƒπ‘*

+ 1βˆ’πœƒβˆž = πœƒ

𝑝*, so πœƒ = 𝑝*

π‘ž*.

By the vector-valued Riesz Interpolation Theorem (proof is identical to real-valued func-tions) (this is in textbooks),

β€–π‘†β€–πΏπ‘Ž* (𝑋*)β†’πΏπ‘Ž* (π‘™π‘›π‘ž* (𝑋

*)) .1

(𝑒𝑑 βˆ’ 1)πœƒ

provided 1π‘Ž*

= πœƒπ‘Ÿ*

+ 1βˆ’πœƒβˆž , or π‘Ž* = π‘Ÿ*

πœƒ= π‘Ÿ*π‘ž*

𝑝*. π‘ž*

𝑝*> 1, as π‘Ÿ ranges from 1 β†’ ∞, so does π‘Ÿ*.

In Pisier’s paper, he says we can get any π‘Ž* that we want. However, this seems to be false.But this does not affect us: If we choose π‘Ÿ = 𝑝, then we get π‘Ž = π‘ž, which is all we needed tofinish the proof.

So in 3.6.12, we take π‘Ÿ* > π‘ž*

𝑝*, and this completes the claim.

Later I will either prove or give a reference for the interpolation portion of the argument.Remember what we proved here and the definition of Enflo type 𝑝. You assign 2𝑛 points

in Banach space, to every vertex of the hyper cube. We deduce the volume of parallelpipedfor any number of points. The open question is do we need to pass to this smaller value,and we needed it to get the integral to converge.

90

Page 91: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Chapter 4

Grothendieck’s Inequality

1 Grothendieck’s Inequality

Next week is student presentations, let’s give some facts about Grothendieck’s inequality.Let’s prove the big Grothendieck theorem (this has books and books of consequences). Ap-plications of Grothendieck’s inequality is a great topic for a separate course.

The big Grothendieck Inequality is the following:

Theorem 4.1.1. Big Grothendeick.There exists a universal constant 𝐾𝐺 (the Grothendieck constant) such that the followingholds. Let 𝐴 = (π‘Žπ‘–π‘—) βˆˆπ‘€π‘šΓ—π‘›(R), and π‘₯1, . . . , π‘₯π‘š, 𝑦1, . . . , 𝑦𝑛 be unit vectors in a Hilbert space𝐻. Then there exist signs πœ€1, . . . , πœ€π‘›, 𝛿1, . . . , 𝛿𝑛 ∈ {Β±1}𝑛 such that if you look at the bilinearform βˆ‘

𝑖,𝑗

π‘Žπ‘–π‘—βŸ¨π‘₯𝑖, π‘¦π‘—βŸ© ≀ 𝐾𝐺

βˆ‘π‘–,𝑗

π‘Žπ‘–π‘—πœ€π‘–π›Ώπ‘—

So whenever you have a matrix, and two sets of high-dimensional vectors in a Hilbertspace, there is at most a universal constant (less than 2 or 3) times the same thing but withsigns.

Let’s give a geometric interpretation. Consider the following convex bodies in (R𝑛)2. So 𝐴is the convex hull of all matrices of the form π‘π‘œπ‘›π‘£({(πœ€π‘–π›Ώπ‘—) : πœ€1, Β· Β· Β· , πœ€π‘š, 𝛿1, Β· Β· Β· , 𝛿𝑛 ∈ {Β±1}}) βŠ†π‘€π‘šΓ—π‘›(R) ∼= Rπ‘šπ‘›. These matrices all have entries Β±1. This is a polytope.

𝐡 = π‘π‘œπ‘›π‘£({⟨π‘₯𝑖, π‘¦π‘—βŸ© : π‘₯𝑖, 𝑦𝑗 unit vectors in a Hilbert space }) βŠ† π‘€π‘šΓ—π‘›(R). It’s obviousthat 𝐡 contains 𝐴 since 𝐴 is a restriction to Β±1 entries.

Grothendieck says the latter half of the containment 𝐴 βŠ† 𝐡 βŠ† 𝐾𝐺𝐴. If there is a pointoutside 𝐾𝐺𝐴, you can find a separating hyperplane, which is a matrix. If 𝐡 is not in 𝐾𝐺𝐴,then there exists ⟨π‘₯𝑖, π‘¦π‘—βŸ© ∈ 𝐾𝐺𝐴 which means by separation there exists a matrix π‘Žπ‘–π‘— suchthat

βˆ‘π‘Žπ‘–π‘—βŸ¨π‘₯𝑖, 𝑦𝑗 > 𝐾𝐺

βˆ‘π‘Žπ‘–π‘—π‘π‘–π‘— for all 𝑐𝑖𝑗 ∈ 𝐴, which is a contradiction by Grothendieck’s

theorem taking 𝑐𝑖𝑗 = πœ€π‘–π›Ώπ‘—.

We previously proved π‘™βˆž β†’ 𝑙2, the following is a harder theorem:

91

Page 92: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Lemma 4.1.2. Every linear operator 𝑇 : π‘™π‘›βˆž β†’ 𝑙𝑛1 satisfies for all π‘₯1, Β· Β· Β· , π‘₯π‘š ∈ β„“π‘›βˆž.(π‘›βˆ‘π‘–=1

‖𝑇π‘₯𝑖‖2𝑙𝑛1

)1/2

≀ 𝐾𝐺 Β· β€–π‘‡β€–π‘™π‘›βˆžβ†’π‘™π‘›1max

β€–π‘₯β€–1≀1,π‘₯π‘–βˆˆπ‘™π‘›1

(π‘›βˆ‘π‘–=1

⟨π‘₯, π‘₯π‘–βŸ©2)1/2

As an exercise, see how the lemma follows from Grothendieck, and how the little Grothendieckinequality follows from the big one in the case where π‘Žπ‘–π‘— is positive definite (so it’s just a

special case; there the best constant isΓˆπœ‹/2).

Note that maxβ€–π‘₯β€–1≀1⟨π‘₯, π‘₯π‘–βŸ© just gives β€–π‘₯π‘–β€–βˆž. But Grothendieck says for any number ofπ‘₯1, Β· Β· Β·π‘₯π‘š, we can for free improve our norm bound by a constant universal factor.

In the little Grothendieck inequality, what we did was prove the above fact for an operatorfor 𝑙1 β†’ 𝑙2, and we deduced Pietch domination from that conclusion: There is a probabilitymeasure on the unit ball of ℓ𝑛1 such that ‖𝑇π‘₯β€–21 ≀ 𝐾2

πΊβ€–π‘‡β€–β„“π‘›βˆžβ†’β„“π‘›1

βˆ«π΅β„“π‘›

1

βŸ¨π‘¦*, π‘₯⟩2π‘‘πœ‡(𝑦*). If you

know there exists such a probability measure, you know the above lemma, since you just dothis for each π‘₯𝑖, and this is at most the maximum so you can just multiply.

This is an amazing fact though: Look at 𝑇 : β„“π‘›βˆž β†’ ℓ𝑛1 , and look at 𝐿2(πœ‡, ). This issaying that if you think of an element in β„“βˆž as a bounded function on the 𝐿2 unit ball of itsdual. You can think like this for any Banach space. Then, we have a diagram mapping fromℓ𝑛2 β†’ 𝐿2(πœ‡) by identity into a Hilbert space subspace 𝐻 of 𝐿2(πœ‡), and 𝑆 : 𝐻 β†’ ℓ𝑛1 . Yougot these operators to factor through a Hilbert space 𝐻, so we form a commutative diagramand the norm of 𝑆 is at most ‖𝑆‖ ≀ 𝐾𝐺‖𝑇‖.

This is how this is used. These theorems give you for free a Hilbert space out of nothing,and we use this a lot. After the presentations, from just this duality consequence, I’ll proveone or two results in harmonic analysis and another result in geometry, and it really lookslike magic. You can end up getting things like Parseval for free. We already saw the powerof this in Restricted Invertibility.

Let’s begin a proof of Grothendieck’s inequality.

Proof. We will give a proof getting 𝐾𝐺 ≀ πœ‹2 log(1+

√2)

= 1.7 Β· Β· Β· (this is not from the original

theorem). We don’t actually know what 𝐾𝐺 is. People don’t really care what 𝐾𝐺 is beyondthe first digit. Grothendieck was interested in the worst configuration of points on the sphere,but it would not tell us what the configuration of points is. We want to know the actualnumber, or some description of the points. We know this is not the best bound since itdoesn’t give the worst configuration of points.

You do the following: The input was points π‘₯1, Β· Β· Β· , π‘₯π‘š, 𝑦1, Β· Β· Β· , 𝑦𝑛 ∈ unit sphere in Hilbertspace 𝐻. We will construct a new Hilbert space 𝐻 β€² and new unit vectors π‘₯′𝑖, 𝑦

′𝑗 such that if

𝑧 is uniform over the sphere of 𝐻 β€² according to surface area measure, then the expectation

of Esign(βŸ¨π‘§, π‘₯β€²π‘–βŸ©) * sign(βŸ¨π‘§, π‘¦β€²π‘—βŸ©) =2 log(1+

√2)

πœ‹βŸ¨π‘₯𝑖, π‘¦π‘—βŸ© for all 𝑖, 𝑗.

So we have a sphere with points π‘₯𝑖, 𝑦𝑗, and there is a way to nonlinearly transform into asphere in the same dimension with π‘₯′𝑖, 𝑦

′𝑗. On the new sphere, if you take a uniformly random

direction on the sphere, and look at a hyperplane defined by 𝑧, then the sign just indicateswhether π‘₯′𝑖, 𝑦

′𝑗 are on the same or opposite sides. Let’s call πœ€π‘– the first β€œrandom sign”, and

𝛿𝑗 the second β€œrandom sign”. Then, Eβˆ‘π‘Žπ‘–π‘—πœ€π‘–π›Ώπ‘— = 2 log(1+

√2)

πœ‹

βˆ‘π‘Žπ‘–π‘—βŸ¨π‘₯𝑖, π‘¦π‘—βŸ©. So there exists

92

Page 93: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

an instance (random construction) of πœ€π‘–, 𝛿𝑗 such thatβˆ‘π‘Žπ‘–π‘—βŸ¨π‘₯𝑖, π‘¦π‘—βŸ© ≀ πœ‹

2 log(1+√2)

βˆ‘π‘Žπ‘–π‘—βŸ¨π‘₯𝑖, π‘¦π‘—βŸ©.

If you succeed in making a bigger constant bound when bounding the expectation of themultiplication of the signs, you get a better Grothendieck constant. For the argument wewill give, we will see that the exact number we get will be the best constant.

Undergrad Presentations.

2 Grothendieck’s Inequality for Graphs (by Arka and

Yuval)

We’re going to talk about upper bounds for the Grothendieck constant for quadratic formson graphs.

Theorem 4.2.1. Grothendieck’s Inequality.

π‘›βˆ‘π‘–=1

π‘šβˆ‘π‘–=1

π‘Žπ‘–π‘—βŸ¨π‘₯𝑖, π‘¦π‘–βŸ© ≀ 𝐾𝐺

π‘›βˆ‘π‘–=1

π‘šβˆ‘π‘–=1

π‘Žπ‘–π‘—πœ€π‘–π›Ώπ‘—

where π‘₯𝑖, 𝑦𝑗 are on the sphere. This is a bipartite graph where π‘₯𝑖s form one side and 𝑦𝑗sform the other side. Then you assign arbitrary weights π‘Žπ‘–π‘—. So you can consider arbitrarygraphs on 𝑛 matrices.

For every π‘₯𝑖, 𝑦𝑗, there exist signs πœ€π‘–, 𝛿𝑗 such thatβˆ‘π‘–,π‘—βˆˆ{1,Β·Β·Β· ,𝑛}

π‘Žπ‘–π‘—βŸ¨π‘₯𝑖, π‘¦π‘—βŸ© ≀ 𝐢 log π‘›βˆ‘

π‘Žπ‘–π‘—πœ€π‘–π›Ώπ‘—

We will prove an exact bound on the complete graph, where the Grothendieck constantis 𝐢 log 𝑛.

For a general graph, we will prove that 𝐾(𝐺) = c𝑂(logΘ(𝐺)), where we will define theΘ function of a graph later.

Definition 4.2.2: 𝐾(𝐺) is the least constant 𝐾 s.t. for all matrices 𝐴 : 𝑉 Γ— 𝑉 β†’ R,

sup𝑓 :π‘‰β†’π‘†π‘›βˆ’1

βˆ‘(𝑒,𝑣)∈𝐸

𝐴(𝑒, 𝑣)βŸ¨π‘“(𝑒), 𝑓(𝑣)⟩ ≀ 𝐾supπœ™:𝑉→𝐸

βˆ‘(𝑒,𝑣)∈𝐸

𝐴(𝑒, 𝑣)πœ™(𝑒)πœ™(𝑣)

Definition 4.2.3: The Gram representation constant of 𝐺.Denote by 𝑅(𝐺) the infimum over constants 𝑅 s.t. for every 𝑓 : 𝑉 β†’ π‘†π‘›βˆ’1. There exists𝐹 : 𝑉 β†’ 𝐿[0,1]

∞ such that for every 𝑣 ∈ 𝑉 we have ‖𝐹 (𝑣)β€–βˆž ≀ 𝑅 and βŸ¨π‘“(𝑒), 𝑓(𝑣)⟩ =⟨𝐹 (𝑒), 𝐹 (𝑣)⟩ =

∫ 10 𝐹 (𝑒)(𝑑)𝐹 (𝑣)(𝑑)𝑑𝑑. For (𝑒, 𝑣) which is an edge, you find the constant 𝑅

so that you can embed functions 𝑓 into 𝐹 such that the β„“βˆž norm is less than an explicitconstant.

93

Page 94: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Lemma 4.2.4. Let 𝐺 be a loopless graph. Then 𝐾(𝐺) = 𝑅(𝐺)2.

Proof. Fix 𝑅 > 𝑅(𝐺) and 𝑓 : 𝑉 β†’ π‘†π‘›βˆ’1. Then there exists 𝐹 : 𝑉 β†’ 𝐿∞[0, 1] such that forevery 𝑣 ∈ 𝑉 , we have ‖𝐹 (𝑣)β€–βˆž ≀ 𝑅 and for (𝑒, 𝑣) ∈ 𝐸 βŸ¨π‘“(𝑒), 𝑓(𝑣)⟩ ≀ ⟨𝐹 (𝑒), 𝐹 (𝑣)⟩. Then,βˆ‘(𝑒,𝑣)∈𝐸

𝐴(𝑒, 𝑣)βŸ¨π‘“(𝑒), 𝑓(𝑣)⟩ =βˆ‘

(𝑒,𝑣)∈𝐸𝐴(𝑒, 𝑣)⟨𝐹 (𝑒), 𝐹 (𝑣)⟩ =

∫ βˆ‘(𝑒,𝑣)∈𝐸

𝐴(𝑒, 𝑣)𝐹 (𝑒)(𝑑)𝐹 (𝑣)(𝑑)𝑑𝑑

β‰€βˆ«sup𝑔:𝑉→[βˆ’π‘…,𝑅]

βˆ‘π΄(𝑒, 𝑣)𝑔(𝑒)𝑔(𝑣)𝑑𝑑

We use definition and linearity to reverse the sum and integral. Now we use the looplessassumption to get to the definition of the right side of the Grothendieck inequality. Wejust need to fix the [βˆ’π‘…,𝑅] to [βˆ’1, 1]. We then get the last term above is equivalent to𝑅2supπœ™:𝑉→[βˆ’1,1]

βˆ‘(𝑒,𝑣)∈𝐸 𝐴(𝑒, 𝑣)πœ™(𝑒)πœ™(𝑣), which completes the proof.

Now, for each 𝑓 : 𝑉 β†’ π‘†π‘›βˆ’1, consider 𝑀(𝑓) βŠ† R|𝐸| = (βŸ¨π‘“(𝑒), 𝑓(𝑣))(𝑒,𝑣)∈𝐸. For eachπœ™ : 𝑉 β†’ {βˆ’1, 1}, define 𝑀(πœ™) = (πœ™(𝑒), πœ™(𝑣))(𝑒,𝑣) ∈ Rπœ™. Now we use the convex geometryinterpretation of Grothendieck from the last class. Mimicking it, we write

conv {𝑀(πœ™) : πœ™ : 𝑉 β†’ {βˆ’1, 1}} βŠ† convβŒ‹π‘€(𝑓) : 𝑓 : 𝑉 β†’ π‘†π‘›βˆ’1

{The implication of Grothendieck’s inequality gives us

convβŒ‹π‘€(𝑓) : 𝑓 : 𝑉 β†’ π‘†π‘›βˆ’1

{βŠ† 𝐾(𝐺) Β· conv {𝑀(πœ™) : πœ™ : 𝑉 β†’ {βˆ’1, 1}}

So there exist weights {πœ†π‘” : 𝑔 : 𝑉 β†’ {βˆ’1, 1}} which satisfyβˆ‘π‘ž:𝑉→{βˆ’1,1} πœ†π‘” = 1, πœ†π‘” β‰₯ 0

and for all (𝑒, 𝑣) ∈ 𝐸 βŸ¨π‘“(𝑒), 𝑓(𝑣)⟩ = βˆ‘π‘ž:𝑣→{βˆ’1,1} πœ†π‘”π‘”(𝑒)𝑔(𝑣)𝐾(𝐺).

Now consider

𝐹 (𝑒) =È𝐾(𝑔) Β· 𝑔(𝑒)[πœ†1 + Β· Β· Β·+ πœ†π‘”βˆ’1, πœ†1 + Β· Β· Β·+ πœ†π‘”]

Then,

⟨𝐹 (𝑒), 𝐹 (𝑣)⟩ =βˆ‘

𝑔:{𝑉→{βˆ’1,1}}𝐾(𝐺)𝑔(𝑒)𝑔(𝑣)πœ†π‘”

Then we consider our interval becomes [βˆ’ΓˆπΎ(𝐺),

È𝐾(𝐺)], and thus 𝑅(𝐺) ≀

È𝐾(𝐺) and

𝑅(𝐺)2 ≀ 𝐾(𝐺). This proof can be modified for graphs with loops, but this is not quitetrue.

An obvious corollary is that if 𝐻 is a subgraph of 𝐺, then 𝑅(𝐻) ≀ 𝑅(𝐺), and 𝐾(𝐻) ≀𝐾(𝐺). This inequality is not obvious from the Grothendieck inequality directly, but isobvious going with our point of view.

Lemma 4.2.5. Let πΎβˆ˜π‘› denote the complete graph on 𝑛-vertices with loops. Then 𝑅(𝐾∘

𝑛) =c𝑂(

√log 𝑛).

94

Page 95: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. Let 𝜎 be the normalized surface measure on π‘†π‘›βˆ’1. By computation, there exists 𝑐 s.t.

𝜎�⌈π‘₯ ∈ π‘†π‘›βˆ’1 : β€–π‘₯β€–βˆž ≀ 𝑐

√log𝑛𝑛

}οΏ½β‰₯ 1βˆ’ 1

2𝑛, which can be calculated through integration.

When you get a function 𝑓 on the sphere to the 𝑛 βˆ’ 1 dimensional sphere, you want tofind a rotation so that all these vectors have low coordinate vector value. We basically usethe union bound. For each of the 𝑛 vectors, the probability you get a vector with low valuedlast coordinate is 1βˆ’ 1/(2𝑛), do this for all the vectors and you get probability greater than1/2. Then you can magnify this to get almost surely.

For every π‘₯ ∈ π‘†π‘›βˆ’1, the random variable on the orthogonal group 𝑂(𝑛) given by π‘ˆ β†’ π‘ˆπ‘₯is uniformly distributed on π‘†π‘›βˆ’1. Thus for every 𝑓 : 𝑉 β†’ π‘†π‘›βˆ’1, there is a rotation π‘ˆ ∈ 𝑂(𝑛)

such that βˆ€π‘£ ∈ 𝑉 β€–π‘ˆ(𝑓(𝑣))β€–βˆž ≀ π‘βˆš

log𝑛𝑛

.

We want 𝐹 (𝑣) to be equal to the π‘—π‘‘β„Ž coordinate of π‘ˆπ‘“(𝑣) on interval of length 1/𝑛. I.e.,Let 𝐹 (𝑒)(𝑑) = (π‘ˆ(𝑓(𝑣)))

βˆšπ‘› on π‘—βˆ’1

𝑛≀ 𝑑 𝑗

𝑛.

Thus 𝑅 ≀ π‘βˆšlog 𝑛.

Now I will prove the thing mentioned at the beginning:

Theorem 4.2.6. 𝐾(𝐺) ≀ logπœ’(𝐺), where πœ’(𝐺) is the chromatic number.

This theorem generalizes since on bipartite graphs πœ’(𝐺) is a constant.One thing we want to observe about the Grothendieck inequality is that it only cares

about the Hilbert space structure. So we will just prove this in one specific (nice) Hilbertspace and then we’ll be done. Fix a probability space (Ξ©, 𝑃 ) such that 𝑔1, Β· Β· Β· , 𝑔𝑛 be i.i.d.standard Gaussians on Ξ© (for instance, Ξ© is the infinite product of R).

Now define the Gaussian Hilbert space 𝐻 = {βˆ‘βˆžπ‘–=1 π‘Žπ‘–π‘”π‘– :

βˆ‘π‘Ž2𝑖 <∞} βŠ† 𝐿2(Ξ©). Every

function in here will have mean zero since these are linear combinations of Gaussians. Notethat the unit ball 𝐡(𝐻) consists of all Gaussian distributions with mean zero and varianceat most 1 (since the variance is norm squared).

Let Ξ“ be the left hand side of the Grothendieck inequality: sup𝑓 :𝑉→{βˆ’1,1}βˆ‘

(𝑒,𝑣)∈𝐸 𝐴(𝑒, 𝑣)βŸ¨π‘“(𝑒), 𝑓(𝑣)⟩,and let Ξ” = sup𝑓 :𝑉→{βˆ’1,1}

βˆ‘(𝑒,𝑣)∈𝐸 𝐴(𝑒, 𝑣)πœ™(𝑒)πœ™(𝑣).

Definition 4.2.7: Truncation.For all 𝑀 > 0, πœ“ ∈ 𝐿2(Ξ©), define the truncation of πœ“ at 𝑀 to be

πœ“π‘€(π‘₯) =

⎧βŽͺβŽͺ⎨βŽͺβŽͺβŽ©πœ“(π‘₯) |πœ“(π‘₯)| ≀𝑀

𝑀 πœ“(π‘₯) β‰₯𝑀

βˆ’π‘€ πœ“(π‘₯) ≀ βˆ’π‘€

We’re just cutting things off at the interval [βˆ’π‘€,𝑀 ]. So fix 𝑓 maximizing Ξ“.

Lemma 4.2.8. There exists Hilbert space 𝐻, β„Ž : 𝑉 β†’ 𝐻, 𝑀 > 0 with β€–β„Ž(𝑣)β€–2𝐻 ≀ 1/2for all vertices 𝑣 ∈ 𝑉 , 𝑀 .

Èlogπœ’(𝐺), and we can now write Ξ“ as a sum over all edges

Ξ“ =βˆ‘

(𝑒,𝑣)𝐴(𝑒, 𝑣)βŸ¨π‘“(𝑒)𝑀 , 𝑓(𝑣)π‘€βŸ©+βˆ‘(𝑒,𝑣)∈𝐸 𝐴(𝑒, 𝑣)βŸ¨β„Ž(𝑒), β„Ž(𝑣)⟩, where the second term is anerror term.

95

Page 96: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. Let π‘˜ = πœ’(𝐺). A fact for all 𝑠 : 𝑉 β†’ 𝑙2 s.t. βŸ¨π‘ (𝑒), 𝑠(𝑣)⟩ = βˆ’1π‘˜βˆ’1

for all (𝑒, 𝑣) ∈ 𝐸, and‖𝑠(𝑒)β€– = 1.

Define another Hilbert space π‘ˆ = 𝑙2 βŠ• R. Define 𝑑, 𝑑, two functions from 𝑉 β†’ π‘ˆ . Theyare defined as follows:

𝑑(𝑒) = (

ΓŠπ‘˜ βˆ’ 1

π‘˜π‘ (𝑒))βŠ• (

1βˆšπ‘˜π‘’1)

𝑑(𝑒) = (βˆ’ΓŠπ‘˜ βˆ’ 1

π‘˜π‘ (𝑒))βŠ• (

1βˆšπ‘˜π‘’1)

Now, 𝑑(𝑒), 𝑑(𝑒) are unit vectors for all (𝑒, 𝑣) ∈ 𝐸. Then βŸ¨π‘‘(𝑒), 𝑑(𝑣)⟩ = ⟨(𝑑)(𝑒), 𝑑(𝑣)⟩ = 0. Wealso have βŸ¨π‘‘(𝑒), 𝑑(𝑣)⟩ = 2

π‘˜.

Our goal is to prove that such a function exists. Set 𝐻 β€² = π‘ˆ βŠ— 𝐿2(Ξ©). We now writedown a function

β„Ž(𝑒) =1

4𝑑(𝑒)βŠ— (𝑓(𝑒) + 𝑓(𝑒)𝑀) + π‘˜π‘‘(𝑒)βŠ— (𝑓(𝑒)βˆ’ 𝑓(𝑒)𝑀)

It turns out that this function does everything we want. A couple of key facts: It’s easy tocheck that the definition of Ξ“ holds, defined in terms of β„Ž(𝑒) (for the error term). Checkingβ€–β„Ž(𝑣)β€–2𝐻 ≀ 1/2 holds is also possible. You can bound β€–β„Ž(𝑒)β€–2 ≀ (1/2 + π‘˜β€–π‘“(𝑒)βˆ’ 𝑓(𝑒)𝑀‖)2after evaluating inner products. Now we use th eproperty of the Hilbert space. 𝑓(𝑒) is in theball of 𝐻, with mean zero and variance at most 1. Then, ‖𝑓(𝑣)βˆ’ 𝑓(𝑣)𝑀‖2 is the probabilitythat the Gaussian falls outside the interval [βˆ’π‘€,𝑀 ]. Therefore, from basic probabilitytheory,

‖𝑓(𝑣)βˆ’ 𝑓(𝑣)𝑀‖2 β‰€βˆš2πœ‹βˆ« ∞

𝑀

π‘₯2π‘’βˆ’π‘₯2/2𝑑π‘₯ ≀ 2π‘€π‘’βˆ’π‘€

2/2

where the last part follows by calculus. This is great since if we pick 𝑀 ∼√log π‘˜. So this

will decay super quickly. Picking 𝑀 = 8√log π‘˜ gives that the entire thing will be ≀ 1/2.

That proves the lemma, so we’re done.

This lemma is actually all we need. Suppose we have proved this. Now we can say the fol-lowing. We can bound the first term by taking expectations:

βˆ‘(𝑒,𝑣)∈𝐸 𝐴(𝑒, 𝑣)βŸ¨π‘“(𝑒)𝑀 , 𝑓(𝑣)π‘€βŸ© =

Eβˆ‘

(𝑒,𝑣)∈𝐸 𝐴(𝑒, 𝑣)𝑓(𝑒)𝑀𝑓(𝑣)𝑀 ≀𝑀2Ξ” by the definition of Ξ”.

For the second term, we writeβˆ‘

(𝑒,𝑣)∈𝐸 𝐴(𝑒, 𝑣)βŸ¨β„Ž(𝑒), β„Ž(𝑣)⟩ ≀ (maxπ‘£βˆˆπ‘‰ β€–β„Ž(𝑣)β€–2)Ξ“ ≀ 12Ξ“.

by re-scaling. You need to pull out each of these maxima separately, using linearity each time.You can also multiply each thing by

√2 makes it be in 𝐡(𝐻). You just have to multiply by√

2 and divide by√2 and they’re not dependent anymore. By the conditions of the lemma,

our β„Ž has norm at most 1/2, so the This entire thing is Hilbert space independent, so thewhole thing is < Ξ“

2. Thus Ξ“ ≀𝑀2Ξ”+ 1

2Ξ“ =β‡’ Ξ“ ≀ 2𝑀2Ξ”.

This implies Ξ“ . logπœ’(𝐺)Ξ”, which is exactly what we wanted to say, since the lefthand side of the Grothendieck inequality is the first term, and the right hand side of theGrothendieck inequality is the second term with Grothendieck constant logπœ’(𝐺).

96

Page 97: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

3 Noncommutative Version of Grothendieck’s Inequal-

ity - Thomas and Fred

First we’ll phrase the classical Grothendieck inequality in terms of an optimization problem.Given 𝐴 βˆˆπ‘€π‘›(R), we can consider

maxπœ€π‘–,π›Ώπ‘—βˆˆ

π‘›βˆ‘π‘–,𝑗=1

π΄π‘–π‘—πœ€π‘–π›Ώπ‘—

In general this is hard to solve, so we do a semidefinite relaxation

sup𝑑 dimensionssupπ‘₯,π‘¦βˆˆ(π‘†π‘‘βˆ’1)𝑛

π‘›βˆ‘π‘–,𝑗=1

π΄π‘–π‘—βŸ¨π‘₯𝑖, π‘¦π‘—βŸ©

which implies that it’s polynomially solveable, and Grothendieck inequality ensures you’reonly a constant factor off from the best.

We can give a generalization to tensors. Given 𝑀 βˆˆπ‘€π‘›(𝑀𝑛(R)) consider

sup𝑒,π‘£βˆˆπ‘‚π‘›

π‘›βˆ‘π‘–,𝑗,π‘˜,𝑙=1

π‘€π‘–π‘—π‘˜π‘™π‘ˆπ‘–π‘—π‘‰π‘˜π‘™

Set 𝑀𝑖𝑖𝑗𝑗 = 𝐴𝑖𝑗, then we obtain

supβˆ‘

π΄π‘–π‘—π‘ˆπ‘–π‘–π‘‰π‘—π‘— = supπ‘₯,π‘¦βˆˆ{βˆ’1,1}𝑛

π‘›βˆ‘π‘–,𝑗=1

𝐴𝑖𝑗π‘₯𝑖𝑦𝑗 = maxπœ€,π›Ώβˆˆ{βˆ’1,1}𝑛

βˆ‘π΄π‘–π‘—πœ€π‘–π›Ώπ‘—

We relax to SDP over R of 𝑀 :

supπ‘‘βˆˆN sup𝑋,π‘Œ βˆˆπ‘‚π‘›(R𝑑)

π‘›βˆ‘π‘–,𝑗,π‘˜,𝑙=1

π‘€π‘–π‘—π‘˜π‘™βŸ¨π‘‹π‘–π‘—, π‘Œπ‘˜π‘™βŸ©

Recall that π‘ˆ ∈ 𝑂𝑛 means that

π‘›βˆ‘π‘˜=1

π‘ˆπ‘–π‘˜π‘ˆπ‘—π‘˜ =π‘›βˆ‘π‘˜=1

π‘ˆπ‘˜π‘–π‘ˆπ‘˜π‘— = 𝛿𝑖𝑗

If 𝑋 βˆˆπ‘€π‘›(R𝑑), let 𝑋𝑋*, 𝑋*𝑋 βˆˆπ‘€π‘›(R) defined by

(𝑋𝑋*)𝑖𝑗 =π‘›βˆ‘π‘˜=1

βŸ¨π‘‹π‘–π‘˜, π‘Œπ‘—π‘˜βŸ©

(𝑋*𝑋)𝑖𝑗 =π‘›βˆ‘π‘˜=1

βŸ¨π‘‹π‘˜π‘–, π‘Œπ‘˜π‘—βŸ©

Then 𝑂𝑛(R𝑑) =βŒ‹π‘‹ :𝑀𝑛(R𝑑) : 𝑋𝑋* = 𝑋*𝑋 = 𝐼

{.

97

Page 98: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

We want to say something about when we relax the search space, we get within a constantfactor of the non-relaxed version of the program. We will prove this in the complex case.

We will write

OptC(𝑀) = supπ‘ˆ,𝑉 βˆˆπ‘ˆπ‘›|

π‘›βˆ‘π‘–,𝑗,π‘˜,𝑙=1

π‘€π‘–π‘—π‘˜π‘™π‘ˆπ‘–π‘—π‘‰ π‘˜π‘™|

The fact that Opt is less than the SDP, Pisier proved thirty years after Grothendieck con-jectured it. What we will actually prove is that the SDP solution is at most twice theoptimal:

SDPC(𝑀) ≀ 2OptC(𝑀)

Theorem 4.3.1. Fix 𝑛, 𝑑 ∈ N and πœ€ ∈ (0, 1). Suppose we have 𝑀 ∈ 𝑀𝑛(𝑀𝑛(C)) and𝑋, π‘Œ ∈ π‘ˆπ‘›(C𝑑) such that

|π‘›βˆ‘

𝑖,𝑗,π‘˜,𝑙=1

π‘€π‘–π‘—π‘˜π‘™βŸ¨π‘‹π‘–π‘—, π‘Œπ‘˜π‘™βŸ©| β‰₯ (1βˆ’ πœ€)SDPC(𝑀)

So what we’re saying is suppose some SDP algorithm gave a solution satisfying this frominput 𝑋, π‘Œ ∈ π‘ˆπ‘›(C𝑑). Then we can give a rounding algorithm which will output 𝐴,𝐡 ∈ π‘ˆπ‘›such that

E|π‘›βˆ‘

𝑖,𝑗,π‘˜,𝑙=1

π‘€π‘–π‘—π‘˜π‘™π΄π‘–π‘—π΅π‘˜π‘™| β‰₯ (1/2βˆ’ πœ€)SDPC(𝑀)

Now what’s the algorithm? It’s a slightly clever version of projection. We have 𝑋, π‘Œ βˆˆπ‘ˆπ‘›(C), and we want to get to actual unitary matrices. First sample 𝑧 ∈ {1,βˆ’1, 𝑖,βˆ’π‘–}𝑑(complex unit cube). Then π‘₯𝑧 = 1√

2βŸ¨π‘‹, π‘§βŸ© (take inner products columnwise), similarly

π‘Œπ‘§ = 1√2βŸ¨π‘Œ, π‘§βŸ©. Now we take the Polar Decomposition 𝐴 = π‘ˆΞ£π‘‰ * where π‘ˆ, 𝑉 are unitary.

The polar decomposition is just 𝐴 = (π‘ˆπ‘‰ *)(𝑉 Σ𝑉 *), and then the first guy is unitary, andthe second guy is PSD (think of it as π‘’π‘–πœƒ * π‘Ÿ).

Then we have (𝐴,𝐡) = (π‘ˆπ‘§|𝑋𝑧|𝑖𝑑, 𝑉𝑧|π‘Œπ‘§|βˆ’π‘–π‘‘) where 𝑑 is sampled from the hyperbolic secantdistribution. It’s very similar to a normal distribution. The PDF is more precisely

πœ™(𝑑) =1

2sech(

πœ‹

2𝑑) =

1

π‘’πœ‹π‘‘/2 + π‘’βˆ’πœ‹π‘‘/2

When you have a positive definite matrix, you can raise it to imaginary powers, so we’rerotating only the positive semidefinite part, and keeping the rotation side (π‘ˆπ‘‰ *). Note thatthe (𝐴,𝐡) are unitary. (You can also think of this as raising diagonals to 𝑖𝑑, these necessarilyhave eigenvalue of magnitude 1).

4-20-16For 𝑋, π‘Œ ∈ π‘ˆπ‘›(C𝑑), do the following.

1. Sample 𝑧 ∈ {Β±1,±𝑖} uniformly. Sample 𝑑 from the hyperbolic secant distribution.

2. Let 𝑋𝑧 =1√2⟨π‘₯, π‘§βŸ©, π‘Œπ‘ = 1√

2βŸ¨π‘Œ, π‘§βŸ©.

98

Page 99: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

3. Let (𝐴,𝐡) := (π‘ˆπ‘ |𝑋𝑍 |𝑖𝑑, 𝑉𝑍 |π‘Œπ‘ |βˆ’π‘–π‘‘) ∈ π‘ˆπ‘› Γ— π‘ˆπ‘› where 𝑋𝑍 = π‘ˆπ‘ |𝑋𝑍 |, π‘Œπ‘ = π‘ˆπ‘ |π‘Œπ‘ |(polar decomposition).

𝑀(𝑋, π‘Œ ) =βˆ‘π‘›π‘–,𝑗,π‘˜,𝑙=1π‘€π‘–π‘—π‘˜π‘™ βŸ¨π‘‹π‘–π‘—, π‘Œπ‘˜π‘™βŸ©. We want to show

E𝑑[𝑀(𝐴,𝐡)] β‰₯(1

2βˆ’ πœ€

)𝑀(𝑋, π‘Œ ) β‰₯ (

1

2βˆ’ πœ€)SDPC(𝑀)

We want to show that the rounded solution still has large norm.It’s possible that |𝑋𝑍 | has 0 eigenvalues. One solution is to add some Gaussian noise to

the original 𝑋, π‘Œ because the set of non-invertible matrices is an algebraic set of measure 0.Alternatively, replace the eigenvalues by πœ€β†’ 0.

We have

E𝑧[𝑀(𝑋𝑧, π‘Œπ‘§)] =1

2E𝑧

⎑⎣ π‘‘βˆ‘π‘Ÿ,𝑠=1

π‘§π‘Ÿπ‘§π‘ π‘›βˆ‘

𝑖,𝑗,π‘˜,𝑙=1

(𝑋𝑖𝑗)π‘Ÿ(π‘Œπ‘˜π‘™)𝑠

⎀⎦ =1

2𝑀(𝑋, π‘Œ ).

Now we analyze step 3. The characteristic function of the hyperbolic secant distributionis, for π‘Ž > 0,

E[π‘Žπ‘–π‘‘] =βˆ«π‘Žπ‘–π‘‘π‘’(𝑑) 𝑑𝑑 =

2π‘Ž

π»π‘Ž2

by doing a contour integral. Then

E𝑑[π‘Žπ‘–π‘‘] = 2π‘Žβˆ’ E

𝑑[π‘Ž2+𝑖𝑑] (4.1)

E𝑑[𝐴𝑖𝑑] = 2π΄βˆ’ E

𝑑[𝐴2+𝑖𝑑] (4.2)

E𝑑[π΄βŠ—π΅] = E

𝑑[(π‘ˆπ‘§|𝑋𝑧|𝑖𝑑)βŠ— (𝑉𝑧|π‘Œπ‘§|𝑖𝑑)] (4.3)

= (π‘ˆπ‘§ βŠ— 𝑉𝑧)E𝑑[(|𝑋𝑧| βŠ— |π‘Œπ‘‘|)𝑖𝑑] (4.4)

= 2𝑋𝑧 βŠ— π‘Œπ‘§ βˆ’ E𝑑[(π‘ˆπ‘§|𝑋𝑧|2+𝑖𝑑)βŠ— (𝑉𝑧|π‘Œπ‘§|

2βˆ’π‘–π‘‘)] (4.5)

We apply 𝑀 .

E𝑧,𝑑[𝑀(𝐴,𝐡)] =𝑀(𝑋, π‘Œ )βˆ’ E

𝑧,𝑑[𝑀(π‘ˆπ‘§|𝑋𝑧|2+𝑖𝑑, 𝑉𝑧|π‘Œπ‘§|2βˆ’π‘–π‘‘)]

Because 𝐴,𝐡 βˆˆπ‘€π‘›(C), we can write 𝑀(𝐴,𝐡) in terms of the tensor product,

𝑀(𝐴,𝐡) =π‘›βˆ‘

𝑖,𝑗,π‘˜,𝑙=1

π΄π‘–π‘—π΅π‘˜π‘™ (4.6)

=π‘›βˆ‘

𝑖,𝑗,π‘˜,𝑙=1

π‘€π‘–π‘—π‘˜π‘™(π΄βŠ—π΅)(𝑖𝑗),(π‘˜π‘™) (4.7)

Claim 4.3.2 (Key claim). For all 𝑑 ∈ R,E𝑧

�𝑀(π‘ˆπ‘§|𝑋𝑧|2+𝑖𝑑, 𝑉𝑧|π‘Œπ‘§|2βˆ’π‘–π‘‘)

�≀ 1

2SDPC(𝑀).

99

Page 100: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof. We have 𝐹 (𝑑), 𝐺(𝑑) βˆˆπ‘€π‘›(C{Β±1,±𝑖}𝑑). where

(𝐹 (𝑑)π‘—π‘˜)𝑧 =1

2𝑑(π‘ˆπ‘§|𝑋𝑧|2+𝑖𝑑)π‘—π‘˜ (4.8)

(𝐺(𝑑)π‘—π‘˜)𝑧 =1

2𝑑(𝑉𝑧|π‘Œπ‘§|2βˆ’π‘–π‘‘)π‘—π‘˜.𝑀(𝐹 (𝑑), 𝐺(𝑑)) =

1

4π‘‘βˆ‘

π‘§βˆˆ{Β±1,±𝑖}𝑑𝑀(π‘ˆπ‘§|𝑋𝑧|2+𝑖𝑑, 𝑉𝑧|π‘Œπ‘§|2βˆ’π‘–π‘‘) (4.9)

= E𝑧[𝑀(π‘ˆπ‘§|𝑋𝑧|2+𝑖𝑑, 𝑉𝑧|π‘Œπ‘§|2βˆ’π‘–π‘‘)]. (4.10)

Lemma 4.3.3. 𝐹 (𝑑), 𝐺(𝑑) satisfy

max {‖𝐹 (𝑑)𝐹 (𝑑)*β€– , ‖𝐹 (𝑑)*𝐹 (𝑑)β€– , ‖𝐺(𝑑)𝐺(𝑑)*β€– , ‖𝐺(𝑑)*𝐺(𝑑)β€–}

Lemma 4.3.4. Suppose 𝑋, π‘Œ ∈ ℳ𝑛(C𝑑) and max{‖𝑋𝑋*β€– , ‖𝑋*𝑋‖ , β€–π‘Œ π‘Œ *β€– , β€–π‘Œ *π‘Œ β€–}. Thenthere exist 𝑅, 𝑆 ∈ π‘ˆπ‘›(C𝑑+2𝑛2

) such that𝑀(𝑅, 𝑆) =𝑀(𝑋, π‘Œ ) ≀ 1 for every𝑀 βˆˆπ‘€π‘›(𝑀𝑛(C)).

From the two lemmas, there exist 𝑅(𝑑), 𝑆(𝑑) ∈ π‘ˆπ‘›(C𝑑+2𝑛2) such that 𝑀(𝑅(𝑑), 𝑆(𝑑)) =

𝑀(√2𝐹 (𝑑),

√2𝐺(𝑑)). Then

|𝑀(𝐹 (𝑑), 𝐺(𝑑))| = 1

2|𝑀(𝑅(𝑑), 𝑆(𝑑))| ≀ 1

22SDPC(𝑀) (4.11)

𝐹 (𝑑)𝐹 (𝑑)* =1

4π‘‘βˆ‘

π‘§βˆˆ{Β±1,±𝑖}π‘‘π‘ˆπ‘§|𝑋𝑧|*π‘ˆ*

𝑧 (4.12)

= E𝑧[π‘ˆπ‘§|𝑋𝑧|*π‘ˆ*

2 ] = E𝑧[(𝑋𝑧𝑋

*𝑧 )]. (4.13)

Claim 4.3.5. For π‘Š ∈ 𝑀𝑛(C𝑑), define for each 𝑣 ∈ [𝑑], (π‘Šπ‘Ÿ)𝑖𝑗 = (π‘Šπ‘–π‘—)π‘Ÿ, π‘Šπ‘§ = βŸ¨π‘Š, π‘§βŸ©.Then

E𝑧[(π‘Šπ‘§π‘Š

*𝑧 )

2] = (π‘Šπ‘Š *)2 +π‘‘βˆ‘π‘Ÿ=1

π‘Šπ‘Ÿ(π‘Š*π‘Š βˆ’π‘Š *

π‘Ÿπ‘Šπ‘Ÿ)π‘Š*π‘Ÿ .

Note π‘Šπ‘§ =βˆ‘π‘‘π‘Ÿ=1Ξ£π‘Ÿπ‘Šπ‘Ÿ and

π‘Šπ‘Š * =π‘‘βˆ‘π‘Ÿ=1

π‘Šπ‘Ÿπ‘Š*π‘Ÿ π‘Š *π‘Š =

π‘‘βˆ‘π‘Ÿ=1

π‘Š *π‘Ÿπ‘Šπ‘Ÿ. (4.14)

We compute

E𝑧[(π‘Šπ‘§π‘Š

*𝑧 )

2] = E𝑧

βŽ‘βŽ£π‘,π‘ž,π‘Ÿ,π‘ βˆ‘=1

π‘‘π‘§π‘π‘§π‘žπ‘§π‘Ÿπ‘§π‘ π‘Šπ‘π‘Š*π‘žπ‘Šπ‘Ÿπ‘Š

*𝑠

⎀⎦ (4.15)

=π‘‘βˆ‘π‘=1

π‘Šπ‘π‘Š*π‘π‘Šπ‘π‘Š

*𝑝 +

π‘›βˆ‘π‘, π‘ž = 1𝑝 = π‘ž

(π‘Šπ‘π‘Š*π‘žπ‘Šπ‘žπ‘Š

*π‘ž +π‘Šπ‘π‘Š

*π‘žπ‘Šπ‘žπ‘Š

*𝑝 ) (4.16)

100

Page 101: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

π‘‘βˆ‘π‘,π‘ž=1

π‘Šπ‘π‘Š*π‘π‘Šπ‘žπ‘Š

*π‘ž +

π‘‘βˆ‘π‘,π‘ž=1

π‘Šπ‘π‘Š*π‘žπ‘Šπ‘žπ‘Š

*𝑝 βˆ’

π‘‘βˆ‘π‘=1

π‘Šπ‘π‘Š*π‘π‘Šπ‘π‘Š

*𝑝

=

οΏ½π‘‘βˆ‘π‘=1

π‘Šπ‘π‘Š*𝑝

οΏ½2

+π‘‘βˆ‘π‘=1

π‘Šπ‘

οΏ½π‘‘βˆ‘π‘ž=1

π‘Š *π‘žπ‘Šπ‘ž

οΏ½π‘Š *𝑝 βˆ’

π‘‘βˆ‘π‘=1

π‘Šπ‘π‘Š*π‘π‘Šπ‘π‘Š

*𝑝 . (4.17)

Apply the claim with π‘Š = 1√2𝑋. Recall that 𝑋𝑋* = 𝑋*𝑋 = 𝐼. Then

𝐹 (𝑑)𝐹 (𝑑)* +1

4

π‘‘βˆ‘π‘Ÿ=1

π‘‹π‘Ÿπ‘‹*π‘Ÿπ‘‹π‘Ÿπ‘‹

*π‘Ÿ =

1

2𝐼 = 𝐹 (𝑑)𝐹 (𝑑)* +

1

4

π‘‘βˆ‘π‘Ÿ=1

𝑋*π‘Ÿπ‘‹π‘Ÿπ‘‹

*π‘Ÿπ‘‹π‘Ÿ

and similarly for 𝐺.

Proof of Lemma 2. Let 𝐴 = 𝐼 βˆ’π‘‹π‘‹*, 𝐡 = 𝐼 βˆ’π‘‹*𝑋. Note 𝐴,𝐡 βͺ° 0, Tr(𝐴) = Tr(𝐡). Wehave

𝐴 =π‘›βˆ‘π‘–=1

πœ†π‘–(𝑣𝑖𝑣*𝑖 ) (4.18)

𝐡 =π‘›βˆ‘π‘—=1

πœ‡π‘—(𝑣𝑗𝑣*𝑗 ) (4.19)

𝜎 =π‘›βˆ‘π‘–=1

πœ†π‘— =π‘›βˆ‘π‘—=1

πœ‡π‘— (4.20)

𝑅 = 𝑋 βŠ•

�𝑛⨁

𝑖,𝑗=1

ΓŠπœ†π‘–πœ‡π‘—πœŽ

(𝑒𝑖𝑣*𝑗 )

οΏ½βŠ•π‘‚π‘€π‘›(C𝑛2 ) βˆˆπ‘€π‘›(C𝑑 βŠ• C𝑛

2 βŠ• C𝑛2

) (4.21)

𝑆 = π‘Œ βŠ•π‘‚π‘€π‘›(C𝑛2 ) βŠ•

�𝑛⨁

𝑖,𝑗=1

ΓŠπœ†π‘–πœ‡π‘—πœŽ

(𝑒𝑖𝑣*𝑗 )

οΏ½(4.22)

Check 𝑅 ∈ π‘ˆπ‘›(C𝑑+2𝑛2),

𝑅𝑅* = 𝑋𝑋* + 𝐴 (4.23)

𝑅*𝑅 = 𝑋*𝑋 +𝐡 (4.24)

𝑀(𝑅, 𝑆) =𝑀(𝑋, π‘Œ ). (4.25)

The 𝜎 disappears becauseβˆ‘πœ‡π‘— = 𝜎.

The factor 2 in the noncommutative inequality is sharp. That the answer is 2 (ratherthan some strange constant) means there is something going on algebraically. The hyperbolicsecant is the unique distribution that makes this work.

4-27

101

Page 102: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

4 Improving the Grothendieck constant

Theorem 4.4.1 (Krivine (1977)).

𝐾𝐺 ≀ πœ‹

2 ln(1 +√2)

≀ 1.7...

The strategy is preprocessing.

There exist new vectors π‘₯β€²1, . . . , π‘₯′𝑛, 𝑦

β€²1, . . . , 𝑦

′𝑛 ∈ S2π‘›βˆ’1 such that if 𝑧 ∈ S2π‘›βˆ’1 is chosen

uniformly at random, then for all 𝑖, 𝑗,

E𝑧(sign βŸ¨π‘§, π‘₯β€²π‘–βŸ©)⏟ ⏞

πœ€π‘–

(sign¬𝑧, 𝑦′𝑗

))⏟ ⏞

𝛿𝑗

=2 ln(1 +

√2)

πœ‹βŸ ⏞ 𝑐

⟨π‘₯𝑖, π‘¦π‘—βŸ© .

Then βˆ‘π‘Žπ‘–π‘— ⟨π‘₯𝑖, π‘¦π‘—βŸ© = E

[πœ‹

2 ln(1 +√2)

βˆ‘π‘Žπ‘–π‘—πœ€π‘–π›Ώπ‘—

].

We will take vectors, transform them in this way, take a random point on the sphere,take a hyperplane orthogonal, and then see which side the points fall on.

Theorem 4.4.2 (Grothendieck’s identity). For π‘₯, 𝑦 ∈ Sπ‘˜ and 𝑧 uniformly random on Sπ‘˜,

E𝑧[sign(⟨π‘₯, π‘§βŸ©) sign(βŸ¨π‘¦, π‘§βŸ©)] = 2

πœ‹sinβˆ’1(⟨π‘₯, π‘¦βŸ©).

Proof. This is 2-D plane geometry.The expression is 1 if both π‘₯, 𝑦 are on the same side of the line, and βˆ’1 if the line cuts

between the angle.

Once we have the idea to pre-process, the rest of the proof is natural.

102

Page 103: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Proof.

E𝑧

οΏ½sign(βŸ¨π‘§, π‘₯β€²π‘–βŸ©) sign(

¬𝑧, 𝑦′𝑗

))οΏ½=

2

πœ‹sinβˆ’1(

Β¬π‘₯′𝑖, 𝑦

′𝑗

)) = 𝑐 ⟨π‘₯𝑖, π‘¦π‘—βŸ© (4.26)Β¬

π‘₯′𝑖, 𝑦′𝑗

)= sin

οΏ½ πœ‹π‘

2⏟ ⏞ 𝑒

⟨π‘₯𝑖, π‘¦π‘—βŸ©οΏ½. (4.27)

We write the Taylor series expansion for sin.Β¬π‘₯′𝑖, 𝑦

′𝑗

)= sin(𝑒 ⟨π‘₯𝑖, π‘¦π‘—βŸ©) (4.28)

=βˆžβˆ‘π‘˜=0

(𝑒 ⟨π‘₯𝑖, π‘¦π‘—βŸ©)2π‘˜+1

(2π‘˜ + 1)!(βˆ’1)π‘˜ (4.29)

=βˆžβˆ‘π‘˜=0

(βˆ’1)π‘˜π‘’2π‘˜+1

(2π‘˜ + 1)!

⟨π‘₯βŠ—(2π‘˜+1)𝑖 , 𝑦

βŠ—(2π‘˜+1)𝑗

⟩. (4.30)

(For π‘Ž, 𝑏 ∈ β„“2, π‘ŽβŠ— 𝑏 = (π‘Žπ‘–π‘π‘—), π‘ŽβŠ—2 = (π‘Žπ‘–π‘Žπ‘—), 𝑏

βŠ—2 = (𝑏𝑖𝑏𝑗), βŸ¨π‘ŽβŠ—2, π‘βŠ—2⟩ = βˆ‘π‘Žπ‘–π‘Žπ‘—π‘π‘–π‘π‘— = βŸ¨π‘Ž, π‘βŸ©2.)

Define an infinite direct sum corresponding to these coordinates.Define

π‘₯′𝑖 =βˆžβ¨π‘˜=0

οΏ½(βˆ’1)π‘˜π‘’

2π‘˜+12È

(2π‘˜ + 1)!π‘₯βŠ—2π‘˜+1𝑖

οΏ½(4.31)

𝑦′𝑗 =βˆžβ¨π‘˜=0

�𝑒

2π‘˜+12È

(2π‘˜ + 1)!π‘₯βŠ—2π‘˜+1𝑖

�∈

βˆžβ¨π‘˜=1

(Rπ‘š)βŠ—2π‘˜+1 (4.32)

Β¬π‘₯′𝑖, 𝑦

′𝑗

)=

βˆžβˆ‘π‘˜=0

(βˆ’1)π‘˜π‘’2π‘˜+1

(2π‘˜ + 1)!

Β¬π‘₯βŠ—2π‘˜+1𝑖 , π‘¦βŠ—2π‘˜+1

𝑗

). (4.33)

The infinite series expansion of sin generated an infinite space for us.We check

β€–π‘₯′𝑖‖2=

βˆžβˆ‘π‘˜=0

𝑒(2π‘˜+1)

(2π‘˜ + 1)!= sinh(𝑒) =

𝑒𝑒 βˆ’ π‘’βˆ’π‘’

2= 1. (4.34)

We don’t care about the construction, we care about the identity. All I want is to findπ‘₯′𝑖, 𝑦

′𝑗 with given

Β¬π‘₯′𝑖, 𝑦

′𝑗

); this can be done by a SDP. We’re using this infinite argument just

to show existence.A different way to see this is given by the following picture. Take a uniformly random

line and look at the orthogonal projection of points on the line. Is it positive or negative. Apriori we have vectors in R𝑛 that don’t have an order.

We want to choose orientations. A natural thing to do is to take a random projectiononto R and use the order on R. Krevine conjectured his bound was optimal.

103

Page 104: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

What is so special about positive or negative? We can partition it into any 2 measurablesets. Any partition into 2 measurable sets produces a sign. It’s a nontrivial fact that nopartition beats this constant. This is a fact about measure theory.

The moment you choose a partition, it forced the construction of the π‘₯β€², 𝑦′. The wholeidea is the partition; then the preprocessing is uniquely determined; it’s how we reverse-engineered the theorem.

Here’s another thing you can do. What’s so special about a random line? The wholeproblem is about finding an orientation.

Consider a floor (plane) chosen randomly, look at shadows. If you partition the planeinto 2 half-planes, you get back the same constant. We can generalize to arbitrary partitionsof the plane. This is equivalent to an isoperimetric problem. In many such problems the bestthing is a half-space. We can look at higher-dimensional projections. It seems unnatural todo thisβ€”except that you gain!

In R2 there is a more clever partition that beats the halfspace! Moreover, as you takethe increase the dimension, eventually you will converge to the Grothendieck constant. Thepartitions look like fractal objects!

5 Application to Fourier series

This is a classical theorem about Fourier series. Helgason generalized it to a general abeliangroup.

Definition 4.5.1: Let 𝑆 = R/Z. The space of continuous functions is 𝐢(𝑆 β€²). Given π‘š :Zβ†’ R (the multiplier), define

Ξ›π‘š : 𝐢(Sβ€²) β†’ β„“βˆž (4.35)

Ξ›π‘š(𝑓) = (π‘š(𝑛) 𝑓(𝑛))π‘›βˆˆZ. (4.36)

For which multipliers π‘š is Ξ›π‘š(𝑓) ∈ β„“1 for every 𝑓 ∈ 𝐢(S1)?A more general question is, what are the possible Fourier coefficients of a continuous

functions? For example, if log 𝑛 works, thenβˆ‘ | 𝑓(𝑛)||π‘š(𝑛)| <∞. This is a classical topic.

An obvious sufficient condition is that π‘š ∈ β„“2, by Cauchy-Schwarz and Parseval.βˆ‘π‘›βˆˆZ

|π‘š(𝑛) 𝑓(𝑛)| ≀ (βˆ‘π‘š(𝑛)2

οΏ½ 12(βˆ‘

| 𝑓(𝑛)|2οΏ½ 12⏟ ⏞

‖𝑓‖2β‰€β€–π‘“β€–βˆž

(4.37)

β€–Ξ›π‘š(𝑓)β€–β„“1 ≀ β€–π‘šβ€–2 β€–π‘“β€–βˆž . (4.38)

This theorem says the converse.

Theorem 4.5.2 (Orlicz-Paley-Sidon). If Ξ›π‘š(𝑓) ∈ β„“1 for all 𝑓 ∈ 𝐢(S1), then π‘š ∈ β„“2.

We know the fine line of when Fourier coefficients of continuous functions converge.

104

Page 105: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

We can make this theorem quantitative. Observe that if you know Ξ›π‘š : 𝐢(S1) β†’ β„“1,then Ξ›π‘š has a closed graph (exercise). The closed graph theorem (a linear operator betweenBanach spaces with closed graph is bounded) says that β€–Ξ›π‘šβ€– <∞.

Corollary 4.5.3.βˆ‘ |π‘š(𝑛) 𝑓(𝑛)| ≀ 𝐾 β€–π‘“β€–βˆž.

We show β€–Ξ›π‘šβ€– ≀ β€–π‘šβ€–2 ≀ 𝐾𝐺 β€–Ξ›π‘šβ€–.We will use the following consequence of Grothendieck’s inequality (proof omitted).

Corollary 4.5.4. Let 𝑇 : β„“π‘›βˆž β†’ β„“π‘š1 . For all π‘₯1, . . . , π‘₯𝑛 ∈ β„“βˆž,(βˆ‘π‘–

‖𝑇π‘₯𝑖‖21

) 12

≀ 𝐾𝐺 ‖𝑇‖ supπ‘¦βˆˆβ„“1

π‘›βˆ‘π‘–=1

βŸ¨π‘¦, π‘₯π‘–βŸ©2.

Equivalently, there exists a probability measure πœ‡ on [𝑛] such that ‖𝑇π‘₯β€– ≀ 𝐾𝐺 ‖𝑇‖ ∫ 𝐾𝐺 ‖𝑇‖(∫π‘₯2𝑗 π‘‘πœ‡(𝑗)

οΏ½ 12 .

Proof. Use duality. To get the equivalence, use the same proof as in Piesch Domination.

Proof. Given Ξ›π‘š : 𝐢(S1) β†’ β„“1, there exists πœ‡ on S1 such that for every 𝑓 , letting π‘“πœƒ(π‘₯) =𝑓(π‘’π‘–πœƒπ‘₯), (βˆ‘

|π‘š(𝑛) π‘“πœƒ(𝑛)|οΏ½2 ≀ 𝐾2𝐺 β€–Ξ›π‘šβ€–2

βˆ«π‘†1(π‘“πœƒ(π‘₯))

2 π‘‘πœ‡(π‘₯) (4.39)(βˆ‘|π‘š(𝑛) 𝑓(𝑛)|οΏ½2 ≀ 𝐾2

𝐺 β€–Ξ›π‘šβ€–2βˆ«π‘†1(𝑓(π‘₯))2 π‘‘πœ‡(π‘₯). (4.40)

Apply this to the following trigonometric sum

𝑓(π‘₯) =π‘βˆ‘

𝑛=βˆ’π‘π‘š(𝑛)π‘’βˆ’π‘–π‘›πœƒ,

to get οΏ½π‘βˆ‘

𝑛=βˆ’π‘π‘š(𝑛)2

οΏ½2

≀ 𝐾2𝐺 β€–Ξ›β€–2π‘š

π‘βˆ‘π‘›=βˆ’π‘

π‘š(𝑛)2.

RIP also had this kind of magic.

6 ? presentation

Theorem 4.6.1. Let (𝑋, 𝑑) be a metric space with 𝑋 = 𝐴 βˆͺ𝐡 such that

βˆ™ 𝐴 embeds into β„“π‘Ž2 with distortion 𝐷𝐴, and

βˆ™ 𝐡 embeds into ℓ𝑏2 with distortion 𝐷𝐡.

105

Page 106: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Then 𝑋 embeds into β„“π‘Ž+𝑏+12 with distortion at most 7𝐷𝐴𝐷𝐡 + 5(𝐷𝐴 +𝐷𝐡).

Furthermore, given πœ“ : 𝐴→ β„“π‘Ž2 and πœ™π΅ : 𝐡 β†’ ℓ𝑏2 with β€–πœ™π΄β€–Lip ≀ 𝐷𝐴 and β€–πœ™π΅β€–Lip ≀ 𝐷𝐡,

there is an embedding Ξ¨ : 𝑋 β†’Λ“ β„“π‘Ž+𝑏+12 with distortion 7𝐷𝐴𝐷𝐡 + 5(𝐷𝐴 + 𝐷𝐡) such that

β€–Ξ¨(𝑒)βˆ’Ξ¨(𝑣)β€– β‰₯ β€–πœ™π΄(𝑒)βˆ’ πœ™π΄(𝑣)β€– for all 𝑒, 𝑣 ∈ 𝐴.

For π‘Ž ∈ 𝐴, let π‘…π‘Ž = 𝑑(π‘Ž,𝐡) and for 𝑏 ∈ 𝐡, let 𝑅𝑏 = 𝑑(𝐴, 𝑏).

Definition 4.6.2: For 𝛼 > 0, 𝐴′ βŠ† 𝐴 is an 𝛼-cover for 𝐴 with respect to 𝐡 if

1. For every π‘Ž ∈ 𝐴, there exists π‘Žβ€² ∈ 𝐴′ where π‘…π‘Žβ€² ≀ π‘…π‘Ž and 𝑑(π‘Ž, π‘Žβ€²) ≀ π›Όπ‘…π‘Ž.

2. For distinct π‘Žβ€²1, π‘Žβ€²2 ∈ 𝐴′, 𝑑(π‘Žβ€²1, π‘Ž

β€²2) β‰₯ 𝛼min(π‘…π‘Žβ€²1

, π‘…π‘Žβ€²2).

This is like a net, but the πœ€ of the net is a function of the distance to the set. The requirementis weaker for points that are farther away.

Lemma 4.6.3. For all 𝛼 > 0, there is an 𝛼-cover 𝐴′ for 𝐴 with respect to 𝐡.

Proof. Induct. The base case is πœ‘, which is clear.Assume the lemma holds for |𝐴| < π‘˜; we’ll show it holds for |𝐴| = π‘˜. Let 𝑒 ∈ 𝐴 be the

point closest to 𝐡. Let 𝑍 = π΄βˆ–π΅π›Όπ‘…π‘’(𝑒) =: 𝐡; we have |𝑍| < |𝐴|. Let 𝑍 β€² be a cover of 𝑍,and let 𝐴′ = 𝑍 β€² βˆͺ {𝑒}.

We show that 𝐴′ is an 𝛼-cover. We need to show the two properties.

1. Divide into cases based on whether π‘Ž ∈ 𝐡 or π‘Ž ∈ 𝐡.

2. For π‘Žβ€²1, π‘Žβ€²2 ∈ 𝐴′, if both are in 𝑧′, we’re done.

Otherwise, without loss of generality, π‘Žβ€²1 = 𝑒 and π‘Žβ€²2 ∈ 𝑍 β€².

Lemma 4.6.4. Define 𝑓 : 𝐴′ β†’ 𝐡 to send every point in the 𝛼-cover to a closest point in𝐡, 𝑑(π‘Žβ€², 𝑓(π‘Žβ€²)) = π‘…π‘Žβ€².

Then ‖𝑓‖Lip ≀ 2(1 + 1

𝛼

οΏ½. By the triangle inequality,

𝑑(𝑓(π‘Žβ€²1), 𝑓(π‘Žβ€²2)) ≀ 𝑑(𝑓(π‘Žβ€²1), π‘Ž

β€²1) + 𝑑(π‘Žβ€²1, π‘Ž

β€²2) + 𝑑(π‘Žβ€²2, 𝑓(π‘Ž

β€²2)).

We have

π‘…π‘Žβ€²1 +π‘…π‘Žβ€²2 + 𝑑(π‘Žβ€²1, π‘Žβ€²2) = 2min(π‘…π‘Žβ€²1

, π‘…π‘Žβ€²2) + |π‘…π‘Žβ€²1

+π‘…π‘Žβ€²2+ 𝑑(π‘Žβ€²1, π‘Ž

β€²2) (4.41)

≀ 1

𝛼𝑑(π‘Žβ€²1, π‘Ž

β€²2) + 𝑑(π‘Žβ€²1, π‘Ž

β€²2) + 𝑑(π‘Žβ€²1, π‘Ž

β€²2) (4.42)

= 2(1βˆ’ 1

𝛼

)𝑑(𝛼′

1, 𝛼′2) (4.43)

Let πœ™π΅ : 𝐡 β†’Λ“ ℓ𝑏2 be an embedding with β€–πœ™β€–Lip ≀ 𝐷𝐡; it is noncontracting.

106

Page 107: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Lemma 4.6.5. There is πœ“ : 𝑋 β†’ ℓ𝑏2 such that

1. For all π‘Ž1, π‘Ž2 ∈ 𝐴,

β€–πœ“(π‘Ž1)βˆ’ πœ“(π‘Ž2)β€– ≀ 2(1 +

1

𝛼

)𝐷𝐴𝐷𝐡𝑑(𝛼1, 𝛼2)

2. For all 𝑏1, 𝑏2 ∈ 𝐡,

𝑑(𝑏1, 𝑏2) ≀ β€–πœ“(𝑏1)βˆ’ πœ“(𝑏2)β€– = β€–πœ™π΅(𝑏1)βˆ’ πœ™π΅(𝑏2)β€– ≀ 𝐷𝐡𝑑(𝑏1, 𝑏2).

3. For all π‘Ž ∈ 𝐴, 𝑏 ∈ 𝐡, 𝑑(π‘Ž, 𝑏) βˆ’ (1 + 𝛼)(2𝐷𝐴𝐷𝐡 + 1)π‘…π‘Ž ≀ β€–πœ“(π‘Ž)βˆ’ πœ“(𝑏)β€– ≀ 2(1 +𝛼)(𝐷𝐴𝐷𝐡 + (2 + 𝛼)𝐷𝐡)𝑑(π‘Ž, 𝑏).

Proof. Let 𝑔 = πœ™π΅π‘“πœ™βˆ’1𝐴 .

We have maps

𝐴 οΏ½οΏ½ 𝑐 //

πœ™π΄

οΏ½οΏ½

𝐴′

πœ™π΄

οΏ½οΏ½

𝑓// 𝐡

πœ™π΅

οΏ½οΏ½

πœ™(𝐴) πœ™(𝐴′)𝑐oo // πœ™(𝐡) οΏ½οΏ½

// ℓ𝑏2.

Now

‖𝑔‖Lip ≀ β€–πœ™π΅β€–Lip ‖𝑓‖Lipπœ™βˆ’1𝐴

Lip

≀ 𝐷𝐡2(1 +

1

𝛼

)By the Kirszbraun extension theorem 3.6.1, construct 𝑔 : β„“π‘Ž2 β†’ ℓ𝑏2 with

‖𝑔‖Lip ≀ 𝐷𝐡2(1 +

1

𝛼

).

Define πœ“(π‘₯) =

βŽ§βŽ¨βŽ©π‘”(πœ™π΄(π‘₯)), if π‘₯ ∈ 𝐴

πœ™π΅(π‘₯), if π‘₯ ∈ 𝐡..

We show the three parts.

1.

β€–πœ“(π‘Ž1)βˆ’ πœ“(π‘Ž2)β€– = ‖𝑔(πœ™π΄(π‘Ž1))βˆ’ 𝑔(πœ™π΄(π‘Ž2))β€– (4.44)

≀ ‖𝑔‖Lip β€–πœ™π΄β€–Lip 𝑑(π‘Ž1, π‘Ž2). (4.45)

2. This is clear.

3. Let 𝑏 = 𝑓(π‘Žβ€²). Then 𝑑(π‘Žβ€², 𝑏′) ≀ π‘…π‘Ž, πœ“(π‘Žβ€²) = πœ“(𝑏′). We have

β€–πœ“(π‘Ž)βˆ’ πœ“(𝑏)β€– ≀ β€–πœ“(π‘Ž)βˆ’ πœ“(π‘Žβ€²)β€– βˆ’ β€–πœ“(π‘Žβ€²)βˆ’ πœ“(𝑏′)β€– βˆ’ β€–πœ“(𝑏)βˆ’ πœ“(𝑏′)β€– (4.46)

β€–πœ“(π‘Ž)βˆ’ πœ“(𝑏)β€– ≀ 2(1 +

1

𝛼

)𝐷𝐴𝐷𝐡𝑑(π‘Ž, π‘Ž

β€²) +𝐷𝐡𝑑(𝑏, 𝑏′) (4.47)

𝑑(π‘Ž, π‘Žβ€²) ≀ π›Όπ‘…π‘Ž ≀ 𝛼𝑑(π‘Ž, 𝑏) (4.48)

𝑑(𝑏, 𝑏′) ≀ 𝑑(𝑏, π‘Ž) + 𝑑(π‘Ž, π‘Žβ€²) + 𝑑(π‘Žβ€², 𝑏′) (4.49)

≀ (2 + 𝛼)𝑑(π‘Ž, 𝑏). (4.50)

107

Page 108: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

This shows the first half of (3).

For the other inequality, use the triangle inequality against and get

β€–πœ“(π‘Ž)βˆ’ πœ“(𝑏)β€– β‰₯ β€–πœ“(𝑏)βˆ’ πœ“(𝑏′)β€– βˆ’ β€–πœ“(π‘Žβ€²)βˆ’ πœ“(𝑏′)β€– βˆ’ β€–πœ“(π‘Žβ€²)βˆ’ πœ“(π‘Ž)β€– (4.51)

β‰₯ 𝑑(𝑏, 𝑏′)βˆ’ 2(1 + 𝛼)π·π΄π·π΅π‘…π‘Ž. (4.52)

Let

πœ“π΅ = πœ“ (4.53)

𝛽 = (1 + 𝛼)(2𝐷𝐴𝐷𝐡 + 1) (4.54)

𝛾 =(1

2

)𝛽 (4.55)

πœ“Ξ” : 𝑋 β†’ R (4.56)

πœ“π΄(π‘Ž) = π›Ύπ‘…π‘Ž, π‘Ž ∈ 𝐴 (4.57)

πœ“π΄(𝑏) = βˆ’π›Ύπ‘…π‘, 𝑏 ∈ 𝐡 (4.58)

Ξ¨ : 𝑋 β†’ β„“π‘Ž+𝑏+1 (4.59)

Ξ¨(π‘₯) = πœ“π΄ βŠ• πœ“π΅ βŠ• πœ“Ξ” ∈ β„“π‘Ž+𝑏+12 . (4.60)

For π‘Ž1, π‘Ž2 ∈ 𝐴,

β€–Ξ¨(π‘Ž1)βˆ’Ξ¨(π‘Ž2)β€– β‰₯ ‖Ψ𝐴(π‘Ž1)βˆ’Ξ¨π΄(π‘Ž2)β€– β‰₯ 𝑑(π‘Ž1, π‘Ž2).

For π‘Ž ∈ 𝐴, 𝑏 ∈ 𝐡,

β€–Ξ¨(π‘Ž)βˆ’Ξ¨(𝑏)β€–2 = β€–πœ“π΄(π‘Ž)βˆ’ πœ“π΄(𝑏)β€–2 + β€–πœ“π΅(π‘Ž)βˆ’ πœ“π΅(𝑏)β€–2 + β€–πœ“π΄(π‘Ž)βˆ’ πœ“π΄(𝑏)β€–2 .

β€–πœ“π΄(π‘Ž)βˆ’ πœ“π΄(𝑏)β€– β‰₯ 𝑑(π‘Ž, 𝑏)βˆ’ 𝛽𝑅𝑏 (4.61)

β€–πœ“π΅(π‘Ž)βˆ’ πœ“π΅(𝑏)β€– β‰₯ 𝑑(π‘Ž, 𝑏)βˆ’ π›½π‘…π‘Ž (4.62)

β€–πœ“π΄(π‘Ž)βˆ’ πœ“π΄(𝑏)β€– = 𝛾(π‘…π‘Ž +𝑅𝑏). (4.63)

Claim 4.6.6. We have β€–πœ“(π‘Ž)βˆ’ πœ“(𝑏)β€– β‰₯ 𝑑(π‘Ž, 𝑏).

Proof. Without loss of generality π‘…π‘Ž βŠ† 𝑅𝑏. Consider 3 cases.

1. 𝛽𝑅𝑏 ≀ 𝑑(π‘Ž, 𝑏).

2. π›½π‘…π‘Ž ≀ 𝑑(π‘Ž, 𝑏) ≀ 𝛽𝑅𝑏.

3. 𝑑(π‘Ž, 𝑏) ≀ π›½π‘…π‘Ž.

108

Page 109: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Consider case 2. The other cases are similar.

β€–πœ“(π‘Ž)βˆ’ πœ“(𝑏)β€–2 β‰₯ (𝑑(π‘Ž, 𝑏)βˆ’ π›½π‘…π‘Ž)2 + 𝛽2(π‘…π‘Ž +𝑅𝑏)

2/2 (4.64)

= 𝑑(π‘Ž, 𝑏)2 βˆ’ 2𝛽𝑑(π‘Ž, 𝑏)π‘…π‘Ž +𝛽2

2(3𝑅2

π‘Ž + 2π‘…π‘Žπ‘…π‘ +𝑅2𝑏) (4.65)

β‰₯ 𝑑(π‘Ž, 𝑏)2 βˆ’ 2π›½π‘…π‘Žπ‘…π‘ +𝛽2

2(3𝑅2

π‘Ž + 2π‘…π‘Žπ‘…π‘ +𝑅2𝑏) (4.66)

= 𝑑(π‘Ž, 𝑏)2 + 𝛽2((√3π‘…π‘Ž βˆ’π‘…π‘)

2 + 2(√3 +π‘…π‘Žπ‘…π‘))/2 (4.67)

> 𝑑(π‘Ž, 𝑏) (4.68)

For π‘Ž1, π‘Ž2 ∈ 𝑋,

β€–Ξ¨(π‘Ž1)βˆ’Ξ¨(π‘Ž2)β€– = β€–πœ“π΄(π‘Ž1) + πœ“π΄(π‘Ž2)β€–2 + β€–πœ“π΅(π‘Ž1)βˆ’ πœ“π΅(π‘Ž2)β€–2 + β€–πœ“π΅(π‘Ž1)βˆ’ πœ“π΄(π‘Ž2)β€–

(4.69)

≀�𝐷2𝐴 + 4

(1 +

1

𝛼

)2

𝐷2𝐴𝐷

2𝐡

�𝑑(π‘Ž1, π‘Ž2)

2 + 𝛾2(π‘…π‘Ž1 βˆ’π‘…π‘Ž2) (4.70)

≀�𝐷2𝐴 + 4

(1 +

1

𝛼

)2

𝐷2𝐴𝐷

2𝐡 + 𝛾2

�𝑑(π‘Ž1, π‘Ž2)

2 (4.71)

|π‘…π‘Ž βˆ’π‘…π‘Ž2| ≀ 𝑑(π‘Ž1, π‘Ž2) (4.72)

β€–Ξ¨(𝑏1)βˆ’Ξ¨(𝑏2)β€–2 ≀�𝐷2𝐡 + 4

(1 +

1

𝛼

)2

𝐷2𝐴𝐷

2𝐡 + 𝛾2

�𝑑(π‘Ž, 𝑏)2 (4.73)

|Ξ¨(π‘Ž)βˆ’Ξ¨(𝑏)|2 ≀ β€–πœ“π΄(π‘Ž)βˆ’ πœ“π΅(𝑏)β€–2 + β€–πœ“π΅(π‘Ž)βˆ’ πœ“π΅(𝑏)β€–2 + β€–πœ“π΄(π‘Ž)βˆ’ πœ“(𝑏)β€–2 (4.74)

≀ Β· Β· Β· 𝛾2(π‘…π‘Ž +𝑅𝑏)2 ≀ Β· Β· Β· (4.75)

π‘…π‘Ž, 𝑅𝑏 ≀ 𝑑(π‘Ž, 𝑏).

109

Page 110: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

110

Page 111: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Chapter 5

Lipschitz extension

β€˜

1 Introduction and Problem Setup

We give a presentation of the paper β€œOn Lipschitz Extension From Finite Subsets”, by AssafNaor and Yuval Rabani, (2015). For the convenience of the reader referencing the originalpaper, we have kept the numbering of the lemmas the same.

Consider the setup where we have a metric space (𝑋, 𝑑𝑋) and a Banach space (𝑍, β€– Β· ‖𝑍).For a subset 𝑆 βŠ† 𝑋, consider a 1-Lipschitz function 𝑓 : 𝑆 β†’ 𝑍. Our goal is to extend 𝑓to 𝐹 : 𝑋 β†’ 𝑍 without experiencing too much growth in the Lipschitz constant ‖𝐹‖𝐿𝑖𝑝 over‖𝑓‖𝐿𝑖𝑝.

Definition 5.1.1: 𝑒(𝑋,𝑆, 𝑍) and its variants.Define 𝑒(𝑋,𝑆, 𝑍) to be the infimum over the sequence of 𝐾 satisfying ‖𝐹‖𝐿𝑖𝑝 ≀ 𝐾‖𝑓‖𝐿𝑖𝑝(i.e., 𝑒(𝑋,𝑆, 𝑍) is the least upper bound for

‖𝐹‖𝐿𝑖𝑝

‖𝑓‖𝐿𝑖𝑝for a particular 𝑆, 𝑋, 𝑍).

Then, define 𝑒(𝑋,𝑍) to be the supremum over all subsets 𝑆 for 𝑒(𝑋,𝑆, 𝑍): So of allsubsets, what’s the largest least upper bound for the ratio of Lipschitz constants?

We may also want to consider supremums over 𝑒(𝑋,𝑆, 𝑍) for 𝑆 with a fixed size. Wecan formulate this in two ways. 𝑒𝑛(𝑋,𝑍) is the supremum of 𝑒(𝑋,𝑆, 𝑍) over all 𝑆 such that|𝑆| = 𝑛. We can also describe π‘’πœ–(𝑋,𝑍) as the supremum of 𝑒(𝑋,𝑆, 𝑍) over all 𝑆 which areπœ–-discrete in the sense that 𝑑𝑋(π‘₯, 𝑦) β‰₯ πœ– Β· diam(𝑆) for distinct π‘₯, 𝑦 ∈ 𝑆 and some πœ– ∈ [0, 1].

Definition 5.1.2: Absolute extendability.We define ae(𝑛) to be the supremum of 𝑒𝑛(𝑋,𝑍) over all possible metric spaces (𝑋, 𝑑𝑋) andBanach spaces (𝑍, β€– Β· ‖𝑍). Identically, ae(πœ–) is the supremum of π‘’πœ–(𝑋,𝑍) over all (𝑋, 𝑑𝑋)and (𝑍, β€– Β· ‖𝑍).

From now on, we will primarily discuss subsets 𝑆 βŠ† 𝑋 with size |𝑆| = 𝑛. Bounding thesupremum, the absolute extendability ae(𝑛) < 𝐾 allows us to make general claims about the

111

Page 112: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

extendability of maps from metric spaces into Banach spaces. Any Banach-space valued 1-Lipschitz function defined on metric space (𝑀,𝑑𝑀) can therefore be extended to any metricspace𝑀 β€² such that𝑀 β€² contains𝑀 (up to isometry; as long as you can embed𝑀 in𝑀 β€² withan injective distance preserving map) such that the Lipschitz constant of the extension is atmost 𝐾.

Therefore, for the last thirty years, it has been of interest to understand upper and lowerbounds on ae(𝑛), as we want to understand the asymptotic behavior as 𝑛 β†’ ∞. In the1980s, the following upper and lower bound were given by Johnson and Lindenstrauss andSchechtman: √

log 𝑛

log log 𝑛. ae(𝑛) . log 𝑛

In 2005, the upper bound was improved:√log 𝑛

log log 𝑛. ae(𝑛) .

log 𝑛

log log 𝑛

In this talk, we improve the lower bound for the first time since 1984 toÈlog 𝑛 . ae(𝑛) .

log 𝑛

log log 𝑛

1.1 Why do we care?

This improvement is of interest primarily not because of the removal of a√log log 𝑛 term in

the denominator. It is due to the fact that the approach taken to get the lower bound pro-vided by Johnson-Lindenstrauss 1984 has an inherent limitation. The approach of Johnson-Lindenstrauss to get the lower bound is to prove the nonexistance of linear projections ofsmall norm. By considering a specific case for 𝑓,𝑋, 𝑆, 𝑍, we can get a lower bound on ae(𝑛).Consider a Banach space (π‘Š, β€–Β·β€–π‘Š ) and let π‘Œ βŠ† π‘Š be a π‘˜-dimensional linear subspace ofπ‘Šwith π‘πœ– an πœ–-net in the unit sphere of π‘Œ , and then define π‘†πœ– = π‘πœ–βˆͺ{0}. Fix πœ– ∈ (0, 1/2). Wetake 𝑓 : π‘†πœ– β†’ π‘Œ to be the identity mapping, and wish to find an extension to 𝐹 : π‘Š β†’ π‘Œ .Then, in our setup, we let 𝑋 = π‘Š , 𝑆 = π‘†πœ–, 𝑍 = π‘Œ . We seek to bound the magnitude of theLipschitz constant of 𝐹 , call it 𝐿. Johnson-Lindenstrauss prove that for πœ– . 1

π‘˜2, there exists

a linear projection 𝑃 : π‘Š β†’ π‘Œ with ‖𝑃‖ . 𝐿. We can now proceed to lower bound 𝐿 bylower-bounding ‖𝑃‖ for all 𝑃 . The classical Kadec’-Snobar theorem says that there alwaysexists a projection with ‖𝑃‖ ≀

βˆšπ‘˜. Therefore, the best (largest) possible lower bound we

could get will be 𝐿 &βˆšπ‘˜ by Kadec’-Snobar. But this is bad:

Taking 𝑛 = |π‘†πœ–|, by bounds on πœ–-nets we get π‘˜ ≍ log𝑛log(1/πœ–)

which implies

𝐿 &

√log 𝑛

log(1/πœ–)

In order to get the lower bound on ae(𝑛) of√log 𝑛, we must take πœ– to be a universal constant.

However, from a lemma by Benyamini (in our current setting), 𝐿 . π‘’πœ–(𝑋,𝑍) . 1/πœ– = 𝑂(1),which means that any lower bound we get on 𝐿 will be too small (and won’t even tend to∞). Therefore, we must make use of nonlinear theory to get the

βˆšπ‘› lower bound on ae(𝑛).

112

Page 113: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

1.2 Outline of the Approach

Let us formally state the theorem, and then give the approach to the proof.

Theorem 5.1.3. Theorem 1.For every 𝑛 ∈ N we have ae(𝑛) &

√log 𝑛.

We give a metric space 𝑋, a Banach space 𝑍, a subset 𝑆 βŠ† 𝑋, a function 𝑓 : 𝑆 β†’ 𝑍such that 𝑓 extends to 𝐹 : 𝑋 β†’ 𝑍 where ‖𝐹‖𝐿𝑖𝑝 ≀ 𝐾‖𝑓‖𝐿𝑖𝑝.

Let 𝑉𝐺 be the vertices of a finite graph 𝐺 with distance metric the shortest path metric𝑑𝐺 where edges all have length 1.

We define our metric space 𝑋 = (𝑉𝐺, π‘‘πΊπ‘Ÿ(𝑆)) where πΊπ‘Ÿ(𝑆) is the π‘Ÿ-magnification of theshortest path metric on 𝑉𝐺. 𝑆 is an 𝑛-vertex subset (𝑆, π‘‘πΊπ‘Ÿ(𝑆)). Our Banach space 𝑍 =(R𝑋0 , β€–Β·β€–π‘Š1(𝑋,π‘‘πΊπ‘Ÿ(𝑆)

) is equipped with the Wasserstein-1 norm induced by the π‘Ÿ-magnification

of the shortest path metric on the graph. Note that R𝑋0 is just weight distributions on thevertices of 𝑋 which sum to zero in the image. Our 𝑓 : 𝑆 β†’ R𝑆0 βŠ† 𝑍, and we extend to𝐹 : 𝑋 β†’ 𝑍. We will show how to choose π‘Ÿ and |𝑆| optimally to get the result.

The rest of my section of the talk will give the requisite definitions and lemmas tounderstand the full proof.

2 π‘Ÿ-Magnification

Definition 5.2.1: π‘Ÿ-magnification of a metric space.Given metric space (𝑋, 𝑑𝑋) and π‘Ÿ > 0, for every subset 𝑆 βŠ† 𝑋 we define π‘‹π‘Ÿ(𝑆) as a metricspace on the points of 𝑋 equipped with the following metric:

π‘‘π‘‹π‘Ÿ(𝑆)(π‘₯, 𝑦) = 𝑑𝑋(π‘₯, 𝑦) + π‘Ÿ|{π‘₯, 𝑦} ∩ 𝑆|

and where π‘‘π‘‹π‘Ÿ(𝑆)(π‘₯, π‘₯) = 0. All this is saying is that when we have distinct points π‘₯, 𝑦 ∈ 𝑆,we have the metric is 2π‘Ÿ+𝑑𝑋(π‘₯, 𝑦), when one point is in 𝑆 and one point is outside, we haveπ‘Ÿ + 𝑑𝑋(π‘₯, 𝑦), and when both π‘₯, 𝑦 are outside, the metric is unchanged.

The significance of this definition is as follows: It’s easier for functions on 𝑆 to be Lipschitz(we enlarge the denominator) without affecting functions on 𝑋 βˆ– 𝑆. Thus, there are morepotential 𝑓 we can draw from which satisfy 1-Lipschitzness which can have potentially largeLipschitz extensions (i.e., large 𝐾) since we don’t make it easier to be Lipschitz on 𝑋 βˆ– 𝑆(which we must deal with in the extension space).

However, we can’t make π‘Ÿ too large: the minimum distance between π‘₯, 𝑦 in 𝑆 becomesclose to diam(𝑆) under π‘Ÿ-magnification as π‘Ÿ increases. Let us assume the minimum distancebetween π‘₯, 𝑦 is 1 (as it would be in an undirected graph with an edge between π‘₯, 𝑦 underthe shortest path metric). Particularly, for distinct π‘₯, 𝑦 ∈ 𝑆, since diam(𝑆, π‘‘π‘‹π‘Ÿ(𝑆)) = 2π‘Ÿ +

diam(𝑆, 𝑑𝑋),

π‘‘π‘‹π‘Ÿ(𝑆)(π‘₯, 𝑦) β‰₯ 2π‘Ÿ + 1 =2π‘Ÿ + 1

2π‘Ÿ + diam(𝑆, 𝑑𝑋)Β· diam(𝑆, π‘‘π‘‹π‘Ÿ(𝑆))

113

Page 114: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Then recall that π‘’πœ–(𝑋,𝑍) is the supremum over 𝑆 such that are πœ–-discrete, where here,πœ– = 2π‘Ÿ+1

2π‘Ÿ+diam(𝑆,𝑑𝑋). Earlier we saw a bound that

π‘’πœ–(𝑋,𝑍) . 1/πœ– =2π‘Ÿ + diam(𝑆, 𝑑𝑋)

2π‘Ÿ + 1≀ 1 + diam(𝑆, 𝑑𝑋)

π‘Ÿ

Thus, if we make π‘Ÿ too large, we again are bounding π‘’πœ–(𝑋,𝑍) . 1 = 𝑂(1), which means ourchoice of 𝑋 and 𝑍 is not good to get a large lower bound (again, we’re not even going to∞).

Thus we must balance our choice of π‘Ÿ appropriately.

3 Wasserstein-1 norm

Now we come to the second part of our choice of 𝑍. Note that we will define R𝑋0 to be theset of functions on the points of 𝑋 such that for each 𝑓 ∈ R𝑋0 ,

βˆ‘π‘₯βˆˆπ‘‹ 𝑓(π‘₯) = 0. We use 𝑒π‘₯ to

denote the indicator weight map with 1 at point π‘₯ and 0 everywhere else.

Definition 5.3.1: Wasserstein-1 Norm.The Wasserstein-1 norm is the norm induced by the following origin-symmetric convex bodyin finite metric space (𝑋, 𝑑𝑋):

𝐾(𝑋,𝑑𝑋) = conv

βŒ‰π‘’π‘₯ βˆ’ 𝑒𝑦𝑑𝑋(π‘₯, 𝑦)

: π‘₯, 𝑦 ∈ 𝑋, π‘₯ = 𝑦

Β«This is a unit ball on R𝑋0 . We denote the induced norm by β€– Β· β€–π‘Š1(𝑋,𝑑𝑋).

We can give an equivalent (proven with the Kantorovich-Rubinstein duality theorem)definition of the Wasserstein-1 distance:

Definition 5.3.2: Wasserstein-1 distance and norm.Let Ξ (πœ‡, 𝜈) be all measures on πœ‹ on 𝑋 ×𝑋 such thatβˆ‘

π‘¦βˆˆπ‘‹πœ‹(𝑦, 𝑧) = 𝜈(𝑧)

for all 𝑧 ∈ 𝑋 and βˆ‘π‘§βˆˆπ‘‹

πœ‹(𝑦, 𝑧) = πœ‡(𝑦)

for all 𝑦 ∈ 𝑋. Then, the Wasserstein-1 distance (earthmover) is

π‘Š 𝑑𝑋1 (πœ‡, 𝜈) = inf

πœ‹βˆˆΞ (πœ‡,𝜈)

βˆ‘π‘₯,π‘¦βˆˆπ‘‹

𝑑𝑋(π‘₯, 𝑦)πœ‹(π‘₯, 𝑦)

In the case that πœ‡ = 𝜈, we automatically have (πœ‡Γ— 𝜈)/πœ‡(𝑋) ∈ Ξ (πœ‡, 𝜈) (normalizing by onemeasure trivially gives the other), so Ξ  is nonempty. The norm induced by this metric for𝑓 ∈ R𝑋0 is

β€–π‘“β€–π‘Š1(𝑋,𝑑𝑋) = π‘Š 𝑑𝑋1 (𝑓+, π‘“βˆ’)

114

Page 115: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

where 𝑓+ = max(𝑓, 0) and π‘“βˆ’ = max(βˆ’π‘“, 0) (sinceβˆ‘π‘₯βˆˆπ‘‹ 𝑓(π‘₯) = 0, we need to make sure

that πœ‡, 𝜈 are both nonnegative measure). Futhermore, we haveβˆ‘π‘₯βˆˆπ‘‹ 𝑓

+(π‘₯) =βˆ‘π‘₯βˆˆπ‘‹ 𝑓

βˆ’(π‘₯)(same total mass), which means that Ξ (𝑓+, π‘“βˆ’) is nonempty.

Definition 5.3.3: β„“1(𝑋) norm.Note that using standard notation, we can also define an β„“1 norm on 𝑓 by

‖𝑓‖ℓ1(𝑋) =βˆ‘π‘₯βˆˆπ‘‹

|𝑓(π‘₯)| =βˆ‘π‘₯βˆˆπ‘‹

𝑓+(π‘₯) + π‘“βˆ’(π‘₯)

Thus, for our restrictions on 𝑓 ,βˆ‘π‘₯βˆˆπ‘‹ 𝑓

+(π‘₯) =βˆ‘π‘₯βˆˆπ‘‹ 𝑓

βˆ’(π‘₯) = ‖𝑓‖ℓ1(𝑋)/2.

Now we give a simple lemma which gives bounds for the Wasserstein-1 norm induced bythe π‘Ÿ-magnification of a metric on 𝑋.

Lemma 5.3.4. Lemma 7.For (𝑋, 𝑑𝑋) a finite metric space, we have

1. ‖𝑒π‘₯ βˆ’ π‘’π‘¦β€–π‘Š1(𝑋,𝑑𝑋) = 𝑑𝑋(π‘₯, 𝑦) for every π‘₯, 𝑦 ∈ 𝑋.

2. For all 𝑓 ∈ R𝑋0 ,1

2min

π‘₯,π‘¦βˆˆπ‘‹;π‘₯ =𝑦𝑑𝑋(π‘₯, 𝑦)‖𝑓‖ℓ1(𝑋) ≀ β€–π‘“β€–π‘Š1(𝑋,𝑑𝑋) ≀

1

2diam(𝑋, 𝑑𝑋)‖𝑓‖ℓ1(𝑋)

3. For every π‘Ÿ > 0, 𝑆 βŠ† 𝑋, for all 𝑓 ∈ R𝑆0 ,

π‘Ÿβ€–π‘“β€–β„“1(𝑆) ≀ β€–π‘“β€–π‘Š1(𝑆,π‘‘π‘‹π‘Ÿ(𝑆)) β‰€οΏ½π‘Ÿ + diam(𝑋, 𝑑𝑋)

2

�‖𝑓‖ℓ1(𝑆)

Proof. 1. This follows directly from the unit ball interpretation of the Wasserstein-1 norm,since 𝑒π‘₯βˆ’π‘’π‘¦

𝑑𝑋(π‘₯,𝑦)is by the first definition on the unit ball.

2. Let π‘š = minπ‘₯,𝑦 𝑑𝑋(π‘₯, 𝑦) > 0. For distinct π‘₯, 𝑦 ∈ 𝑋, we have

maxπ‘₯,π‘¦βˆˆπ‘‹;π‘₯ =𝑦

𝑒π‘₯ βˆ’ 𝑒𝑦𝑑𝑋(π‘₯, 𝑦)

β„“1(𝑋)

≀ maxπ‘₯,π‘¦βˆˆπ‘‹;π‘₯ =𝑦

‖𝑒π‘₯ βˆ’ 𝑒𝑦‖ℓ1(𝑋)

π‘š=

2

π‘š

since 0 < π‘š ≀ 𝑑𝑋(π‘₯, 𝑦) and 1 + 1 = 2. Therefore 𝑒π‘₯βˆ’π‘’π‘¦π‘‘π‘‹(π‘₯,𝑦)

∈ 2π‘šπ΅β„“1(𝑋). These elements

span 𝐾𝑋,𝑑𝑋 , so we have 𝐾𝑋,𝑑𝑋 βŠ† 2π‘šπ΅β„“1(𝑋) and we get the first inequality. The second

inequality follows from

β€–π‘“β€–π‘Š1(𝑋,𝑑𝑋) = infπœ‹βˆˆΞ (𝑓+,π‘“βˆ’)

βˆ‘π‘₯,π‘¦βˆˆπ‘‹

𝑑𝑋(π‘₯, 𝑦)πœ‹(π‘₯, 𝑦) ≀ diam(𝑋, 𝑑𝑋)βˆ‘π‘₯βˆˆπ‘‹

𝑓+(π‘₯) = diam(𝑋, 𝑑𝑋)‖𝑓‖ℓ1(𝑋)/2

3. This inequality is a special case of the previous inequality. We have that for π‘‹π‘Ÿ(𝑆),

π‘š β‰₯ 2π‘Ÿ (so 2π‘š

≀ 1π‘Ÿ) and 1

2diam(𝑋, π‘‘π‘‹π‘Ÿ(𝑆)) ≀ 12

(2π‘Ÿ + diam(𝑋, 𝑑𝑋)

οΏ½=οΏ½π‘Ÿ + diam(𝑋,𝑑𝑋)

2

οΏ½.

Plugging in these estimates give the inequality.

115

Page 116: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

4 Graph theoretic lemmas

4.1 Expanders

We will need several properties of edge-expanders in our proof of the main theorem. For thissection, we fix 𝑛, 𝑑 β‰₯ 3 and let 𝐺 be a connected 𝑛-vertex 𝑑-regular graph. We can imagine𝑑 = 3 in this section, all that matters is that 𝑑 is fixed.

First we record two basic average bounds on distance in the shortest-path metric, denoted𝑑𝐺.

Lemma 5.4.1. Average shortest-path metric and π‘Ÿ-magnified average shortest-path bounds.

1. 𝑑𝐺 lower bound: For nonempty 𝑆 βŠ† 𝑉𝐺,

1

|𝑆|2βˆ‘π‘₯,π‘¦βˆˆπ‘†

𝑑𝐺(π‘₯, 𝑦) β‰₯log |𝑆|4 log 𝑑

2. π‘‘πΊπ‘Ÿ(𝑆) equality: For some 𝑆 βŠ† 𝑉𝐺 and π‘Ÿ > 0,

1

|𝐸𝐺|βˆ‘

(π‘₯,𝑦)∈𝐸𝐺

π‘‘πΊπ‘Ÿ(π‘₯,𝑦) = 1 +2π‘Ÿ|𝑆|𝑛

Proof. 1. The smallest nonzero distance in 𝐺 is at least 1. Thus, the average is boundedbelow by 1

|𝑆|2 |𝑆|(|𝑆|βˆ’1) = 1βˆ’ 1|𝑆| since 𝐺 is connected (shortest case is complete graph

on 𝑛 vertices). Then, 1 βˆ’ 1/π‘Ž β‰₯ (log π‘Ž)/4 log 3 for π‘Ž ∈ [15] (𝑑 = 3 maximizes), sowe proceed assuming |𝑆| β‰₯ 16. Let’s bound the distance in the average. Since 𝐺 is𝑑-regular, for every π‘₯ ∈ 𝑉𝐺 the number of vertices 𝑦 such that 𝑑𝐺(π‘₯, 𝑦) ≀ π‘˜ βˆ’ 1 is atmost

βˆ‘π‘˜βˆ’1𝑖=0 𝑑

𝑖. The rest of the vertices are farther away. Since 1 + Β· Β· Β· + π‘‘π‘˜βˆ’2 < π‘‘π‘˜βˆ’1

we have #{𝑦 : 𝑑𝐺(π‘₯, 𝑦) ≀ π‘˜ βˆ’ 1} ≀ 2π‘‘π‘˜βˆ’1. Choosing π‘˜ = 1 + ⌊log𝑑(|𝑆|/4)βŒ‹ gives that

2π‘‘π‘˜βˆ’1 ≀ |𝑆|2. Therefore

1

|𝑆|2βˆ‘π‘₯,π‘¦βˆˆπ‘†

𝑑𝐺(π‘₯, 𝑦) β‰₯1

|𝑆|2* |𝑆| * |𝑆|/2 * (π‘˜ βˆ’ 1) =

π‘˜ βˆ’ 1

2=

log(|𝑆|/4)2 log 𝑑

β‰₯ log |𝑆|4 log 𝑑

since |𝑆| β‰₯ 16.

2. Let 𝐸1 be edges completely contained in 𝑆 and 𝐸2 be edges partially contained in 𝑆.Because 𝐺 is 𝑑-regular, 2|𝐸1|+ |𝐸2| = 𝑑|𝑆| (2 vertices in 𝑆 for 𝐸1, only 1 vertex for 𝐸2,then divide by 𝑑 for overcounting since each vertex in 𝑆 hits 𝑑 other vertices, and wecount exactly the edges which have at least one vertex in 𝑆). Note that |𝐸𝐺| = 𝑑𝑛/2 bydouble-counting vertices. Then for each edge in 𝐸1 we add 2π‘Ÿ, for each edge in 𝐸2 we

116

Page 117: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

add π‘Ÿ, and otherwise we add 0 to the base distance of an edge, which is 1. Therefore,

1

|𝐸𝐺|βˆ‘

(π‘₯,𝑦)∈𝐸𝐺

π‘‘πΊπ‘Ÿ(π‘₯,𝑦) =((0 + 1)|𝐸𝐺 βˆ– (𝐸1 βˆͺ 𝐸2)|+ (π‘Ÿ + 1)|𝐸2|+ (2π‘Ÿ + 1)|𝐸1|)

|𝐸𝐺|

= 1 +π‘Ÿ(2|𝐸1|+ |𝐸2|)

|𝐸𝐺|= 1 +

π‘Ÿπ‘‘|𝑆|𝑑𝑛/2

= 1 +2π‘Ÿ|𝑆|𝑛

Now we introduce the definition of edge expansion.

Definition 5.4.2: Edge expansion πœ‘(𝐺).𝐺 is a connected 𝑛-vertex 𝑑-regular graph. Consider 𝑆, 𝑇 βŠ† 𝑉𝐺 disjoint subsets. Let𝐸𝐺(𝑆, 𝑇 ) βŠ† 𝐸𝐺 denote the set of edges which bridge 𝑆 and 𝑇 . Then the edge-expansionπœ‘(𝐺) is defined by

πœ‘(𝐺) = sup

βŒ‰πœ‘ : |𝐸𝐺(𝑆, 𝑉𝐺 βˆ– 𝑆)| β‰₯ πœ‘

|𝑆|(π‘›βˆ’ |𝑆|)𝑛2

|𝐸𝐺|,βˆ€π‘† βŠ† 𝑉𝐺, πœ‘ ∈ [0,∞)

Β«We give an equivalent formulation of edge expansion via the cut-cone decomposition:

Lemma 5.4.3. Edge-Expansion: Cut-cone Decomposition of Subsets of β„“1.πœ‘(𝐺) is the largest πœ‘ such that for all β„Ž : 𝑉𝐺 β†’ β„“1,

πœ‘

𝑛2

βˆ‘π‘₯,π‘¦βˆˆπ‘‰πΊ

β€–β„Ž(π‘₯)βˆ’ β„Ž(𝑦)β€–1 ≀1

|𝐸𝐺|βˆ‘

(π‘₯,𝑦)∈𝐸𝐺

β€–β„Ž(π‘₯)βˆ’ β„Ž(𝑦)β€–1

Proof. We will assume this for this talk.

Now we combine Lemma 7 and the cut-cone decomposition to get Lemma 8:

Lemma 5.4.4. Lemma 8.Fix 𝑛 ∈ N and πœ‘ ∈ (0, 1]. Suppose 𝐺 is an 𝑛-vertex graph with edge expansion πœ‘(𝐺) β‰₯ πœ‘.For all nonempty 𝑆 βŠ† 𝑉𝐺 and π‘Ÿ > 0, every 𝐹 : 𝑉𝐺 β†’ R𝑆0 satisfies

1

𝑛2

βˆ‘π‘₯,π‘¦βˆˆπ‘‰πΊ

‖𝐹 (π‘₯)βˆ’πΉ (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀2π‘Ÿ + diam(𝑆, 𝑑𝐺)

(2π‘Ÿ + 1)πœ‘Β· 1

|𝐸𝐺|βˆ‘

(π‘₯,𝑦)∈𝐸𝐺

‖𝐹 (π‘₯)βˆ’πΉ (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))

Basically, whereas before we were bounding average norms (normal and π‘Ÿ-magnified)in the domain space 𝑉𝐺, we are now bounding average norms in the image space using theWasserstein-1 norm induced by the π‘Ÿ-magnification of the shortest path metric on the graph.

Proof. First, plug in β„Ž = 𝐹 into the cut-cone decomposition, where the norm is now definedby β„“1(𝑆). Then, we know that diam(𝑆, π‘‘πΊπ‘Ÿ(𝑆)) = 2π‘Ÿ+diam(𝑆, 𝑑𝐺) and the smallest positivedistance in (𝑆, π‘‘πΊπ‘Ÿ(𝑆)) is 2π‘Ÿ + 1. Applying Lemma 7, every π‘₯, 𝑦 ∈ 𝑉𝐺 satisfy

2π‘Ÿ + 1

2‖𝐹 (π‘₯)βˆ’πΉ (𝑦)β€–β„“1(𝑆) ≀ ‖𝐹 (π‘₯)βˆ’πΉ (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀

2π‘Ÿ + diam(𝑆, 𝑑𝐺)

2‖𝐹 (π‘₯)βˆ’πΉ (𝑦)β€–β„“1(𝑆)

117

Page 118: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Plugging these estimates directly into the cut-cone decomposition

1

𝑛2

βˆ‘π‘₯,π‘¦βˆˆπ‘‰πΊ

‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–β„“1(𝑆) ≀1

πœ‘|𝐸𝐺|βˆ‘

(π‘₯,𝑦)∈𝐸𝐺

‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–β„“1(𝑆)

gives the desired result.

4.2 An Application of Menger’s Theorem

Here, we want to bound below the number of edge-disjoint paths.

Lemma 5.4.5. Lemma 9.Let 𝐺 be an 𝑛-vertex graph and 𝐴,𝐡 βŠ† 𝑉𝐺 be disjoint. Fix πœ‘ ∈ (0,∞) and suppose πœ‘(𝐺) β‰₯ πœ‘.Then

#βŒ‹edge-disjoint paths joining𝐴 and𝐡

{β‰₯ πœ‘|𝐸𝐺|

2𝑛·min{|𝐴|, |𝐡|}

Proof. Let π‘š be the maximal number of edge-disjoint paths joining 𝐴 and 𝐡. By Menger’stheorem from classical graph theory, there exists a subset of edges 𝐸* βŠ† 𝐸𝐺 with |𝐸*| = π‘šsuch that every path in 𝐺 joining π‘Ž ∈ 𝐴 to 𝑏 ∈ 𝐡 contains an edge from 𝐸*.

Now consider the graph 𝐺* = (𝑉𝐺, 𝐸𝐺 βˆ– 𝐸*). In this graph, there are no paths between𝐴 and 𝐡. Now, if we let 𝐢 βŠ† 𝑉𝐺 be the union of all connected components of 𝐺* containingan element of 𝐴, then 𝐴 βŠ† 𝐢 and 𝐡 ∩ 𝐢 = βˆ…. Since we covered the maximal possiblevertices reachable from 𝐴 with 𝐢, all edges between 𝐢 and 𝑉𝐺 βˆ– 𝐢 are in 𝐸*. Therefore,|𝐸𝐺(𝐢, 𝑉𝐺 βˆ– 𝐢)| ≀ |𝐸*| = π‘š. By the definition of expansion,

π‘š β‰₯ |𝐸𝐺(𝐢, π‘‰πΊβˆ–πΆ)| β‰₯ πœ‘max{|𝐢|, π‘›βˆ’ |𝐢|} Β·min{|𝐢|, π‘›βˆ’ |𝐢|}

𝑛2|𝐸𝐺| β‰₯

πœ‘min{|𝐴|, |𝐡|} Β· |𝐸𝐺|2𝑛

since max{|𝐢|, π‘›βˆ’ |𝐢|} β‰₯ 𝑛/2 and since 𝐴 βŠ† 𝐢,𝐡 βŠ† 𝑉𝐺 βˆ– 𝐢, we have min{|𝐢|, π‘›βˆ’ |𝐢|} β‰₯min{|𝐴|, |𝐡|}.

5 Main Proof

We fix 𝑑, 𝑛 ∈ N, πœ‘ ∈ (0, 1) and let 𝐺 be a 𝑑-regular graph on 𝑛 vertices with πœ‘(𝐺) β‰₯ πœ‘. Wealso fix a nonempty subset 𝑆 βŠ‚ 𝑉𝐺 and π‘Ÿ > 0 and define a mapping

𝑓 : (𝑆, π‘‘πΊπ‘Ÿ(𝑆)) ↦→(R𝑆0 , β€– Β· β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))

οΏ½s.t. 𝑓(π‘₯) = 𝑒π‘₯ βˆ’

1

|𝑆|βˆ‘π‘§βˆˆπ‘†

𝑒𝑧 βˆ€ π‘₯ ∈ 𝑆

Then suppose we have that some 𝐹 : 𝑉𝐺 ↦→ R𝑆0 extends 𝑓 and for some 𝐿 ∈ (0,∞) we have

‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀ πΏπ‘‘πΊπ‘Ÿ(𝑆)(π‘₯, 𝑦)

118

Page 119: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

For π‘₯ ∈ 𝑉𝐺, 𝑠 ∈ (0,∞) define

𝐡𝑆(π‘₯) = {𝑦 ∈ 𝑉𝐺 : ‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀ 𝑠}

i.e.e 𝐡𝑠 is the inverse image of 𝐹 of the ball (in the Wasserstein 1-norm) of radius 𝑠 centeredat 𝐹 (π‘₯). By the lemma consequence of Menger’s Theorem, since πœ‘ β‰₯ πœ‘(𝐺) we have

π‘š β‰₯ πœ‘π‘‘

4min{|π‘†βˆ–π΅π‘ (π‘₯)|, |𝐡𝑠(π‘₯)|} (1)

forπ‘š edge-disjoint paths between π‘†βˆ–π΅π‘ (π‘₯) and 𝐡𝑠(π‘₯), i.e. we can find indices π‘˜1, . . . , π‘˜π‘š ∈ Nand vertex sets {𝑧𝑗,1, . . . , 𝑧𝑗,π‘˜π‘—}π‘šπ‘—=1 ∈ 𝑉𝐺 s.t. {𝑧𝑗,1}π‘šπ‘—=1 βŠ‚ π‘†βˆ–π΅π‘ (π‘₯), {𝑧𝑗,π‘˜π‘—}π‘šπ‘—=1 βŠ‚ 𝐡𝑠(π‘₯) (i.e. thebeginnings and ends of paths are in different disjoint subsets) and such that {{𝑧𝑗,1, 𝑧𝑗,𝑖+1 :𝑗 ∈ {1, . . . ,π‘š}∧ 𝑖 ∈ {1, . . . , π‘˜π‘—βˆ’1}} are distinct edges in 𝐸𝐺 (i.e. edge-disjointedness). Nowtake an index subset 𝐽 βŠ‚ {1, . . . ,π‘š} s.t. {𝑧𝑗,1}π‘—βˆˆπ½ are distinct and {𝑧𝑗,1}π‘—βˆˆπ½ = {𝑧𝑖,1}π‘šπ‘–=1. For𝑗 ∈ 𝐽 denote by 𝑑𝑗 the number of 𝑖 ∈ {1, . . . ,π‘š} for which 𝑧𝑗,1 = 𝑧𝑖,1. Then since 𝐺 is𝑑-regular and {{𝑧𝑖,1, 𝑧𝑖,2}}π‘šπ‘–=1 are distinct, max

π‘—βˆˆπ½π‘‘π‘— ≀ 𝑑. Since

βˆ‘π‘—βˆˆπ½

𝑑𝑗 = π‘š, from (1) we have

that

|𝐽 | β‰₯ π‘š

𝑑β‰₯ πœ‘

4min{|π‘†βˆ–π΅π‘ (π‘₯)|, |𝐡𝑠(π‘₯)|} (2)

The quantity |𝐽 | can be upperbounded as follows:

Lemma 5.5.1. Lemma 10: |𝐽 | ≀ max{𝑑16(π‘ βˆ’π‘Ÿ), 16𝑛𝐿𝑑 log 𝑑

log𝑛

(1 + 2π‘Ÿ|𝑆|

𝑛

)}Proof. Assume |𝐽 | ≀ 𝑑16(π‘ βˆ’π‘Ÿ) (otherwise we are done). This is equivalent to

π‘ βˆ’ π‘Ÿ <log |𝐽 |16 log 𝑑

Now since {𝑧𝑗,1}π‘—βˆˆπ½ βŠ‚ 𝑆 and 𝐹 (π‘₯) = 𝑓(π‘₯) βˆ€ π‘₯ ∈ 𝑆 and is an isometry on (𝑆, π‘‘πΊπ‘Ÿ(𝑆)), by thedefinition of the π‘Ÿ-magnified metric

‖𝐹 (𝑧𝑖,1)βˆ’ 𝐹 (𝑧𝑗,1)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) = π‘‘πΊπ‘Ÿ(𝑆)(𝑧𝑖,1, 𝑧𝑗,1) = 2π‘Ÿ + 𝑑𝐺(𝑧𝑖,1, 𝑧𝑗,1)

This gives usβˆ‘π‘—βˆˆπ½

‖𝐹 (𝑧𝑗,1)βˆ’ 𝐹 (𝑧𝑗,π‘˜π‘—)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) =1

2|𝐽 |βˆ‘π‘–,π‘—βˆˆπ½

(‖𝐹 (𝑧𝑖,1)βˆ’ 𝐹 (𝑧𝑖,π‘˜π‘–)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) + ‖𝐹 (𝑧𝑗,1)βˆ’ 𝐹 (𝑧𝑗,π‘˜π‘—)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))

οΏ½β‰₯ 1

2|𝐽 |βˆ‘π‘–,π‘—βˆˆπ½

(‖𝐹 (𝑧𝑖,1)βˆ’ 𝐹 (𝑧𝑗,1)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) + ‖𝐹 (𝑧𝑧𝑖,π‘˜π‘–)βˆ’ 𝐹 (𝑧𝑗,π‘˜π‘—)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))

οΏ½Since {𝑧𝑗,π‘˜π‘—}π‘—βˆˆπ½ βŠ‚ 𝐡𝑠(π‘₯), by the definition of 𝐡𝑠(π‘₯) we have

‖𝐹 (𝑧𝑖,π‘˜π‘–)βˆ’πΉ (𝑧𝑗,π‘˜π‘—)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀ ‖𝐹 (𝑧𝑖,π‘˜π‘–)βˆ’πΉ (π‘₯)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))+‖𝐹 (π‘₯)βˆ’πΉ (𝑧𝑗,π‘˜π‘—)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀ 2𝑠 βˆ€ 𝑖, 𝑗 ∈ 𝐽

119

Page 120: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

so the previous inequality can be further continued asβˆ‘π‘—βˆˆπ½

‖𝐹 (𝑧𝑗,1)βˆ’πΉ (𝑧𝑗,π‘˜π‘—)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) β‰₯1

2|𝐽 |βˆ‘π‘–,π‘—βˆˆπ½

𝑑𝐺(𝑧𝑖,1, 𝑧𝑗,1)βˆ’(π‘ βˆ’π‘Ÿ)|𝐽 | β‰₯ |𝐽 | log |𝐽 |8 log 𝑑

βˆ’(π‘ βˆ’π‘Ÿ)|𝐽 | > |𝐽 | log |𝐽 |16 log 𝑑

where the second inequality above is a property of the expander graph. The same quantitycan be bounded from above using the Lipschitz condition

βˆ‘π‘—βˆˆπ½

‖𝐹 (𝑧𝑗,1)βˆ’ 𝐹 (𝑧𝑗,π‘˜π‘—)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀ πΏβˆ‘π‘—βˆˆπ½

π‘‘πΊπ‘Ÿ(𝑆)(𝑧𝑗,1, 𝑧𝑗,π‘˜π‘—) ≀ πΏβˆ‘π‘—βˆˆπ½

π‘˜π‘—βˆ’1βˆ‘π‘–=1

π‘‘πΊπ‘Ÿ(𝑆)(𝑧𝑗,1, 𝑧𝑗,𝑖+1)

By the edge-disjointedness of the paths (specifically since {{𝑧𝑗,1, 𝑧𝑗,𝑖+1} : 𝑗 ∈ 𝐽 ∧ 𝑖 ∈{1, . . . , π‘˜π‘— βˆ’ 1}}) are distinct edges in 𝐸𝐺, we have that

βˆ‘π‘—βˆˆπ½

π‘˜π‘—βˆ’1βˆ‘π‘–=1

π‘‘πΊπ‘Ÿ(𝑆)(𝑧𝑗,1, 𝑧𝑗,𝑖+1) β‰€βˆ‘

{𝑒,𝑣}∈𝐸𝐺

π‘‘πΊπ‘Ÿ(𝑆)(𝑒, 𝑣) =𝑛𝑑

2

οΏ½1 +

2π‘Ÿ|𝑆|𝑛

οΏ½Everything together gives

𝐿𝑛𝑑

2

οΏ½1 +

2π‘Ÿ|𝑆|𝑛

οΏ½β‰₯ |𝐽 | log |𝐽 |

16 log 𝑑

By the simple fact that π‘Ž log π‘Ž ≀ 𝑏 =β‡’ π‘Ž ≀ 2𝑏log 𝑏

for π‘Ž ∈ [1,∞), 𝑏 ∈ (1,∞), using π‘Ž = |𝐽 |and 𝑏 = 8𝐿𝑛𝑑 log 𝑑

(1 + 2π‘Ÿ|𝑆|

𝑛

)β‰₯ 𝑛 we have that

|𝐽 | ≀ 16𝑛𝐿𝑑 log 𝑑

log 𝑛

οΏ½1 +

2π‘Ÿ|𝑆|𝑛

οΏ½completing the proof.

The lemma has two corollaries, both depending on the following condition:

𝑑16(π‘ βˆ’π‘Ÿ) ≀ πœ‘|𝑆|8

𝐿 ≀ πœ‘|𝑆| log 𝑛128

(1 + 2π‘Ÿ|𝑆|

𝑛

)𝑛𝑑 log 𝑑

(3)

Corollary 5.5.2. Corollary 11: maxπ‘₯βˆˆπ‘‰πΊ

|𝐡𝑠(π‘₯)| < |𝑆|2

Proof. Pick an π‘₯ ∈ 𝑉𝐺. If 𝐡𝑠(π‘₯)βˆ©π‘† is nonempty, then again using the standard estimate onexpander graphs we have that βˆƒ 𝑦, 𝑧 ∈ 𝐡𝑠(π‘₯) ∩ 𝑆 s.t.

𝑑𝐺(𝑦, 𝑧) β‰₯log |𝐡𝑠(π‘₯) ∩ 𝑆|

4 log 𝑑

Since 𝑦, 𝑧 ∈ 𝑆 and 𝐹 (π‘₯) = 𝑓(π‘₯) βˆ€ π‘₯ ∈ 𝑆 and furthermore 𝐹 is an isometry on (𝑆, π‘‘πΊπ‘Ÿ(𝑆)), wehave similarly to before that

𝑑𝐺(𝑦, 𝑧)+2π‘Ÿ = π‘‘πΊπ‘Ÿ(𝑆)(𝑦, 𝑧) = ‖𝐹 (𝑦)βˆ’πΉ (𝑧)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀ ‖𝐹 (𝑦)βˆ’πΉ (π‘₯)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))+‖𝐹 (π‘₯)βˆ’πΉ (𝑧)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀ 2𝑠

120

Page 121: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

where we have used first the triangle inequality and second the fact that 𝑦, 𝑧 ∈ 𝐡𝑠(π‘₯). Thenusing the first inequality in the proof we have

|𝐡𝑠(π‘₯) ∩ 𝑆| ≀ 𝑑8(π‘ βˆ’π‘Ÿ) β‰€βˆšπœ‘|𝑆|8

≀ 2|𝑆|5

where the second inequality comes from the first assumption. Now the above inequalityimplies that |π‘†βˆ–π΅π‘ (π‘₯)| β‰₯ 3|𝑆|

5, which combined with Lemma 10 yields

min

βŒ‰3|𝑆|5, |𝐡𝑠(π‘₯)|

Β«< max

{4𝑑16(π‘ βˆ’π‘Ÿ)

πœ‘,64𝑛𝐿𝑑 log 𝑑

πœ‘ log 𝑛

οΏ½1 +

2π‘Ÿ|𝑆|𝑛

οΏ½}However, the two original assumptions together imply that 3|𝑆|

5is in fact greater than either

value in the maximum, so

|𝐡𝑠(π‘₯)| ≀ max

{4𝑑16(π‘ βˆ’π‘Ÿ)

πœ‘,64𝑛𝐿𝑑 log 𝑑

πœ‘ log 𝑛

οΏ½1 +

2π‘Ÿ|𝑆|𝑛

οΏ½}≀ |𝑆|

2

Corollary 5.5.3. Corollary 12: 𝐿 β‰₯ πœ‘π‘ 

2(1+

diam(𝐺,𝑑𝐺)

2π‘Ÿ

οΏ½(1+ 2π‘Ÿ|𝑆|

𝑛 )

Proof. By the definition of𝐡𝑠(π‘₯) for π‘₯ ∈ 𝑉𝐺 and 𝑦 ∈ π‘‰πΊβˆ–π΅π‘ (π‘₯) we have ‖𝐹 (π‘₯)βˆ’πΉ (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) >𝑠, so

1

𝑛2

βˆ‘π‘₯,π‘¦βˆˆπ‘‰πΊ

‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) β‰₯1

𝑛2

βˆ‘π‘₯βˆˆπ‘‰πΊ

βˆ‘π‘¦βˆˆπ‘‰πΊβˆ–π΅π‘ (π‘₯)

‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))

β‰₯ 𝑠

𝑛2

βˆ‘π‘₯βˆˆπ‘‰πΊ

(π‘›βˆ’ |𝐡𝑠(π‘₯)|) β‰₯ 𝑠

οΏ½1βˆ’

maxπ‘₯βˆˆπ‘‰πΊ

|𝐡𝑠(π‘₯)|

𝑛

οΏ½β‰₯ 𝑠

2

where the last inequality comes from Corollary 11. Then since Lemma 8 gives

1

𝑛2

βˆ‘π‘₯,π‘¦βˆˆπ‘‰πΊ

‖𝐹 (π‘₯)βˆ’ 𝐹 (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆)) ≀2π‘Ÿ + diam(𝑆, 𝑑𝐺)

(2π‘Ÿ + 1)πœ‘|𝐸𝐺|βˆ‘

π‘₯,π‘¦βˆˆπ‘‰πΊβ€–πΉ (π‘₯)βˆ’ 𝐹 (𝑦)β€–π‘Š1(𝑆,π‘‘πΊπ‘Ÿ(𝑆))

≀𝐿(1 + diam(𝐺,𝑑𝐺)

2π‘Ÿ

)πœ‘|𝐸𝐺|

βˆ‘{π‘₯,𝑦}∈𝐸𝐺

π‘‘πΊπ‘Ÿ(𝑆)(π‘₯, 𝑦)

=𝐿(1 + diam(𝐺,𝑑𝐺)

2π‘Ÿ

) (1 + 2π‘Ÿ|𝑆|

𝑛

)πœ‘

where the ineqaulity on the second line is true since 𝐹 is an isometry on (𝑆, 𝑑𝐺) and fromthe Lipschitz constant of 𝑓 and the equality is average length of the π‘Ÿ-magnification ofthe expander graph 𝐺. Combining these inequalities with the previous ones yields thecorollary.

121

Page 122: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

MAT529 Metric embeddings and geometric inequalities

Theorem 5.5.4. Theorem 13: if 0 < π‘Ÿ ≀ diam(𝐺, 𝑑𝐺) then

𝐿 β‰₯πΆπœ‘

1 + π‘Ÿ|𝑆|𝑛

min

⎧⎨⎩ |𝑆| log 𝑛𝑛𝑑 log 𝑑

,16π‘Ÿ2 log 𝑑+ π‘Ÿ log

(πœ‘|𝑆|8

)diam(𝐺, 𝑑𝐺) log 𝑑

⎫⎬⎭Proof. Assume 16π‘Ÿ log 𝑑+log

(πœ‘|𝑆|8

)> 0 (otherwise we are done) and choose 𝑠 = π‘Ÿ+

log(πœ‘|𝑆|8 )

16 log 𝑑

s.t. 𝑠 > 0 and 𝑑16(π‘ βˆ’π‘Ÿ) = πœ‘|𝑆|8. Then the first inequality in (3) is satisfied, so either the second

fails and 𝐿 thus has that expression as a lower bound, or both are satisfied and we have thelower bound in Corollary 12.

Theorem 5.5.5. Theorem 1: for every 𝑛 ∈ N we have ae(𝑛) β‰₯𝐢

√log 𝑛.

Proof. Substituting πœ‘ ≍ 1 and diam(𝐺, 𝑑𝐺) ≍ log𝑛log 𝑑

(the ≍ indicates asymptotically for large

𝑛), and using the assumption 0 < π‘Ÿ ≀ diam(𝐺, 𝑑𝐺), the lower bound given in Theorem 13becomes

𝐿 β‰₯𝐢1

1 + π‘Ÿ|𝑆|𝑛

min

βŒ‰ |𝑆| log 𝑛𝑛𝑑 log 𝑑

,π‘Ÿ(π‘Ÿ log 𝑑+ log |𝑆|)

log 𝑛

Β«Taking 𝑆 βŠ‚ 𝑉𝐺 s.t. |𝑆| =

βŒŠπ‘›βˆšπ‘‘ log π‘‘βˆšlog𝑛

βŒ‹(we must have 𝑛

𝑔𝑒𝑑𝑑 to ensure |𝑆| ≀ 𝑛) and π‘Ÿ ≍ log π‘‘βˆšπ‘‘ log 𝑑

, which gives a lower bound of 𝐿 β‰₯𝐢

√logπ‘›βˆšπ‘‘ log 𝑑

.

Therefore

π‘’βŒŠπ‘›βˆš

𝑑 log π‘‘βˆšlog𝑛

βŒ‹(πΊπ‘Ÿ(𝑆),π‘Š1(𝑆, π‘‘πΊπ‘Ÿ(𝑆))) β‰₯𝐢

√log π‘›βˆšπ‘‘ log 𝑑

which completes the proof for fixed 𝑑.

122

Page 123: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

List of Notations

𝛿𝑋 β†’Λ“π‘Œ (πœ€) supremum over all those 𝛿 > 0 such that for every 𝛿-net 𝒩𝛿 in 𝐡𝑋 , πΆπ‘Œ (𝒩𝛿) β‰₯(1βˆ’ πœ€)πΆπ‘Œ (𝑋) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

R𝜎 {(π‘₯1, . . . , π‘₯𝑛) ∈ Rπ‘š : π‘₯𝑖 = 0 if 𝑖 ∈ 𝜎} . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

srank(𝐴) stable rank,�

‖𝐴‖𝑆2

β€–π΄β€–π‘†βˆž

οΏ½2. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

‖𝐴‖𝑆𝑝Schatten-von Neumann 𝑝-norm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

πΆπ‘Œ (𝑋) infimum of 𝐷 where 𝑋 imbeds into π‘Œ with distortion 𝐷 . . . . . . . . . . . . . . . . . . . . . . . . 16

123

Page 124: Metric embeddings and geometric inequalitiesnaor/mat529.pdfThe topic of this course is geometric inequalities with applications to metric embeddings; and we are actually going to do

Index

Bourgain’s almost extension theorem, 49

Corson-Klee Lemma, 13cotype π‘ž, 9

discretization modulus, 15, 16distortion, 15

Enflo type, 82Enflo type 𝑝, 11

finitely representable, 7

Jordan-von Neumann Theorem, 9

Kadec’s Theorem, 10Kahane’s inequality, 9Ky Fan maximum principle, 23

local properties, 7

nonlinear Hahn-Banach Theorem, 73nontrivial type, 8

parallelogram identity, 9Pietsch Domination Theorem, 30Poisson kernel, 50, 57preprocessing, 102

Rademacher type, 82restricted invertibility principle, 19rigidity theorem, 10

Sauer-Shelah, 33Schatten-von Neumann 𝑝-norm, 20shattering set, 33stable 𝑝th rank, 22stable rank, 21symmetric norm, 25

type, 8

uniformly homeomorphic, 10

124