Lecture 1 Annotated
Lecture 1 Annotated
1
Motivation For this Class
2
Motivation For this Class
2
Motivation For this Class
2
A New Paradigm for Algorithm Design
3
A New Paradigm for Algorithm Design
3
A New Paradigm for Algorithm Design
For example:
4
A New Paradigm for Algorithm Design
For example:
4
A New Paradigm for Algorithm Design
For example:
{
• How can Twitter rapidly detect if an incoming Tweet is an
exact duplicate of another Tweet made in the last year?
Given that no machine can store all Tweets made in a year.
• How can Google estimate the number of unique search
queries that are made in a given week? Given that no
machine can store the full list of queries.
4
A New Paradigm for Algorithm Design
For example:
4
Motivation for This Class
5
Motivation for This Class
5
Motivation for This Class
6
What We’ll Cover
6
What We’ll Cover
6
What We’ll Cover
6
What We’ll Cover
7
What We’ll Cover
7
What We’ll Cover
o
Section 2: Spectral Methods
7
What We’ll Cover
=
dimensionality reduction.
• The singular value decomposition (SVD) and its applications to
PCA, low-rank approximation, LSI, MDS, …
• Spectral graph theory. Spectral clustering, community detection,
J ]
network visualization.
• Computing the SVD on large matrices via iterative methods.
7
What We’ll Cover
– Michael Stonebraker
7
What We’ll Cover
8
What We’ll Cover
Section 3: Optimization
8
What We’ll Cover
Section 3: Optimization
8
What We’ll Cover
Section 3: Optimization
E
• Gradient descent. Analysis for convex functions.
• Stochastic and online gradient descent.
• Focus on convergence analysis.
8
What We’ll Cover
Section 3: Optimization
8
Important Topics We Won’t Cover
9
Important Topics We Won’t Cover
• Systems/Software Tools.
9
Important Topics We Won’t Cover
• Systems/Software Tools.
9
Important Topics We Won’t Cover
• Systems/Software Tools.
9
Important Topics We Won’t Cover
• Systems/Software Tools.
9
Style of the Course
10
Style of the Course
10
Style of the Course
10
Style of the Course
10
Style of the Course
10
Style of the Course
See Moodle page for this link if you lose it, or search my name
and follow the link from my homepage.
Moodle will be used for weekly quizzes, but the course page for
mostly everything else.
11
Personnel
[
• Forsad Al Houssain
• An La
• Mohit Yadav
• The sections will closely parallel each other, and share the
same TAs.
• You may attend Prof. McGregor’s lectures on Zoom if it is
helpful.
• See Moodle for the Zoom link.
13
Piazza and Participation
14
Piazza and Participation
15
Homework
16
Homework
[
We will have 5 problem sets, which you may complete in
groups of up to 3 students.
16
Homework
16
Weekly Quizzes
17
Weekly Quizzes
[
• Will also include free response check-in questions to get
your feedback on how the course is going, what material
from the past week you find most confusing, interesting,
etc.
17
Grading
Grade Breakdown:
• Problem Sets (5 total): 40%, weighted equally.
• Weekly Quizzes: 10%, weighted equally. drop lowest
• Midterm (October 20th, in class): 25%.
• Final (December 14th, 10:30am - 12:30pm): 25%.
• Extra Credit: Up to 5% for participation, and more
available on problem sets and exams.
18
Grading
Grade Breakdown:
• Problem Sets (5 total): 40%, weighted equally.
• Weekly Quizzes: 10%, weighted equally.
• Midterm (October 20th, in class): 25%.
• Final (December 14th, 10:30am - 12:30pm): 25%.
• Extra Credit: Up to 5% for participation, and more
available on problem sets and exams.
Academic Honesty:
• A first violation cheating on a homework, quiz, or other
assignment will result in a 0 on that assignment.
• A second violation, or cheating on an exam will result in
failing the class.
• For fairness, I adhere very strictly to these policies.
18
Disability Services and Accomodations
19
Disability Services and Accomodations
20
Section 1: Randomized Methods & Sketching
21
Some Probability Review
22
Some Probability Review
a -
Consider a random X variable taking values in some finite set
S ⊂ R. E.g., for a random dice roll, S = {1, 2, 3, 4, 5, 6}.
22
Some Probability Review
=
3.5
§ y
22
Some Probability Review
f-oo
- →
f-(I-3.5ft I f .(63.5¥?
12.3.55t .
22
Some Probability Review
22
Independence
§ x
£
=
24
Independence
Y i , T
Uses -
Pr(X = s ∩ Y = t) = Pr(X = s) · Pr(Y = t).
-
E T
Y t
P r ( A I B ) : P r (A)
Pruss?;¥Y¥÷Du,
Grass n e t n 2-= D :
25
Linearity of Expectation and Variance
F
E[X + Y] = E[X] + E[Y]
and
Var[X + Y] = Var[X] + Var[Y].
26
Linearity of Expectation
X :Y
E[X + Y] = E[X] + E[Y]
, dwny¥
I tC x + T ) s
1 $f i x ] s z . LEX
27
Linearity of Expectation
27
Linearity of Expectation
27
Linearity of Expectation
E S ET
E[X + Y] = E[X] + E[Y] for any random variables X and Y.
Proof:
""
E[X + Y] = Pr(X = s ∩ Y = t) · (s + t)
- .
= . s∈S t∈T
27
Linearity of Expectation
s.ES?+PrCx=snY:D
-
P r( K s )
27
Linearity of Expectation
27
Linearity of Expectation
= Pr(X = s) · s + Pr(Y = t) · t
-s∈S t∈T
(law of total probability)
HID t EM
27
Linearity of Expectation
27