0% found this document useful (0 votes)
4 views

02_Introduction_To_Probability

UCL Machine Vision notes

Uploaded by

weinana1005
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views

02_Introduction_To_Probability

UCL Machine Vision notes

Uploaded by

weinana1005
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 28

Computer vision: models,

learning and inference


Chapter 2
Introduction to probability
Photo by JaypeeOnline (copyright, all rights reserved)

• What color are stop signs?


2
Random variables
• A random variable x denotes a quantity that is
uncertain
• May be result of experiment (flipping a coin) or real
world measurements (measuring temperature)
• Observing more instances of x, we get different values
• Some values occur more than others and this
information is captured by a probability distribution

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 3
Discrete Random Variables

Photo by DaylandS

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 4
Continuous Random Variable

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 5
Joint Probability
• Consider two random variables x and y
• If we observe multiple paired instances, then some
combinations of outcomes are more likely than others

• This is captured in the joint probability distribution


• Written as Pr(x,y)
• Can read Pr(x,y) as “probability of x and y”

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 6
Joint Probability

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 7
Marginalization
We can recover the probability distribution of any variable in a joint distribution
by integrating (or summing) over the other variables

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 8
Marginalization
We can recover the probability distribution of any variable in a joint distribution
by integrating (or summing) over the other variables

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 9
Marginalization
We can recover the probability distribution of any variable in a joint distribution
by integrating (or summing) over the other variables

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 10
Marginalization
We can recover the probability distribution of any variable in a joint distribution
by integrating (or summing) over the other variables

Works in higher dimensions as well – leaves joint distribution between


whatever variables are left

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 11
Conditional Probability
• Conditional probability of x given that y=y1 is relative
propensity of variable x to take different outcomes given that
y is fixed to be equal to y1.
• Written as Pr(x|y=y1)

12
Computer vision: models, learning and inference. ©2011 Simon J.D. Prince
Conditional Probability
• Conditional probability can be extracted from joint probability
• Extract appropriate slice and normalize

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 13
Conditional Probability

• More usually written in compact form

• Can be re-arranged to give

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 14
Conditional Probability

• This idea can be extended to more than two


variables

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 15
Bayes’ Rule
From before:

Combining:

Re-arranging:

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 16
Bayes’ Rule Terminology
Likelihood – propensity for Prior – what we know
observing a certain value of about y before seeing x
x given a certain value of y

Posterior – what we Evidence –a constant to


know about y after ensure that the left hand
seeing x side is a valid distribution
Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 17
Independence
• If two variables x and y are independent then variable x tells
us nothing about variable y (and vice-versa)

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 18
Independence
• If two variables x and y are independent then variable x tells
us nothing about variable y (and vice-versa)

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 19
Independence
• When variables are independent, the joint factorizes into a
product of the marginals:

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 20
Expectation
Expectation tells us the expected or average value of some
function f [x], taking into account the distribution of x

Definition:

"Thermometer" by matsuyuki is licensed


under CC BY-SA 2.0

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 21
Expectation
Expectation tells us the expected or average value of some
function f [x], taking into account the distribution of x

Definition in two dimensions:

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 22
Expectation: Common Cases

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 23
Expectation: Rules

Rule 1:

Expected value of a constant is the constant

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 24
Expectation: Rules

Rule 2:

Expected value of constant times function is constant times


expected value of function

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 25
Expectation: Rules

Rule 3:

Expectation of sum of functions is sum of expectation of


functions

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 26
Expectation: Rules

Rule 4:

Expectation of product of functions in variables x and y


is product of expectations of functions if x and y are independent

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27
Conclusions
• Rules of probability are compact and simple

• Concepts of marginalization, joint and conditional


probability, Bayes rule and expectation underpin all of the
models we use

• One remaining concept – conditional expectation –


discussed later

Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 28

You might also like