Logistic Regression
Logistic Regression
Regression
Classica&on
Machine
Learning
Classica(on
Email:
Spam
/
Not
Spam?
Online
Transac&ons:
Fraudulent
(Yes
/
No)?
Tumor:
Malignant
/
Benign
?
0:
Nega&ve
Class
(e.g.,
benign
tumor)
1:
Posi&ve
Class
(e.g.,
malignant
tumor)
Andrew Ng
(Yes) 1
Malignant
?
(No)
0
Tumor Size
Tumor Size
Andrew Ng
Classica&on:
y
=
0
or
1
can
be
>
1
or
<
0
Logis&c
Regression:
Andrew Ng
Logis&c
Regression
Hypothesis
Representa&on
Machine
Learning
1
0.5
Sigmoid
func&on
Logis&c
func&on
Andrew Ng
Andrew Ng
Logis&c
Regression
Decision
boundary
Machine
Learning
Logis(c regression
z
Suppose
predict
if
predict if
Andrew Ng
Decision
Boundary
x2
3
2
1
1
x1
Predict if
Andrew Ng
-1
x1
Predict if
-1
x2
x1
Andrew Ng
Logis&c
Regression
Cost
func&on
Machine
Learning
Training
set:
m
examples
Andrew Ng
Cost
func(on
Linear
regression:
non-convex
convex
Andrew Ng
If y = 1
Andrew Ng
If y = 0
Andrew Ng
Logis&c
Regression
Simplied
cost
func&on
and
gradient
descent
Machine
Learning
Andrew Ng
To
t
parameters
:
To
make
a
predic&on
given
new
:
Output
Andrew Ng
Gradient Descent
Want
:
Repeat
Andrew Ng
Gradient Descent
Want
:
Repeat
Andrew Ng
Logis&c
Regression
Advanced
op&miza&on
Machine
Learning
Op(miza(on
algorithm
Cost
func&on
.
Want
.
Given
,
we
have
code
that
can
compute
-
-
(for
)
Gradient
descent:
Repeat
Andrew Ng
Op(miza(on
algorithm
Given
,
we
have
code
that
can
compute
-
-
(for
)
Op&miza&on
algorithms:
- Gradient
descent
- Conjugate
gradient
- BFGS
- L-BFGS
Advantages:
- No
need
to
manually
pick
- Oeen
faster
than
gradient
descent.
Disadvantages:
- More
complex
Andrew Ng
Example:
Andrew Ng
theta =
];
];
];
];
Andrew Ng
Logis&c
Regression
Mul&-class
classica&on:
One-vs-all
Machine
Learning
Mul(class
classica(on
Email
foldering/tagging:
Work,
Friends,
Family,
Hobby
Andrew Ng
Binary classica&on:
x2
Mul&-class classica&on:
x2
x1
x1
Andrew Ng
One-vs-all (one-vs-rest):
x2
x2
x1
Class
1:
Class
2:
Class
3:
x2
x1
x1
x2
x1
Andrew Ng
One-vs-all
Train
a
logis&c
regression
classier
for
each
class
to
predict
the
probability
that
.
On
a
new
input
,
to
make
a
predic&on,
pick
the
class
that
maximizes
Andrew Ng