Introduction To Neural Networks and Machine Learning Lecture 1: What Are Neural Networks?
Introduction To Neural Networks and Machine Learning Lecture 1: What Are Neural Networks?
axon
body
dendritic
tree
Synapses
When a spike travels along an axon and arrives at a
synapse it causes vesicles of transmitter chemical to be
released
There are several kinds of transmitter
The transmitter molecules diffuse across the synaptic cleft
and bind to receptor molecules in the membrane of the postsynaptic neuron thus changing their shape.
This opens up holes that allow specific ions in or out.
The effectiveness of the synapse can be changed
vary the number of vesicles of transmitter
vary the number of receptor molecules.
Synapses are slow, but they have advantages over RAM
Very small
They adapt using locally available signals (but how?)
Idealized neurons
To model things we have to idealize them (e.g. atoms)
Idealization removes complicated details that are not
essential for understanding the main principles
Allows us to apply mathematics and to make
analogies to other, familiar systems.
Once we understand the basic principles, its easy to
add complexity to make the model more faithful
It is often worth understanding models that are known to
be wrong (but we mustnt forget that they are wrong!)
E.g. neurons that communicate real values rather
than discrete spikes of activity.
Linear neurons
These are simple but computationally limited
If we can make them learn we may get insight
into more complicated neurons
bias
i th input
y b xi wi
output
index over
input connections
y
0
weight on
ith input
b xi wi
i
z xi wi
i
1 if
0 otherwise
y
0
threshold
z j b j xi wij
i
yj
z j if z j 0
0 otherwise
y
0
threshold
Sigmoid neurons
These give a real-valued
output that is a smooth
and bounded function of
their total input.
Typically they use the
logistic function
They have nice
derivatives which
make learning easy
(see lecture 3).
If we treat y as a
probability of producing a
spike, we get stochastic
binary neurons.
z b xi wi
i
z
1 e
0.5
0
0
0 1 2 3 4 5 6 7 8 9
The image
The image
The image
The image
The image
The image
The image
Types of connectivity
Feedforward networks
These compute a series of
transformations
Typically, the first layer is the
input and the last layer is the
output.
Recurrent networks
These have directed cycles
in their connection graph.
They can have complicated
dynamics.
More biologically realistic.
output units
hidden units
input units
Reinforcement learning
Learn action to maximize payoff
Not much information in a payoff signal
Payoff is often delayed
Unsupervised learning
Create an internal representation of the input
e.g. form clusters; extract features
How do we know if a representation is good?