Human Level Concept Learning Through Probabilistic Program Induction
Human Level Concept Learning Through Probabilistic Program Induction
Introduction
People when presented with new things, learn to generalize it so quickly, for eg: If we
have learnt to drive in some car, you could generalise to drive other cars too, it need
be the same as the one we learnt in, but this is not the same with machines. Machine
learning algorithms would require potentially hundreds or even thousands of images to
learn from before they could achieve human like accuracy.
This paper presents computational model that captures human learning capabilities for a
large class of simple visual concepts, such as character’s of the world’s alphabets. It
then represents concepts as simple programs that best explain the given examples
under the Bayesian Criterion. On a challenging one-shot learning classification tasks,
it achieves human level performance. This model has even passed the Turing tests in
some cases.
The generative process for types P(y) and tokens P(q(m)|y) are described by the
pseudocode in Fig. 3B and detailed along with the image model P(I(m)|q(m)) in section
S2.The model learns to learn by fitting each conditional distribution to a background set
of characters from 30 alphabets, using both the image and the stroke data, and this
image set was also used to pretrain the alternative deep learning models.
3. Results
● One shot learning for alphabets(experimental setup), humans error rate was at 4.5% and
BPL was around 3.3%, where a Deep CNN got around 13.5% error rate.
● BPL benefits by modeling the underlying causal process in learning concepts.
● Human capacity for one-shot learning is much more than classification, can lead to
sparse generations as well.
● A siamese DNN customised for this one shot learning was compared
● Even with less experience (priori) these models were able to beat DNN’s.
● Generating totally foreign concepts that are meaningful, is termed creativity. So, some
foreign examples were trained and yet again BPL gave good results.
4. Discussion
● Principles of composition, causality and learning to learn machines will help in creating
better one-shot learning algorithms.
● BPL still sees less info than humans do, like symmetry, parallel lines e.t.c
● This can be extended beyond visual perception.
● Audio recognition is a good application for this algorithm.
● Comparing how children parse and generalize behaviour at different stages of learning,
will help in comparing different BPL (adult learning) algorithms.