Lesson 1 - Introduction to AI
Lesson 1 - Introduction to AI
ARTIFICIAL INTELLIGENCE
BUI NGOC DUNG
Information (if available)
Traditionally, computer scientists and engineers have been more interested in the engineering goal, while
psychologists, philosophers and cognitive scientists have been more interested in the scientific goal. It makes
good sense to be interested in both, as there are common techniques and the two approaches can feed off
each other. In this module we shall attempt to keep both goals in mind.
ROOTS OF AI
PHILOSOPHY
Time Event
~400 BC Socrates asks for an algorithm to distinguish piety from non-piety.
~350 BC Aristole formulated different styles of deductive reasoning, which could mechanically generate conclusions from
initial premises, e.g.
Modus Ponens
𝐼𝑓 𝐴 ⇒ 𝐵 𝑎𝑛𝑑 𝐴 𝑡ℎ𝑒𝑛 𝐵
𝐼𝑓 𝐴 𝑖𝑚𝑝𝑙𝑖𝑒𝑠 𝐵 𝑎𝑛𝑑 𝐴 𝑖𝑠 𝑡𝑟𝑢𝑒 𝑡ℎ𝑒𝑛 𝐵 𝑖𝑠 𝑡𝑟𝑢𝑒
1596 – 1650 Rene Descartes idea of mind-body dualism – part of the mind is exempt from physical laws. Otherwise, how do we
have free will?
1646 – 1716 Wilhelm Leibnitz was one of the first to take the materialist position which holds that the mind operates by ordinary
physical processes – this has the implication that mental processes can potentially be carried out by machines.
LOGIC / MATHEMATICS
Time Event
1777 Earl Stanhope’s Logic Demonstrator was a machine that was able to solve syllogisms, numerical problems in a
logical form, and elementary questions of probability.
1815 – 1864 George Boole introduced his formal language for making logical inference in 1847 – Boolean algebra.
1848 – 1925 Gottlob Frege produced a logic that is essentially the first-order logic that today forms the most basic knowledge
representation system.
1906 – 1978 Kurt Gödel showed in 1931 that there are limits to what logic can do. His Incompleteness Theorem showed that in any
formal logic powerful enough to describe the properties of natural numbers, there are true statements whose truth
cannot be established by any algorithm.
1995 Roger Penrose tries to prove the human mind has non-computable capabilities.
COMPUTATION
Time Event
1869 William Jevon’s Logic Machine could handle Boolean Algebra and Venn Diagrams, and was able to solve logical
problems faster than human beings.
1912 – 1954 Alan Turing tried to characterise exactly which functions are capable of being computed. Unfortunately, it is difficult
to give the notion of computation a formal definition. However, the Church-Turing thesis, which states that a Turing
machine is capable of computing any computable function, is generally accepted as providing a sufficient
definition. Turing also showed that there were some functions which no Turing machine can compute (e.g. Halting
Problem).
1903 – 1957 John von Neumann proposed the von Neuman architecture which allows a description of computation that is
independent of the particular realisation of the computer.
~1960s Two important concepts emerged: Intractability (when solution time grows at least exponentially) and Reduction
(to ‘easier’ problems).
PSYCHOLOGY / COGNITIVE SCIENCE
Modern Psychology / Cognitive Psychology / Cognitive Science is the science which studies how the mind operates,
how we behave, and how our brains process information.
Language is an important part of human intelligence. Much of the early work on knowledge representation was tied to
language and informed by research into linguistics.
It is natural for us to try to use our understanding of how human (and other animal) brains lead to intelligent behavior
in our quest to build artificial intelligent systems. Conversely, it makes sense to explore the properties of artificial
systems (computer models/simulations) to test our hypotheses concerning human systems.
Many sub-fields of AI are simultaneously building models of how the human system operates, and artificial systems for
solving real world problems, and are allowing useful ideas to transfer between them.
BIOLOGY / NEUROSCIENCE
Our brains (which give rise to our intelligence) are made up of tens of billions of neurons, each connected to hundreds
or thousands of other neurons. Each neuron is a simple processing device (e.g. just firing or not firing depending on the
total amount of activity feeding into it). However, large networks of neurons are extremely powerful computational
devices that can learn how best to operate.
The field of Connectionism or Neural Networks attempts to build artificial systems based on simplified networks of
simplified artificial neurons. The aim is to build powerful AI systems, as well as models of various human abilities.
Neural networks work at a sub-symbolic level, whereas much of conscious human reasoning appears to operates at a
symbolic level.
Artificial neural networks perform well at many simple tasks, and provide good models of many human abilities.
However, there are many tasks that they are not so good at, and other approaches seem more promising in those
areas.
EVOLUTION
One advantage humans have over current machines/computers is that they have a long evolutionary history.
Charles Darwin (1809 – 1882) is famous for his work on evolution by natural selection. The idea is that fitter individuals
will naturally tend to live longer and produce more children, and hence after many generations a population will
automatically emerge with good innate properties.
This has resulted in brains that have much structure, or even knowledge, built in at birth. This gives them at the
advantage over simple artificial neural network systems that have to learn everything. Computers are finally becoming
powerful enough that we can simulate evolution and evolve good AI systems. We can now even evolve systems (e.g.
neural networks) so that they are good at learning.
A related field called genetic programming has had some success in evolving programs, rather than programming
them by hand.
TECHNIQUES OF AI
Representation Learning
Search Rules
TECHNIQUES OF AI
Representation
All AI systems have an important feature of knowledge representation. The rule-based systems, frame-based systems, and
semantic networks make use of a sequence of if-then rules, while the artificial neural networks make use of connections along
with connection weights.
Learning
All AI systems have capability of learning, using which they automatically build up the knowledge from the environment, e.g.,
acquiring the rules for a rule-based expert system, or determining the appropriate connection weights in an artificial neural
network.
Rules
The rules of an AI-based system can be implicit or explicit. When explicit, the rules are created by a knowledge engineer, say, for
an expert system, and when implicit, they can be, for example, in the form of connection weights in a neural network.
Search
The search can be in many forms, for example, searching the sequence of states that lead to solution faster, or searching for an
optimum set of connection weights in an ANN by minimizing the fitness function.
SUB-FIELDS OF AI
LEARNING SYSTEM
LEARNING SYSTEM
with respect to some task (T). Thus, the performance measure (P)
Tom Mitchell
SPAM E-MAIL DETECTION
Big Data
Open-Source Software
Cloud Computing
Interdisciplinary Collaboration
Increased Investment
ADVANCEMENTS IN COMPUTING POWER