Chapter 1
Chapter 1
Introduction
What is AI?
Example:
My goals – (i) do not get wet if rain; (ii) do not looked stupid
(such as bring an umbrella when not raining)
My knowledge/belief – weather forecast for rain and I believe
it
My rational behaviour – bring an umbrella
The outcome of my behaviour: If rain, then my rational
behaviour achieves both goals; If no rain, then my rational
behaviour fails to achieve the 2nd goal
The successfulness of “behave rationally” is limited by my
knowledge and belief.
The State of The Art
Intelligent Agents
Rationality
Rational at any given time depends on four things:
• The performance measure that defines the criterion of success.
• The agent's prior knowledge of the environment.
• The actions that the agent can perform.
• The agent's percept sequence to date.
What is the driving environment that the taxi will face? Any taxi
driver must deal with a variety of roads, ranging from rural lanes
and urban alleys to 12-lane freeways. The roads contain other
traffic, pedestrians, stray animals, road works, police cars, puddles,
and potholes. The taxi must also interact with potential and actual
passengers.
3. Properties of Task Environments:
The state of the environment, the way time is handled, and agents
percepts & actions can be discrete or continuous
Ex: Crossword puzzles: discrete state, time, percepts &
actions
Ex: Taxi driving: continuous state, time, percepts & actions
Note:
The simplest environment is fully observable, single-agent,
deterministic, episodic, static and discrete. Ex: simple
vacuum cleaner
The table represents explicitly the agent function Ex: the simple
vacuum cleaner
For the braking problem, the internal state is not too extensive—
just the previous frame from the camera, allowing the agent to
detect when two red lights at the edge of the vehicle go on or off
simultaneously.
For other driving tasks such as changing lanes, the agent needs
to keep track of where the other cars are if it can’t see them all at
once. And for any driving to be possible at all, the agent needs
to keep track of where its keys are.
Updating this internal state information as time goes by requires
two kinds of knowledge to be encoded in the agent program.
First, we need some information about how the world evolves
independently of the agent—for example, that an overtaking car
generally will be closer behind than it was a moment ago.
Second, we need some information about how the agent’s own
actions affect the world—for example, that when the agent turns
the steering wheel clockwise, the car turns to the right, or that
after driving for five minutes northbound on the freeway, one is
usually about five miles north of where one was five minutes
ago.
This knowledge about “how the world works”—whether
implemented in simple Boolean circuits or in complete scientific
theories—is called a model of the world. An agent that uses
such a model is called a model-based agent.
1. Goal-based agents
• These agents are similar to the goal-based agent but provide an extra
component of utility measurement which makes them different by
providing a measure of success at a given state.
• Utility-based agent act based not only goals but also the best way to
achieve the goal.
• The utility function maps each state to a real number to check how
efficiently each action achieves the goals.
5. Learning Agents
After the taxi makes a quick left turn across three lanes, the
critic observes the shocking language used by other drivers.
From this experience, the learning element formulates a rule
saying this was a bad action.
The performance element is modified by adding the new rule.
The problem generator might identify certain areas of behaviour
in need of improvement, and suggest trying out the brakes on
different road surfaces under different conditions.