0% found this document useful (0 votes)
21 views5 pages

sp2014 Midterm

This document is a midterm exam for UW CSE 473 from Fall 2014, consisting of various questions related to artificial intelligence concepts such as agents, search algorithms, constraint satisfaction problems (CSP), and Markov decision processes (MDP). The exam includes multiple choice, true/false questions, and problem-solving tasks requiring students to demonstrate their understanding of the material. Students are instructed to provide clear and concise answers while adhering to specific grading criteria.

Uploaded by

shahram taheri
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views5 pages

sp2014 Midterm

This document is a midterm exam for UW CSE 473 from Fall 2014, consisting of various questions related to artificial intelligence concepts such as agents, search algorithms, constraint satisfaction problems (CSP), and Markov decision processes (MDP). The exam includes multiple choice, true/false questions, and problem-solving tasks requiring students to demonstrate their understanding of the material. Students are instructed to provide clear and concise answers while adhering to specific grading criteria.

Uploaded by

shahram taheri
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

Name: _________________________________________ UW CSE 473 Midterm, Fall 2014

Instructions
Please answer clearly and succinctly. If an explanation is requested, think carefully before
writing. Points may be removed for rambling answers. If a question is unclear or ambiguous,
feel free to make the additional assumptions necessary to produce the answer. State these
assumptions clearly; you will be graded on the basis of the assumption as well as subsequent
reasoning. On multiple choice questions, incorrect answers will incur negative points
proportional to the number of choices. For example a 1 point true-false question will receive 1
point if correct, -1 if incorrect, and zero if left blank. Only make informed guesses.

There are XXXX problems worth KKK points on PPP pages.

1. (1 point) Who are you? Write your name at the top of every page.

2. (1 point each – total of 4). Types of Agents You are developing an


agent that solves crossword puzzle (like the one pictured to the right)
using an exhaustive dictionary of possible words. States are
partially-completed puzzles and actions place a word on the puzzle. On
each line below, we’ve listed two possible environmental aspects; circle
the one which better describes the crossword-puzzle environment.

a) fully observable vs. partially observable

b) single agent vs. multiagent,

c) stochastic vs. deterministic,

d) discrete vs. continuous

3. (1 point each – total of 13) True / False Circle the correct answer.
 
(a) T F Iterative deepening search is guaranteed to expand more nodes than breadth-
first search (on any graph whose root is not the goal).

(b) T F A* search with a heuristic that is not completely admissible may still find the
shortest path to the goal state.

(c) T F Consider a finite, acyclic search space where depth-first search is guaranteed
to eventually find a solution and the root is not a goal. In this situation
iterative deepening search will always explore more nodes than depth-first.

1
Name: _________________________________________ UW CSE 473 Midterm, Fall 2014

(d) T F A pattern database helps an agent avoid wasting time in cycles by storing
previously-expanded states.

(e) T F Random restarts are often used in local search to diminish the problem of
local maxima.

(f) T F Doubling your computer's speed allows you to double the depth of a tree
search given the same amount of time

(g) T F Every CSP with higher order constraints can be rewritten as a binary CSP with
the same number of variables.

(h) T F If a binary CSP has a tree-structured constraint graph, we can find a satisfying
assignment (or prove no satisfying assignment exists) in time that is linear in
the number of variables.

(i) T F Backtracking search on CSPs, while generally much faster than general
purpose search algorithms like A*, still requires exponential time in the worst
case.

(j) T F One reason to use forward checking in a CSP problem is in order to detect
failures quickly and backtrack earlier.

(k) T F An agent that uses Minimax search, which assumes an enemy behaves
optimally, may well achieve a better score when playing against a suboptimal
enemy than the agent would against an optimal enemy.

(l) T F All other things being equal, value iteration will converge in fewer iterations,
when the discount factor, gamma, is smaller.

(m) T F Expectimax search can be used to solve an MDP in a finite horizon setting.

(n) T F The optimal policy for an MDP depends on the MDP’s start state.

2
Name: _________________________________________ UW CSE 473 Midterm, Fall 2014

4. (2 points each – total of 8) Search.


Given the graph to the right, write down
the order in which the states are visited
by the following search algorithms. If a
state is visited more than once, write it
each time. Ties (e.g., which child to first
explore in depth-first search) should be
resolved according to alphabetic order
(i.e. prefer A before Z). Remember to
include the start and goal states in
your answer. Treat the goal state as G
when you break ties. Assume that
algorithms execute the goal check when
nodes are visited, not when their parent
is expanded to create them as children.

(a) Iterative deepening depth first search

____ ____ ____ ____ ____ ____ ____ ____ ____ ____ ____ ____

(b) A* search, where f(n)=g(n)+h(n)

____ ____ ____ ____ ____ ____ ____ ____ ____ ____ ____ ____

3
Name: _________________________________________ UW CSE 473 Midterm, Fall 2014

5. (10 points) Constraint Satisfaction


Three robots (A, B, C) have two hours to complete five tasks (1,2,3,4,5). Each task takes one
hour to complete, each robot can work on only one task at a time,
Robot Tasks
and only one robot may work on a task at a time. Each robot is
A 1,2,3
only equipped to perform certain tasks, as shown in the table.
B 1,2,5
Finally, task 1 must be completed before task 2, and 3 must be
C 2,4,5
completed before 5. We can formulate this problem as a CSP,
using one variable for each task: X1,…, X5, whose possible values are a subset of A1, A2, …
C2, where X5=C2 means that task 5 is done by robot C and Time(X5) is 2. We have written the
domain of each variable in the diagram below.

a) (2 points) write the constraints (either binary or


n-ary forms are ok).

b) (1 point) Complete the drawing above to show the constraint graph.

d) (4 points) Is the initial state arc-consistent?


If not, cross out the values for each variable that would be pruned by running AC-3 .

e) (4 points) Solve the (reduced, arc-consistent) CSP using backtracking search (without
forward checking). Use the minimum remaining values (MRV) variable ordering (breaking
ties in numerical order), and least constraining value (LCV) value ordering (breaking ties in
alpha-numerical). What order are the first variables assigned, and what values are they
given?
The first variable assigned is _______ it’s given value _________

The second variable assigned is _______ it’s given value _________

The third variable assigned is _______ it’s given value _________

4
Name: _________________________________________ UW CSE 473 Midterm, Fall 2014

6 (9 points) MDPs. Consider a setting where every 6 months Apple decides whether to
release a new version of the iPhone or not. Assume the problem can be represented as an
MDP with states (G=Good, M = Mediocre, B = Bad),
To G To M To B
each referring to the public sentiment towards Apple.
From G Take R 0.1 0.9 0.0
The actions are R=Release, D = Don’t release.
From G Take D 0.2 0.8 0.0
Taking an action that lands in state G (from any
From M Take R 0.1 0.9 0.0
other state, including itself) receives reward 2,
From M Take D 0.0 0.3 0.7
landing in state M receives reward 0, and landing in
From B Take R 0.9 0.0 0.1
state B receives reward -1. The discount factor (γ)
From B Take D 0.0 0.5 0.5
is 1. The transitions are as shown in the table:

a) For this MDP, fill in the blank spaces in the value iteration table (1 point for each Q entry)

G M B

V0(state) 0 0 0

Q1(state, R)

Q1(state, D)

V1(state)

Q2(state, R) Not required Not required

Q2(state, D) Not required Not required

V2(state) Not required Not required

b) (1 point) What should Apple do if it is in state M with a horizon of 2?

You might also like