Final Examination (Description and Topic List Subject to Minor Changes)
CSE 415: Introduction to Artificial Intelligence
The University of Washington, Seattle, Winter 2018
Date: Tuesday, March 13 (2:30-4:30PM)
Format: The format of the final exam will be similar to that of the midterm exam. However, the exam will be longer. The topics covered will be drawn from the following, which includes some topics from the first part of the course and some from the second.
Topics:

State-space search
  States, state spaces, operators, preconditions, moves,
  Heuristic evaluation functions,
  Iterative depth-first search, recursive depth-first search,
  Breadth-first search, best-first search, uniform-cost search,
  Iterative deepening, A* search.
  Admissible heuristics, Consistent heuristics
  Genetic search
    Application to the traveling Salesman Problem

Problem formulation
  States, operators, goal criteria
  Rittel and Webber's 10 characteristics of wicked problems

Minimax search for 2-player, zero-sum games
  Static evaluation functions
  Backed up values
  Alpha-beta pruning
  Zobrist hashing
  Expectimax search

Probabilistic reasoning
  Conditional probability
  Priors, likelihoods, and posteriors
  Bayes' rule
  Naive Bayes modeling
  The joint probability distribution
  Marginal probabilities
  Independence of random variables

Markov Decision Processes
  States, actions, transition model, reward function
  Values, Q-states, and Q-values
  Bellman updates
  Policies, policy extraction

Reinforcement Learning
  Model-based vs model-free learning
  Policy evaluation
  Temporal difference learning
  Q-learning
  Epsilon-greedy learning
  Exploration functions for Q-learning
  Application to the Towers-of-Hanoi puzzle and Grid World

Classification using Naive Bayes classifiers
  Classification using Naive Bayes
  Division by P(E) not necessary for classification
  Laplace smoothing: Adding 1 to counts when estimating P(Ei | Cj): why and how

Perceptrons
  How to compute AND, OR, and NOT.
  Simple pattern recognition (e.g., 5 x 5 binary image
    inputs for optical character recognition)
  Training sets, training sequences, and the perceptron 
    training algorithm.
  Linear separability and the perceptron training theorem.
  The meaning of deep convolutional networks

The Future of AI
  Asimov's three laws of robotics, Kurzweil's "singularity"