



Next: Business Studies Up: Michaelmas Term 2007: Part Previous: Michaelmas Term 2007: Part Contents
Artificial Intelligence II
Lecturer: Dr S.B. Holden
No. of lectures: 16
Prerequisite courses: Artificial Intelligence I, Logic and Proof, Algorithms I + II, Mathematical Methods for Computer Science, Discrete Mathematics I + II, Probability
Aims
The aim of this course is to build on Artificial Intelligence I, first by introducing more elaborate methods for knowledge representation and planning within the symbolic tradition, but then by moving beyond the purely symbolic view of AI and presenting methods developed for dealing with the critical concept of uncertainty. The central tool used to achieve the latter is probability theory. The course continues to exploit the primarily algorithmic and computer science-centric perspective that informed Artificial Intelligence I.
The course aims to provide further tools and algorithms required to produce AI systems able to exhibit limited human-like abilities, with an emphasis on the need to obtain richer forms of knowledge representation, better planning algorithms, and systems able to deal with the uncertainty inherent in the environments that most real agents might be expected to perform within.
Lectures
- Further symbolic knowledge representation. Representing
knowledge using First Order Logic (FOL). The situation calculus.
[1 lecture]
- Further planning. Incorporating heuristics into
partial-order planning. Planning graphs. The GRAPHPLAN algorithm.
Planning using propositional logic. [2 lectures]
- Uncertainty and Bayesian networks. Review of probability
as applied to AI. Bayesian networks. Inference in Bayesian networks
using both exact and approximate techniques. Other ways of dealing
with uncertainty. [3 lectures]
- Utility and decision-making. Maximizing expected utility,
decision networks, the value of information. [1 lecture]
- Further supervised learning. Bayes theorem as applied to
supervised learning. The maximum likelihood and maximum a posteriori
hypotheses. Applying the Bayesian approach to neural networks.
[4 lectures]
- Uncertain reasoning over time. Markov processes,
transition and sensor models. Inference in temporal models:
filtering, prediction, smoothing and finding the most likely
explanation. Hidden Markov models. [2 lectures]
- Reinforcement learning. Learning from rewards and punishments.
[2 lectures]
Objectives
At the end of this course students should
- have gained a deeper appreciation for the way in which computer
science has been applied to the problem of AI, and in particular for
more recent techniques concerning knowledge representation,
inference, planning and uncertainty
- know how to model situations using a variety of knowledge
representation techniques
- be able to design problem solving methods based on knowledge
representation, inference, planning, and learning techniques
- know how probability theory can be applied in practice as a
means of handling uncertainty in AI systems
Recommended reading
* Russell, S. & Norvig, P. (2003). Artificial intelligence: a modern approach. Prentice Hall (2nd ed.).
Bishop, S. (1995). Neural networks for pattern recognition. Oxford University Press.




Next: Business Studies Up: Michaelmas Term 2007: Part Previous: Michaelmas Term 2007: Part Contents