Overview and examples of sequential decision problems.
PART I - Stochastic models
1. Review of essential probabilistic tools. Markov chains, Martingales, basic concentration inequalities.
2. Discrete time Markov Decision Processes (MDPs).
2a. Finite time-horizon. Principle of optimality, backward induction.
2b. Infinite time-horizon. Principle of optimality, value / policy iteration, modified policy iteration, linear programming.
3. Solving MDPs - part 1. Exact solutions based on structural properties of the MDP.
4. Solving MDPs - part 2. Some approximation methods.
5. Extensions. Constrained MDPs, Partially Observable MDPs, Decentralized MDPs.
6. Limit theorems. Going from MDPs to deterministic continuous-time control and back.
7. Optimal stopping time problems.
8. Kalman filter.
9. Prediction with expert advice and Multi-Armed Bandit (MAB) problems.
PART II - Adversarial models and Games.
1. Prediction with expert advice and MAB problems in adversarial scenarios.
2. Sequential decision making in games. Internal regret, Correlated equilibria, Convergence to and selection of Nash Equilibria.
3. Recent advances in online optimization.