Help support MIT OpenCourseWare by shopping at Amazon.com! MIT OpenCourseWare offers direct links to Amazon.com to purchase the books cited in this course. Click on the Amazon logo to the left of any citation and purchase the book from Amazon.com, and MIT OpenCourseWare will receive up to 10% of all purchases you make. Your support will enable MIT to continue offering open access to MIT courses. |
Topic coverage will be adapted according to students' interests. Some or all of the following will be covered:
Markov Decision Processes and Dynamic Programming (2-3 weeks)
Simulation-Based Methods (2 weeks)
Value Function Approximation (4 weeks)
Policy Search Methods (2-3 weeks)
Online Learning and Games (2 weeks)
We will see applications throughout the course, including dynamic resource allocation, finance and queuing networks, among others.
Textbooks
Bertsekas, Dimitri P. Dynamic Programming and Optimal Control. 2 vols. Belmont, MA: Athena Scientific, 2007. ISBN: 9781886529083.
Bertsekas, Dimitri P., and John N. Tsitsiklis. Neuro-Dynamic Programming. Belmont, MA: Athena Scientific, 1996. ISBN: 9781886529106.
Individual Papers are also used for many class sessions, as listed in the readings section.
Grading Policy
|
Term Project
Students will be offered the option of working on theory, algorithms and/or applications. Project proposals will be submitted midway through the term, with the final project due at the end of the term.
A 10-15 page project report and 15-20 minute presentation are required.