Sutton Barto Book: Reinforcement Learning: An Introduction Reinforcement Reinforcement Learning: Learning An Introduction Richard s. Sutton and andrew G, Barto MIT Press, Cambridge, MA, 1998 A Bradford book Endorsements Code Solutions Figures Errata Course Slides Hard s Sutton and Anew G, Harto This introductory textbook on reinforcement learning is targeted toward engineers and scientists in artificial intelligence, operations research, neural networks, and control systems, and we hope it will also be of interest to psychologists and neuroscientists If you would like to order a copy of the book, or if you are qualified instructor and would like to see an examination copy, please see the MIT Press home page for this book. Or you might be interested in the reviews at amazon. com. There is also a Japanese translation available The table of contents of the book is given below, with associated HTML. The hTML version has a number of presentation problems, and its text is slightly different from the real book, but it may be useful for some purposes Preface Part i: The problem 1 Introduction o 1.1 Reinforcement learning 1.2 Examples o 1 3 Elements of reinforcement learning o 1. 4 An Extended Example: Tic-Tac-Toe file: //Ibook/thebook. html (1 of 5)[28 08/382 03: 12: 45 3A
Sutton & Barto Book: Reinforcement Learning: An Introduction Reinforcement Learning: An Introduction Richard S. Sutton and Andrew G. Barto MIT Press, Cambridge, MA, 1998 A Bradford Book Endorsements Code Solutions Figures Errata Course Slides This introductory textbook on reinforcement learning is targeted toward engineers and scientists in artificial intelligence, operations research, neural networks, and control systems, and we hope it will also be of interest to psychologists and neuroscientists. If you would like to order a copy of the book, or if you are qualified instructor and would like to see an examination copy, please see the MIT Press home page for this book. Or you might be interested in the reviews at amazon.com. There is also a Japanese translation available. The table of contents of the book is given below, with associated HTML. The HTML version has a number of presentation problems, and its text is slightly different from the real book, but it may be useful for some purposes. ● Preface Part I: The Problem ● 1 Introduction ❍ 1.1 Reinforcement Learning ❍ 1.2 Examples ❍ 1.3 Elements of Reinforcement Learning ❍ 1.4 An Extended Example: Tic-Tac-Toe file:///C|/book/the-book.html (1 of 5) [28/08/1382 03:12:45 ユネヘ]
Sutton Barto Book: Reinforcement Learning: An Introduction o 1.5 Summary o 1.6 History of Reinforcement Learning o 1.7 Bibliographical Remarks e2 Evaluative Feedback 2.1 An n-armed Bandit problem o 2. 2 Action-Value methods o 2.3 Softmax Action Selection o 2.4 Evaluation versus Instruction o 2. 5 Incremental Implementation o 2. 6 Tracking a Nonstationary Problem o 2. 7 Optimistic Initial values o 2. 8 Reinforcement Comparison o 2.9 Pursuit Methods 2.10 Associative searo o 2.11 Conclusion o 2. 12 Bibliographical and Historical Remarks 3 The Reinforcement Learning Problem o 3. 1 The Agent-Environment Interface o 3.2 Goals and Rewards o 3. 3 Returns o 3. 4 A Unified Notation for Episodic and Continual Tasks o 3. 5 The Markov Property o 3.6 Markov Decision Processes o 3.7 Value Functions o 3.8 Optimal Value Functions o 3.9 Optimality and Approximation 3.10 Summary o 3 11 Bibliographical and Historical Remarks Part Il: Elementary Methods 4 Dynamic Programming 4. 1 Policy Evaluation o 4.2 Policy Improvement o 4.3 Policy Iteration o 44 Value iteration e:∥/c|/ ookthe-book. htm(2of5)[2808∧38203:12:45卫
Sutton & Barto Book: Reinforcement Learning: An Introduction ❍ 1.5 Summary ❍ 1.6 History of Reinforcement Learning ❍ 1.7 Bibliographical Remarks ● 2 Evaluative Feedback ❍ 2.1 An n-armed Bandit Problem ❍ 2.2 Action-Value Methods ❍ 2.3 Softmax Action Selection ❍ 2.4 Evaluation versus Instruction ❍ 2.5 Incremental Implementation ❍ 2.6 Tracking a Nonstationary Problem ❍ 2.7 Optimistic Initial Values ❍ 2.8 Reinforcement Comparison ❍ 2.9 Pursuit Methods ❍ 2.10 Associative Search ❍ 2.11 Conclusion ❍ 2.12 Bibliographical and Historical Remarks ● 3 The Reinforcement Learning Problem ❍ 3.1 The Agent-Environment Interface ❍ 3.2 Goals and Rewards ❍ 3.3 Returns ❍ 3.4 A Unified Notation for Episodic and Continual Tasks ❍ 3.5 The Markov Property ❍ 3.6 Markov Decision Processes ❍ 3.7 Value Functions ❍ 3.8 Optimal Value Functions ❍ 3.9 Optimality and Approximation ❍ 3.10 Summary ❍ 3.11 Bibliographical and Historical Remarks Part II: Elementary Methods ● 4 Dynamic Programming ❍ 4.1 Policy Evaluation ❍ 4.2 Policy Improvement ❍ 4.3 Policy Iteration ❍ 4.4 Value Iteration file:///C|/book/the-book.html (2 of 5) [28/08/1382 03:12:45 ユネヘ]
Sutton Barto Book: Reinforcement Learning: An Introduction o 4.5 Asynchronous Dynamic Programming o 4.6 Generalized Policy Iteration o 4.7 Efficiency of Dynamic Programming o 4.8 Summar o 4.9 Historical and Bibliographical Remarks 5 Monte Carlo Methods o 5. 1 Monte Carlo Policy Evaluation o 5.2 Monte Carlo Estimation of Action values o 5.3 Monte Carlo Control o 5.4 On-Policy Monte Carlo Control o 5.5 Evaluating One Policy While Following Another o 5.6 Off-Policy Monte Carlo Control o 5.7 Incremental Implementation o 5.8 Summary o 5.9 Historical and Bibliographical Remarks 6 Temporal Difference Learning o.1 TD Prediction o 6.2 Advantages of TD Prediction Methods o 6.3 Optimality of TD(O) o 6.4 Sarsa: On-Policy TD Control 6.5 Q-learning: Off-Policy TD Control o 6.6 Actor-Critic Methods (s o 6.7 R-Learning for Undiscounted Continual Tasks() o 6.8 Games, After States, and other Special Cases o 6.conclusions o 6.10 Historical and Bibliographical Remarks Part llle: a unified view ·7 Eligibility Traces 7. 1 n-step TD Prediction o 7.2 The Forward View of TDO. o 7.3 The Backward view of TDO o 7. 4 Equivalence of the Forward and Backward views o 7.5 Sarsa o76Q0 e:∥/c|/ ookthe-book. htm(3of5)[2808∧38203:12:45卫
Sutton & Barto Book: Reinforcement Learning: An Introduction ❍ 4.5 Asynchronous Dynamic Programming ❍ 4.6 Generalized Policy Iteration ❍ 4.7 Efficiency of Dynamic Programming ❍ 4.8 Summary ❍ 4.9 Historical and Bibliographical Remarks ● 5 Monte Carlo Methods ❍ 5.1 Monte Carlo Policy Evaluation ❍ 5.2 Monte Carlo Estimation of Action Values ❍ 5.3 Monte Carlo Control ❍ 5.4 On-Policy Monte Carlo Control ❍ 5.5 Evaluating One Policy While Following Another ❍ 5.6 Off-Policy Monte Carlo Control ❍ 5.7 Incremental Implementation ❍ 5.8 Summary ❍ 5.9 Historical and Bibliographical Remarks ● 6 Temporal Difference Learning ❍ 6.1 TD Prediction ❍ 6.2 Advantages of TD Prediction Methods ❍ 6.3 Optimality of TD(0) ❍ 6.4 Sarsa: On-Policy TD Control ❍ 6.5 Q-learning: Off-Policy TD Control ❍ 6.6 Actor-Critic Methods (*) ❍ 6.7 R-Learning for Undiscounted Continual Tasks (*) ❍ 6.8 Games, After States, and other Special Cases ❍ 6.9 Conclusions ❍ 6.10 Historical and Bibliographical Remarks Part III: A Unified View ● 7 Eligibility Traces ❍ 7.1 n-step TD Prediction ❍ 7.2 The Forward View of TD() ❍ 7.3 The Backward View of TD() ❍ 7.4 Equivalence of the Forward and Backward Views ❍ 7.5 Sarsa() ❍ 7.6 Q() file:///C|/book/the-book.html (3 of 5) [28/08/1382 03:12:45 ユネヘ]
Sutton Barto Book: Reinforcement Learning: An Introduction o 7. 7 Eligibility Traces for Actor-Critic Methods o 7. 8 Replacing Traces o 7.9 Implementation Issues o 7.10 Variable o 7 11 Conclusions o 7 12 Bibliographical and Historical Remarks 8 Generalization and Function Approximation o 8.1 Value Prediction with Function Approximation o 8.2 Gradient-Descent Methods 08.3 Linear Methods 8.3 1 Coarse Co 8.3.2 Tile Coding 83. 3 Radial Basis Functions ■834 Kanerva Coding o 8.4 Control with Function Approximation o 8.5 Off-Policy Bootstrapping o 8.6 Should We Bootstrap? o 8/summary o 8.8 Bibliographical and Historical Remarks 9 Planning and Learning o 9. 1 Models and Planning o 9.2 Integrating Planning, Acting, and Learning o 9.3 When the Model is Wrong o 9.4 Prioritized Sweeping o 9.5 Full vs Sample Backups o 9.6 Trajectory Sampling 9.7 Heuristic Search 9. 8 Summary o 9.9 Historical and Bibliographical Remarks 10 Dimensions o 10.1 The Unified View o 10.2 Other frontier dimensions 11 Case Studies o.I TD-Gammon file: //Ibook /thebook. html (4 of 5)[28 08/382 03: 12: 45 3A
Sutton & Barto Book: Reinforcement Learning: An Introduction ❍ 7.7 Eligibility Traces for Actor-Critic Methods (*) ❍ 7.8 Replacing Traces ❍ 7.9 Implementation Issues ❍ 7.10 Variable (*) ❍ 7.11 Conclusions ❍ 7.12 Bibliographical and Historical Remarks ● 8 Generalization and Function Approximation ❍ 8.1 Value Prediction with Function Approximation ❍ 8.2 Gradient-Descent Methods ❍ 8.3 Linear Methods ■ 8.3.1 Coarse Coding ■ 8.3.2 Tile Coding ■ 8.3.3 Radial Basis Functions ■ 8.3.4 Kanerva Coding ❍ 8.4 Control with Function Approximation ❍ 8.5 Off-Policy Bootstrapping ❍ 8.6 Should We Bootstrap? ❍ 8.7 Summary ❍ 8.8 Bibliographical and Historical Remarks ● 9 Planning and Learning ❍ 9.1 Models and Planning ❍ 9.2 Integrating Planning, Acting, and Learning ❍ 9.3 When the Model is Wrong ❍ 9.4 Prioritized Sweeping ❍ 9.5 Full vs. Sample Backups ❍ 9.6 Trajectory Sampling ❍ 9.7 Heuristic Search ❍ 9.8 Summary ❍ 9.9 Historical and Bibliographical Remarks ● 10 Dimensions ❍ 10.1 The Unified View ❍ 10.2 Other Frontier Dimensions ● 11 Case Studies ❍ 11.1 TD-Gammon file:///C|/book/the-book.html (4 of 5) [28/08/1382 03:12:45 ユネヘ]
Sutton Barto Book: Reinforcement Learning: An Introduction o 11.2 Samuel's Checkers Player o 11. 3 The Acrobot o 11.4 Elevator Dispatching o 11.5 Dynamic Channel Allocation o 11.6 Job-Shop Scheduling References Summary of Notation file://clbookthebook html (5 of 5)[28/08/382 03: 12: 45 1*]
Sutton & Barto Book: Reinforcement Learning: An Introduction ❍ 11.2 Samuel's Checkers Player ❍ 11.3 The Acrobot ❍ 11.4 Elevator Dispatching ❍ 11.5 Dynamic Channel Allocation ❍ 11.6 Job-Shop Scheduling ● References ● Summary of Notation file:///C|/book/the-book.html (5 of 5) [28/08/1382 03:12:45 ユネヘ]
Endorsements for Sutton Barto Book: Reinforcement Learning: An Introduction Endorsements for Reinforcement Learning An Introduction by richard s Sutton and andrew G. Barto This is a highly intuitive and accessible introduction to the recent major developments in reinforcement learning, written by two of the fields pioneering contributors Dimitri P. Bertsekas and John N. Tsitsiklis, Professors, Department of Electrical Enginneering and Computer Science, Massachusetts Institute of Technology This book not only provides an introduction to learning theory but also serves as a tremendous sourve of ideas for further development and applications in the real world Toshio Fukuda, Nagoya University, Japan; President, IEEE Robotics and Automation Society Reinforcement learning has al ways been important in the understanding of the driving forces behind biological systems, but in the past two decades it has become increasingly important, owing to the development of mathematical algorithms. Barto and Sutton were the prime movers in leading the development of these algorithms and have described them with wonderful clarity in this new text. I predict it will be the standard text Dana ballard, Professor of Computer Science, University of rochester The widely acclaimed work of Sutton and Barto on reinforcement learning applies some essentials of animal learning, in clever ways, to artificial learning systems. This is a very readable and comprehensive account of the background, algorithms, applications, and future directions of this pioneering and far reaching work Wolfram Schulte, University of fribourg Switzerland file: //Ibook/endorsements. html [28 08/382 03: 12: 57 17A
Endorsements for Sutton & Barto Book: Reinforcement Learning: An Introduction Endorsements for: Reinforcement Learning: An Introduction by Richard S. Sutton and Andrew G. Barto "This is a highly intuitive and accessible introduction to the recent major developments in reinforcement learning, written by two of the field's pioneering contributors" Dimitri P. Bertsekas and John N. Tsitsiklis, Professors, Department of Electrical Enginneering and Computer Science, Massachusetts Institute of Technology "This book not only provides an introduction to learning theory but also serves as a tremendous sourve of ideas for further development and applications in the real world" Toshio Fukuda, Nagoya University, Japan; President, IEEE Robotics and Automation Society "Reinforcement learning has always been important in the understanding of the driving forces behind biological systems, but in the past two decades it has become increasingly important, owing to the development of mathematical algorithms. Barto and Sutton were the prime movers in leading the development of these algorithms and have described them with wonderful clarity in this new text. I predict it will be the standard text." Dana Ballard, Professor of Computer Science, University of Rochester "The widely acclaimed work of Sutton and Barto on reinforcement learning applies some essentials of animal learning, in clever ways, to artificial learning systems. This is a very readable and comprehensive account of the background, algorithms, applications, and future directions of this pioneering and farreaching work." Wolfram Schultz, University of Fribourg, Switzerland file:///C|/book/endorsements.html [28/08/1382 03:12:57 ユネヘ]
Code for Sutton Barto Book: Reinforcement Learning: An Introductio Code for. Reinforcement Learning An Introduction by richard s Sutton and andrew G. Barto Below are links to a variety of software related to examples and exercises in the book, organized by chapters(some files appear in multiple places). See particularly the Mountain Car code. Most of the rest of the code is written in Common Lisp and requires utility routines available here. For the graphics, you will need the the packages for G and in some cases my graphing tool. Even if you can not run this code, it still may clarify some of the details of the experiments. However, there is no guarantee that the examples in the book were run using exactly the software given. This code also has not been extensively tested or documented and is being made available"as is". If you have corrections, extensions, additions or improvements of any kind, please send them to me at rich @richsutton com for inclusion here Chapter 1: Introduction o Tic-Tac-Toe Example (Lisp). In C Chapter 2: Evaluative Feedback o 10-armed Testbed Example, Figure 2. 1 ( Lisp) o Testbed with Softmax Action Selection, Exercise 2.2(Lisp) o Bandits A and B, Figure 2.3(Lisp) o Testbed with Constant Alpha, cf Exercise 2.7 ( Lisp) o Optimistic Initial Values Example, Figure 2. 4(Lisp) o Code Pertaining to Reinforcement Comparison: Filel, File2, File3 Lisp) o Pursuit Methods Example, Figure 2.6 ( Lisp) Chapter 3: The Reinforcement Learning Problem o Pole-Balancing Example, Figure 3. 2(C) o Gridworld Example 3.8. Code for Figures 3.5 and 3.8 Lisp) Chapter 4: Dynamic Programming Policy evaluation, Gridworld Example 4. 1 Figure 4.2 Lisp o Policy Iteration, Jacks Car Rental Example, Figure 4.4 Lisp o Value Iteration, Gambler's Problem Example, Figure 4.6 Lisp) Chapter 5: Monte Carlo Methods o Monte Carlo Policy Evaluation, Blackjack Example 5. 1, Figure 5.2(Lisp) o Monte Carlo es, Blackjack Example 5.3, Figure 5.5 ( Lisp) Chapter 6: Temporal-Difference Learning o TD Prediction in Random Walk. Example 6.2, Figures 6.5 and 6.6 Lisp) o TD Prediction in Random Walk with Batch Training, Example 6.3, Figure 6.8(Lisp) le: /Cook/ode/code. html (1 of 2)[2808/382 03: 12: 58 1]
Code for Sutton & Barto Book: Reinforcement Learning: An Introduction Code for: Reinforcement Learning: An Introduction by Richard S. Sutton and Andrew G. Barto Below are links to a variety of software related to examples and exercises in the book, organized by chapters (some files appear in multiple places). See particularly the Mountain Car code. Most of the rest of the code is written in Common Lisp and requires utility routines available here. For the graphics, you will need the the packages for G and in some cases my graphing tool. Even if you can not run this code, it still may clarify some of the details of the experiments. However, there is no guarantee that the examples in the book were run using exactly the software given. This code also has not been extensively tested or documented and is being made available "as is". If you have corrections, extensions, additions or improvements of any kind, please send them to me at rich@richsutton.com for inclusion here. ● Chapter 1: Introduction ❍ Tic-Tac-Toe Example (Lisp). In C. ● Chapter 2: Evaluative Feedback ❍ 10-armed Testbed Example, Figure 2.1 (Lisp) ❍ Testbed with Softmax Action Selection, Exercise 2.2 (Lisp) ❍ Bandits A and B, Figure 2.3 (Lisp) ❍ Testbed with Constant Alpha, cf. Exercise 2.7 (Lisp) ❍ Optimistic Initial Values Example, Figure 2.4 (Lisp) ❍ Code Pertaining to Reinforcement Comparison: File1, File2, File3 (Lisp) ❍ Pursuit Methods Example, Figure 2.6 (Lisp) ● Chapter 3: The Reinforcement Learning Problem ❍ Pole-Balancing Example, Figure 3.2 (C) ❍ Gridworld Example 3.8, Code for Figures 3.5 and 3.8 (Lisp) ● Chapter 4: Dynamic Programming ❍ Policy Evaluation, Gridworld Example 4.1, Figure 4.2 (Lisp) ❍ Policy Iteration, Jack's Car Rental Example, Figure 4.4 (Lisp) ❍ Value Iteration, Gambler's Problem Example, Figure 4.6 (Lisp) ● Chapter 5: Monte Carlo Methods ❍ Monte Carlo Policy Evaluation, Blackjack Example 5.1, Figure 5.2 (Lisp) ❍ Monte Carlo ES, Blackjack Example 5.3, Figure 5.5 (Lisp) ● Chapter 6: Temporal-Difference Learning ❍ TD Prediction in Random Walk, Example 6.2, Figures 6.5 and 6.6 (Lisp) ❍ TD Prediction in Random Walk with Batch Training, Example 6.3, Figure 6.8 (Lisp) file:///C|/book/code/code.html (1 of 2) [28/08/1382 03:12:58 ユネヘ]
Code for Sutton Barto Book: Reinforcement Learning: An Introduction o TD Prediction in Random Walk(MatLab by Jim Stone R-learning on Access-Control Queuing Task, Example 6.7, Figure 6. 17(Lisp),(C version) Chapter 7: Eligibility Traces o N-step Td on the Random Walk, Example 7. 1, Figure 7. 2: online and offline (lisp). In C o lambda-return Algorithm on the Random Walk, Example 7.2, Figure 7.6(Lisp) o Online TD(ambda) on the Random Walk, Example 7.3, Figure 7.9 (Lisp) Chapter 8: Generalization and Function Approximation o Coarseness of Coarse Coding, Example 8.1, Figure 8.4 (Lisp o Tile Coding. ak. a cmacs o Linear Sarsa(lambda)on the Mountain-Car, a la Example 8.2 o Bairds Counterexample, Example 8.3, Figures 8 12 and 8.13(Lisp) Chapter 9: Planning and Learnin o Trajectory Sampling Experiment, Figure 9.14(Lisp) Chapter 10: Dimensions of Reinforcement Learning Chapter 11: Case Studies o Acrobot(Lisp environment only) o Java Demo of RL Dynamic Channel Assignment For other Rl software see the reinforcement Learning Repository at Michigan State University and here le: /Cook/ode/code. html (2 of 2)[2808/382 03: 12: 58 1]
Code for Sutton & Barto Book: Reinforcement Learning: An Introduction ❍ TD Prediction in Random Walk (MatLab by Jim Stone) ❍ R-learning on Access-Control Queuing Task, Example 6.7, Figure 6.17 (Lisp), (C version) ● Chapter 7: Eligibility Traces ❍ N-step TD on the Random Walk, Example 7.1, Figure 7.2: online and offline (Lisp). In C. ❍ lambda-return Algorithm on the Random Walk, Example 7.2, Figure 7.6 (Lisp) ❍ Online TD(lambda) on the Random Walk, Example 7.3, Figure 7.9 (Lisp) ● Chapter 8: Generalization and Function Approximation ❍ Coarseness of Coarse Coding, Example 8.1, Figure 8.4 (Lisp) ❍ Tile Coding, a.k.a. CMACs ❍ Linear Sarsa(lambda) on the Mountain-Car, a la Example 8.2 ❍ Baird's Counterexample, Example 8.3, Figures 8.12 and 8.13 (Lisp) ● Chapter 9: Planning and Learning ❍ Trajectory Sampling Experiment, Figure 9.14 (Lisp) ● Chapter 10: Dimensions of Reinforcement Learning ● Chapter 11: Case Studies ❍ Acrobot (Lisp, environment only) ❍ Java Demo of RL Dynamic Channel Assignment For other RL software see the Reinforcement Learning Repository at Michigan State University and here. file:///C|/book/code/code.html (2 of 2) [28/08/1382 03:12:58 ユネヘ]
Solutions Manual for Sutton Barto Book: Reinforcement Learning: An Introduction Solutions Manual for Reinforcement Learning An Introduction by richard s Sutton and andrew G. Barto An instructor's manual containing answers to all the non-programming exercises is available to qualified teachers. Send or fax a letter under your university's letterhead to the Text manager at MIT Press Exactly who you should send to depends on your location. Obtain the address as if you were requesting an examination copy Readers using the book for self study can obtain answers on a chapter-by-chapter basis after working on the exercises themselves. Send email to rich @richsutton com with your efforts to answer the exercises for a chapter, and we will send back a postscript file with the answers for that chapter We are also collecting overheads, code, exams, and other material useful for teaching from the book. If you have anything that may be useful in this regard that you would like to share with others, please send it in and welll make it available e∥/c|/ ook/solutions. htm|[2808∧38203:12:59
Solutions Manual for Sutton & Barto Book: Reinforcement Learning: An Introduction Solutions Manual for: Reinforcement Learning: An Introduction by Richard S. Sutton and Andrew G. Barto An instructor's manual containing answers to all the non-programming exercises is available to qualified teachers. Send or fax a letter under your university's letterhead to the Text Manager at MIT Press. Exactly who you should send to depends on your location. Obtain the address as if you were requesting an examination copy. Readers using the book for self study can obtain answers on a chapter-by-chapter basis after working on the exercises themselves. Send email to rich@richsutton.com with your efforts to answer the exercises for a chapter, and we will send back a postscript file with the answers for that chapter. We are also collecting overheads, code, exams, and other material useful for teaching from the book. If you have anything that may be useful in this regard that you would like to share with others, please send it in and we'll make it available. file:///C|/book/solutions.html [28/08/1382 03:12:59 ユネヘ]
Figures for Sutton Barto Book: Reinforcement Learning: An Introduction Fiqures for: Reinforcement Learning An Introduction by richard s Sutton and andrew G. Barto Below are links to postscript files for the figures of the book Page 1 Tic-Tac-Toe Game e Figure 1.1 Tic-Tac-Toe tree Figure 2. 1 10-armed Testbed Results Figure 2.2 Easy and Difficult Regions e Figure 2. 3 Performance on Bandits a and B Figure 2. 4 Effect of Optimistic Initial Action values Figure 2.5 Performance of Reinforcement Comparison Method Figure 2.6 Performance of Pursuit Method Figure 3. 1 Agent-Environment Interaction Figure 3.2 Pole-balancing Example Page 62 Absorbing State Sequence Figure 3.3 Transition Graph for the Recycling robot Example Figure 3. 4 Prediction Backup diagrams Figure 3.5 Gridworld Example Figure 3.6 Golf Example Figure 3. 7"Max"Backup Diagrams Figure 3.8 Solution to Gridworld Example Page 62 4 x 4 Gridworld Example Figure 4.2 Convergence Example(4 x 4 Gridworld) Figure 4. 4 Policy sequence in Jack's Car Rental Example Figure 4.6 Solution to the Gambler's Problem Figure 4. 7 Generalized policy Iteration Page 106 Coconvergence of Policy and value Figure 5. 1 Blackjack Policy Evaluation Figure 5.3 Backup Diagram for Monte Carlo Prediction le: /Cook/figures figures. html( 1 of 4)[2808/138203: 12: 59 17A
Figures for Sutton & Barto Book: Reinforcement Learning: An Introduction Figures for: Reinforcement Learning: An Introduction by Richard S. Sutton and Andrew G. Barto Below are links to postscript files for the figures of the book. ● Page 1 Tic-Tac-Toe Game ● Figure 1.1 Tic-Tac-Toe Tree ● Figure 2.1 10-armed Testbed Results ● Figure 2.2 Easy and Difficult Regions ● Figure 2.3 Performance on Bandits A and B ● Figure 2.4 Effect of Optimistic Initial Action Values ● Figure 2.5 Performance of Reinforcement Comparison Method ● Figure 2.6 Performance of Pursuit Method ● Figure 3.1 Agent-Environment Interaction ● Figure 3.2 Pole-balancing Example ● Page 62 Absorbing State Sequence ● Figure 3.3 Transition Graph for the Recycling Robot Example ● Figure 3.4 Prediction Backup Diagrams ● Figure 3.5 Gridworld Example ● Figure 3.6 Golf Example ● Figure 3.7 "Max" Backup Diagrams ● Figure 3.8 Solution to Gridworld Example ● Page 62 4 x 4 Gridworld Example ● Figure 4.2 Convergence Example (4 x 4 Gridworld) ● Figure 4.4 Policy sequence in Jack's Car Rental Example ● Figure 4.6 Solution to the Gambler's Problem ● Figure 4.7 Generalized Policy Iteration ● Page 106 Coconvergence of Policy and Value ● Figure 5.1 Blackjack Policy Evaluation ● Figure 5.3 Backup Diagram for Monte Carlo Prediction file:///C|/book/figures/figures.html (1 of 4) [28/08/1382 03:12:59 ユネヘ]