Neuro-Dynamic Programming

┬╖
┬╖ Athena Scientific
рел.реж
рез рд╕рдореАрдХреНрд╖рд╛
рдЗ-рдкреБрд╕реНрддрдХ
505
рдкреГрд╖реНрдард╣рд░реВ
рдпреЛрдЧреНрдп

рдпреЛ рдЗ-рдкреБрд╕реНрддрдХрдХрд╛ рдмрд╛рд░реЗрдорд╛

This is historically the first book that fully explained the neuro-dynamic programming/reinforcement learning methodology, a breakthrough in the practical application of neural networks and dynamic programming to complex problems of planning, optimal decision making, and intelligent control.

Neuro-dynamic programming uses neural network approximations to overcome the "curse of dimensionality" and the "curse of modeling" that have been the bottlenecks to the practical application of dynamic programming and stochastic control to complex problems. The methodology allows systems to learn about their behavior through simulation, and to improve their performance through iterative reinforcement.┬а

This book provides the first systematic presentation of the science and the art behind this exciting and far-reaching methodology. It develops a comprehensive analysis of reinforcement learning algorithms, and guides the reader to their successful application through case studies from complex problem areas. It contains material that is not available elsewhere in book form, such as a comprehensive and rigorous analysis of temporal difference methods, Q-learning, and error bounds associated with various methods.

рдореВрд▓реНрдпрд╛рдЩреНрдХрди рд░ рд╕рдореАрдХреНрд╖рд╛рд╣рд░реВ

рел.реж
рез рд╕рдореАрдХреНрд╖рд╛

рд▓реЗрдЦрдХрдХреЛ рдмрд╛рд░реЗрдорд╛

Dimitri P. Bertsekas undergraduate studies were in engineering at the National Technical University of Athens, Greece. He obtained his MS in electrical engineering at the George Washington University, Wash. DC in 1969, and his Ph.D. in system science in 1971 at the Massachusetts Institute of Technology.

Dr. Bertsekas has held faculty positions with the Engineering-Economic Systems Dept., Stanford University (1971-1974) and the Electrical Engineering Dept. of the University of Illinois, Urbana (1974-1979). From 1979 to 2019 he was with the Electrical Engineering and Computer Science Department of the Massachusetts Institute of Technology (M.I.T.), where he served as McAfee Professor of Engineering. In 2019, he was appointed Fulton Professor of Computational Decision Making, and a full time faculty member at the department of Computer, Information, and Decision Systems Engineering at Arizona State University (ASU), Tempe, while maintaining a research position at MIT. His research spans several fields, including optimization, control, large-scale computation, and data communication networks, and is closely tied to his teaching and book authoring activities. He has written numerous research papers, and eighteen books and research monographs, several of which are used as textbooks in MIT and ASU classes. Most recently Dr Bertsekas has been focusing on reinforcement learning, and authored a textbook in 2019, and a research monograph on its distributed and multiagent implementation aspects in 2020.

Professor Bertsekas was awarded the INFORMS 1997 Prize for Research Excellence in the Interface Between Operations Research and Computer Science for his book "Neuro-Dynamic Programming", the 2000 Greek National Award for Operations Research, the 2001 ACC John R. Ragazzini Education Award, the 2009 INFORMS Expository Writing Award, the 2014 ACC Richard E. Bellman Control Heritage Award for "contributions to the foundations of deterministic and stochastic optimization-based methods in systems and control," the 2014 Khachiyan Prize for Life-Time Accomplishments in Optimization, the SIAM/MOS 2015 George B. Dantzig Prize, and the 2022 IEEE Control Systems Award. In 2018, he was awarded, jointly with his coauthor John Tsitsiklis, the INFORMS John von Neumann Theory Prize for the contributions of the research monographs "Parallel and Distributed Computation" and "Neuro-Dynamic Programming". In 2001, he was elected to the United States National Academy of Engineering for "pioneering contributions to fundamental research, practice and education of optimization/control theory, and especially its application to data communication networks."

Dr. Bertsekas' recent books are "Introduction to Probability: 2nd Edition" (2008), "Convex Optimization Theory" (2009), "Dynamic Programming and Optimal Control," Vol. I, (2017), and Vol. II: (2012), "Abstract Dynamic Programming" (2018), "Convex OptimizationAlgorithms" (2015), "Reinforcement Learning and Optimal Control" (2019), and "Rollout, Policy Iteration, and Distributed Reinforcement Learning" (2020), all published by Athena Scientific.

John Tsitsiklis received the B.S. degree in Mathematics (1980), and the B.S. (1980), M.S. (1981), and Ph.D. (1984) degrees in Electrical Engineering, all from the Massachusetts Institute of Technology, Cambridge, Massachusetts, U.S.A.

Since 1984, he has been with the department of Electrical Engineering and Computer Science (EECS) at the Massachusetts Institute of Technology (MIT), where he is currently a Clarence J Lebel Professor of Electrical Engineering. At MIT, he has served as the director of the Laboratory for Information and Decision Systems (LIDS) and also as a co-director of the Operations Research Center (ORC).┬а

His research interests are in the fields of systems, optimization, control, and operations research. He is a coauthor of Parallel and Distributed Computation: Numerical Methods (D. Bertsekas), Neuro-Dynamic Programming (with D. Bertsekas), Introduction to Linear Optimization (with D. Bertsimas), and Introduction to Probability (with D. Bertsekas). He is also a coinventor in seven awarded U.S. patents.

He is a member of the National Academy of Engineering "for contributions to the theory and application of optimization in dynamic and distributed systems." He is a Fellow of the IEEE (1999) and of INFORMS (2007). His distinctions include the ACM Sigmetrics Achievement Award (2016), the INFORMS John von Neumann Theory Prize (2018), and the IEEE Control Systems Award (2018). He holds honorary doctorates from the Universit├й catholique de Louvain, (2008), the Athens University of Economics and Business (2018), and the Harokopio University (2019).

рдпреЛ рдЗ-рдкреБрд╕реНрддрдХрдХреЛ рдореВрд▓реНрдпрд╛рдЩреНрдХрди рдЧрд░реНрдиреБрд╣реЛрд╕реН

рд╣рд╛рдореАрд▓рд╛рдИ рдЖрдлреНрдиреЛ рдзрд╛рд░рдгрд╛ рдмрддрд╛рдЙрдиреБрд╣реЛрд╕реНред

рдЬрд╛рдирдХрд╛рд░реА рдкрдвреНрджреИ

рд╕реНрдорд╛рд░реНрдЯрдлреЛрди рддрдерд╛ рдЯреНрдпрд╛рдмрд▓реЗрдЯрд╣рд░реВ
Android рд░ iPad/iPhone рдХрд╛ рд▓рд╛рдЧрд┐┬аGoogle Play рдХрд┐рддрд╛рдм рдПрдк рдХреЛ рдЗрдиреНрд╕реНрдЯрд▓ рдЧрд░реНрдиреБрд╣реЛрд╕реНред рдпреЛ рддрдкрд╛рдИрдВрдХреЛ рдЦрд╛рддрд╛рд╕реЕрдВрдЧ рд╕реНрд╡рддрдГ рд╕рд┐рдВрдХ рд╣реБрдиреНрдЫ рд░ рддрдкрд╛рдИрдВ рдЕрдирд▓рд╛рдЗрди рд╡рд╛ рдЕрдлрд▓рд╛рдЗрди рдЬрд╣рд╛рдБ рднрдП рдкрдирд┐┬ардЕрдзреНрдпрдпрди рдЧрд░реНрди рджрд┐рдиреНрдЫред
рд▓реНрдпрд╛рдкрдЯрдк рддрдерд╛ рдХрдореНрдкреНрдпреБрдЯрд░рд╣рд░реВ
рддрдкрд╛рдИрдВ Google Play рдорд╛ рдЦрд░рд┐рдж рдЧрд░рд┐рдПрдХреЛ рдЕрдбрд┐рдпреЛрдмреБрдХ рдЖрдлреНрдиреЛ рдХрдореНрдкреНрдпреБрдЯрд░рдХреЛ рд╡реЗрдм рдмреНрд░рд╛рдЙрдЬрд░ рдкреНрд░рдпреЛрдЧ рдЧрд░реЗрд░ рд╕реБрдиреНрди рд╕рдХреНрдиреБрд╣реБрдиреНрдЫред
eReaders рд░ рдЕрдиреНрдп рдЙрдкрдХрд░рдгрд╣рд░реВ
Kobo eReaders рдЬрд╕реНрддрд╛ e-ink рдбрд┐рднрд╛рдЗрд╕рд╣рд░реВрдорд╛ рдлрд╛рдЗрд▓ рдкрдвреНрди рддрдкрд╛рдИрдВрд▓реЗ рдлрд╛рдЗрд▓ рдбрд╛рдЙрдирд▓реЛрдб рдЧрд░реЗрд░ рдЙрдХреНрдд рдлрд╛рдЗрд▓ рдЖрдлреНрдиреЛ рдбрд┐рднрд╛рдЗрд╕рдорд╛ рдЯреНрд░рд╛рдиреНрд╕реНрдлрд░ рдЧрд░реНрдиреБ рдкрд░реНрдиреЗ рд╣реБрдиреНрдЫред рддреА рдлрд╛рдЗрд▓рд╣рд░реВ рдкрдвреНрди рдорд┐рд▓реНрдиреЗ рдЗрдмреБрдХ рд░рд┐рдбрд░рд╣рд░реВрдорд╛ рддреА рдлрд╛рдЗрд▓рд╣рд░реВ рдЯреНрд░рд╛рдиреНрд╕реНрдлрд░ рдЧрд░реНрдиреЗрд╕рдореНрдмрдиреНрдзреА рд╡рд┐рд╕реНрддреГрдд рдирд┐рд░реНрджреЗрд╢рдирд╣рд░реВ рдкреНрд░рд╛рдкреНрдд рдЧрд░реНрди рдорджреНрджрдд рдХреЗрдиреНрджреНрд░ рдорд╛ рдЬрд╛рдиреБрд╣реЛрд╕реНред