HOME > 상세정보

상세정보

Deep reinforcement learning : frontiers of artificial intelligence

Deep reinforcement learning : frontiers of artificial intelligence (5회 대출)

자료유형
단행본
개인저자
Sewak, Mohit.
서명 / 저자사항
Deep reinforcement learning : frontiers of artificial intelligence / Mohit Sewak.
발행사항
Singapore :   Springer,   c2019.  
형태사항
xvii, 203 p. : ill. (chiefly col.) ; 25 cm.
ISBN
9789811382840
내용주기
Introduction to Reinforcement Learning -- Mathematical and Algorithmic understanding of Reinforcement Learning -- Coding the Environment and MDP Solution -- Temporal Difference Learning, SARSA, and Q Learning -- Q Learning in Code -- Introduction to Deep Learning -- Implementation Resources -- Deep Q Network (DQN), Double DQN and Dueling DQN -- Double DQN in Code -- Policy-Based Reinforcement Learning Approaches -- Actor-Critic Models & the A3C -- A3C in Code -- Deterministic Policy Gradient and the DDPG -- DDPG in Code.
서지주기
Includes bibliographical references and index.
000 00000nam u2200205 a 4500
001 000046023294
005 20200409091318
008 200408s2019 si a b 001 0 eng d
020 ▼a 9789811382840
040 ▼a 211009 ▼c 211009 ▼d 211009
082 0 4 ▼a 005.11 ▼2 23
084 ▼a 005.11 ▼2 DDCK
090 ▼a 005.11 ▼b S512d
100 1 ▼a Sewak, Mohit.
245 1 0 ▼a Deep reinforcement learning : ▼b frontiers of artificial intelligence / ▼c Mohit Sewak.
260 ▼a Singapore : ▼b Springer, ▼c c2019.
300 ▼a xvii, 203 p. : ▼b ill. (chiefly col.) ; ▼c 25 cm.
504 ▼a Includes bibliographical references and index.
505 0 ▼a Introduction to Reinforcement Learning -- Mathematical and Algorithmic understanding of Reinforcement Learning -- Coding the Environment and MDP Solution -- Temporal Difference Learning, SARSA, and Q Learning -- Q Learning in Code -- Introduction to Deep Learning -- Implementation Resources -- Deep Q Network (DQN), Double DQN and Dueling DQN -- Double DQN in Code -- Policy-Based Reinforcement Learning Approaches -- Actor-Critic Models & the A3C -- A3C in Code -- Deterministic Policy Gradient and the DDPG -- DDPG in Code.
945 ▼a KLPA

소장정보

No. 소장처 청구기호 등록번호 도서상태 반납예정일 예약 서비스
No. 1 소장처 과학도서관/Sci-Info(2층서고)/ 청구기호 005.11 S512d 등록번호 121252916 (5회 대출) 도서상태 대출가능 반납예정일 예약 서비스 B M

컨텐츠정보

책소개

This book starts by presenting the basics of reinforcement learning using highly intuitive and easy-to-understand examples and applications, and then introduces the cutting-edge research advances that make reinforcement learning capable of out-performing most state-of-art systems, and even humans in a number of applications. The book not only equips readers with an understanding of multiple advanced and innovative algorithms, but also prepares them to implement systems such as those created by Google Deep Mind in actual code.

This book is intended for readers who want to both understand and apply advanced concepts in a field that combines the best of two worlds ? deep learning and reinforcement learning ? to tap the potential of ‘advanced artificial intelligence’ for creating real-world applications and game-winning algorithms.




New feature

This book starts by presenting the basics of reinforcement learning using highly intuitive and easy-to-understand examples and applications, and then introduces the cutting-edge research advances that make reinforcement learning capable of out-performing most state-of-art systems, and even humans in a number of applications. The book not only equips readers with an understanding of multiple advanced and innovative algorithms, but also prepares them to implement systems such as those created by Google Deep Mind in actual code.

 This book is intended for readers who want to both understand and apply advanced concepts in a field that combines the best of two worlds ? deep learning and reinforcement learning ? to tap the potential of ‘advanced artificial intelligence’ for creating real-world applications and game-winning algorithms.





정보제공 : Aladin

목차

Chapter 1: Introduction to Reinforcement LearningChapter Goal: In this chapter the reader will learn about the basic building blocks of Reinforcement Learning like state, actor, action and policy, and some traditional approaches of RL including Dynamic Programming and Monte Carlo. The readers will also learn the difference between On-Policy and Off-Policy methods and when to use which one. The chapter also covers some advanced non-Deep RL techniques. The chapter ends with some hands-on experience of coding some of the stated techniques. No of pages : 30
Chapter 2: Introduction to Deep LearningChapter Goal: In this chapter the readers will understand the building blocks of Deep Learning, and some popular types like DNN, RNN and CNN. The chapter also covers some example application of DL in areas like language modeling and image classification, and the challenges with using these techniques, where RL could help. No of pages: 25
Chapter 3: Introduction to Open AI GymChapter Goal: In this chapter, the readers will learn about the Open AI Gym. We would explore different games and sections within Open AI gym, and then would also explore the Open AI Gym Python API for the important games and section in Open AI Gym. The chapter will be concluded by coding an actual RL application in Open AI Gym using Python. No of pages: 10
Chapter 3: Temporal Difference Learning and TD-LambdaChapter Goal: This chapter begins with introducing the readers to the very effective, and highly in demand algorithm called Temporal Difference Learning, in its simplest form. To draw some intuition, we would cover RL algorithms like SARSA in brief and how does TD Learning compare with it. The chapter then builds upon these concepts, to introduce the advance TD-Lambda formulation of TD Learning which aims to combine insights across difference steps to choose the next action. The chapter concludes with giving the reader some hands-on experience of coding some of the discussed TD algorithm in Python. No of pages: 30
Chapter 4: Attention Mechanism and Reinforcement LearningChapter Goal: This chapter steps into the realm of combining both Deep Learning and Reinforcement Learning, to power many of the most powerful applications and real-life business use-cases. Much of discussions for this chapter will have its roots into cutting-edge research and the innovative applications. The chapter will start with introducing the problems with DL and how Attention Mechanism can solve it. We would next discuss the types of Attention, and then the role of RL in Attention, especially in Hard Attention. The chapter also covers use cases in Advanced NLP and Image Captioning with the role of Attention and RL in them. No of pages: 25
Chapter 5: Deep Q LearningChapter Goal: In the previous chapter we discovered how RL helps Deep Learning networks, especially the one that could benefit using Attention Mechanism, in this chapter we will reverse the role and will discuss how Deep Learning could take RL to the next step and answer many challenges that we highlighted earlier. It starts with the introduction to Policy Optimization and the Q Learning method. From there the chapter logically leads into Deep Q Learning while combining insights from the earlier covered chapter on Deep Learning, and earlier section on Q-Learning. The chapter next covers some of the active research in this area and advancements happening in this very powerful RL technique. No of pages: 25
Chapter 6: Applying Deep Q Learning in CodeChapter Goal: This chapter covers step by step the coding of Deep Q Learning, in a modular and Object-Oriented manner, covering class by class, explaining the complicated parts of the code in detail. No of pages: 15
Chapter 7: Asynchronous Advantage Actor Critic ModelChapter Goal: This chapter introduces the readers to the future of Artificial Intelligence and Reinforcement Learning with the Asynchronous Advantage Actor-Critic Model. This is a fairly new model with much of the recent and advanced researches being done in recent times, and a very powerful one as well, as this is the model behind the most successful applications from Google Deep Mind. But at the same time, this model is also very complex to understand and more complicated to implement. This chapter tries to take a lot of example and intuitions to explain the concepts of Actor-Critic models and then slowly moves to explaining and then mastering the A3C model. No of pages: 25
Chapter 8: Applying A3C Model in CodeChapter Goal: This chapter covers step by step the coding of Deep Q Learning, in a modular and Object-Oriented manner, covering class by class, explaining the complicated parts of the code in detail. No of pages: 15

관련분야 신착자료

Harvard Business Review (2025)