We all learn through trial and error. We avoid the things that cause us to experience pain and failure. We embrace and build on the things that give us reward and success. This common pattern is the foundation of deep reinforcement learning: building machine learning systems that explore and learn based on the responses of the environment. Grokking Deep Reinforcement Learning introduces this powerful machine learning approach, using examples, illustrations, exercises, and crystal-clear teaching. You'll love the perfectly paced teaching and the clever, engaging writing style as you dig into this awesome exploration of reinforcement learning fundamentals, effective deep learning techniques, and practical applications in this emerging field.
About the technology
We learn by interacting with our environment, and the rewards or punishments we experience guide our future behavior. Deep reinforcement learning brings that same natural process to artificial intelligence, analyzing results to uncover the most efficient ways forward. DRL agents can improve marketing campaigns, predict stock performance, and beat grand masters in Go and chess.
About the book
Grokking Deep Reinforcement Learning uses engaging exercises to teach you how to build deep learning systems. This book combines annotated Python code with intuitive explanations to explore DRL techniques. You'll see how algorithms function and learn to develop your own DRL agents using evaluative feedback.
Author(s): Miguel Morales
Edition: 1
Publisher: Manning Publications
Year: 2020
Language: English
Commentary: Vector PDF
Pages: 450
City: Shelter Island, NY
Tags: Deep Learning; Reinforcement Learning; Agent-based AI; Actor-Critic Method; Policy-Gradient Method; General Artifical Intelligence
foreword
preface
acknowledgments
about this book
about the author
Introduction todeep reinforcement learning
What is deep reinforcement learning?
The past, present, and future of deep reinforcement learning
The suitability of deep reinforcement learning
Setting clear two-way expectations
Mathematical foundationsof reinforcement learning
Components of reinforcement learning
MDPs: The engine of the environment
Balancing immediateand long-term goals
The objective of a decision-making agent
Planning optimal sequences of actions
Balancing the gatheringand use of information
The challenge of interpreting evaluative feedback
Strategic exploration
Evaluatingagents’ behaviors
Learning to estimate the value of policies
Learning to estimate from multiple steps
Improvingagents’ behaviors
The anatomy of reinforcement learning agents
Learning to improve policies of behavior
Decoupling behavior from learning
Achieving goals moreeffectively and efficiently
Learning to improve policies using robust targets
Agents that interact, learn, and plan
Introduction to value-baseddeep reinforcement learning
The kind of feedback deep reinforcement learning agents use
Introduction to function approximation for reinforcement learning
NFQ: The first attempt at value-based deep reinforcement learning
More stablevalue-based methods
DQN: Making reinforcement learning more like supervised learning
Double DQN: Mitigating the overestimation of action-value functions
Sample-efficientvalue-based methods
Dueling DDQN: A reinforcement-learning-aware neural network architecture
PER: Prioritizing the replay of meaningful experiences
Policy-gradient andactor-critic methods
REINFORCE: Outcome-based policy learning
VPG: Learning a value function
A3C: Parallel policy updates
GAE: Robust advantage estimation
A2C: Synchronous policy updates
Advancedactor-critic methods
DDPG: Approximating a deterministic policy
TD3: State-of-the-art improvements over DDPG
SAC: Maximizing the expected return and entropy
PPO: Restricting optimization steps
Toward artificialgeneral intelligence
What was covered and what notably wasn’t?
More advanced concepts toward AGI
What happens next?
index