MARC보기
LDR05397cmm u2200553Ki 4500
001000000317339
003OCoLC
00520230525182402
006m d
007cr cnu---unuuu
008191102s2019 enk o 000 0 eng d
015 ▼a GBB9H7912 ▼2 bnb
0167 ▼a 019591497 ▼2 Uk
020 ▼a 1789139708
020 ▼a 9781789139709 ▼q (electronic bk.)
035 ▼a 2278656 ▼b (N$T)
035 ▼a (OCoLC)1125107738
037 ▼a D518365C-A419-4AF3-8B4C-AD04FD491FF8 ▼b OverDrive, Inc. ▼n http://www.overdrive.com
040 ▼a EBLCP ▼b eng ▼e pn ▼c EBLCP ▼d TEFOD ▼d UKMGB ▼d EBLCP ▼d OCLCF ▼d OCLCQ ▼d UKAHL ▼d MERUC ▼d OCLCQ ▼d N$T ▼d 248032
049 ▼a MAIN
050 4 ▼a QA76.9.A43
08204 ▼a 005.1 ▼2 23
1001 ▼a Lonza, Andrea.
24510 ▼a Reinforcement Learning Algorithms with Python : ▼b Learn, Understand, and Develop Smart Algorithms for Addressing AI Challenges / ▼c Andrea Lonza.
260 ▼a Birmingham : ▼b Packt Publishing, Limited, ▼c 2019.
300 ▼a 1 online resource (356 pages)
336 ▼a text ▼b txt ▼2 rdacontent
337 ▼a computer ▼b c ▼2 rdamedia
338 ▼a online resource ▼b cr ▼2 rdacarrier
500 ▼a Implementing REINFORCE with baseline
5050 ▼a Cover; Title Page; Copyright and Credits; Dedication; About Packt; Contributors; Table of Contents; Preface; Section 1: Algorithms and Environments; Chapter 1: The Landscape of Reinforcement Learning; An introduction to RL; Comparing RL and supervised learning; History of RL; Deep RL; Elements of RL; Policy; The value function; Reward; Model; Applications of RL; Games; Robotics and Industry 4.0; Machine learning; Economics and finance; Healthcare; Intelligent transportation systems; Energy optimization and smart grid; Summary; Questions; Further reading
5058 ▼a Chapter 2: Implementing RL Cycle and OpenAI GymSetting up the environment; Installing OpenAI Gym; Installing Roboschool; OpenAI Gym and RL cycles; Developing an RL cycle; Getting used to spaces; Development of ML models using TensorFlow; Tensor; Constant; Placeholder; Variable; Creating a graph; Simple linear regression example; Introducing TensorBoard; Types of RL environments; Why different environments?; Open source environments; Summary; Questions; Further reading; Chapter 3: Solving Problems with Dynamic Programming; MDP; Policy; Return; Value functions; Bellman equation
5058 ▼a Categorizing RL algorithmsModel-free algorithms; Value-based algorithms; Policy gradient algorithms; Actor-Critic algorithms; Hybrid algorithms; Model-based RL; Algorithm diversity; Dynamic programming; Policy evaluation and policy improvement; Policy iteration; Policy iteration applied to FrozenLake; Value iteration; Value iteration applied to FrozenLake; Summary; Questions; Further reading; Section 2: Model-Free RL Algorithms; Chapter 4: Q-Learning and SARSA Applications; Learning without a model; User experience; Policy evaluation; The exploration problem; Why explore?; How to explore
5058 ▼a TD learningTD update; Policy improvement; Comparing Monte Carlo and TD; SARSA; The algorithm; Applying SARSA to Taxi-v2; Q-learning; Theory; The algorithm; Applying Q-learning to Taxi-v2; Comparing SARSA and Q-learning; Summary; Questions; Chapter 5: Deep Q-Network; Deep neural networks and Q-learning; Function approximation; Q-learning with neural networks; Deep Q-learning instabilities; DQN; The solution; Replay memory; The target network; The DQN algorithm; The loss function; Pseudocode; Model architecture; DQN applied to Pong; Atari games; Preprocessing; DQN implementation; DNNs
5058 ▼a The experienced bufferThe computational graph and training loop; Results; DQN variations; Double DQN; DDQN implementation; Results; Dueling DQN; Dueling DQN implementation; Results; N-step DQN; Implementation; Results; Summary; Questions; Further reading; Chapter 6: Learning Stochastic and PG Optimization; Policy gradient methods; The gradient of the policy; Policy gradient theorem; Computing the gradient; The policy; On-policy PG; Understanding the REINFORCE algorithm; Implementing REINFORCE; Landing a spacecraft using REINFORCE; Analyzing the results; REINFORCE with baseline
520 ▼a With this book, you will understand the core concepts and techniques of reinforcement learning. You will take a look into each RL algorithm and will develop your own self-learning algorithms and models. You will optimize the algorithms for better precision, use high-speed actions and lower the risk of anomalies in your applications.
5880 ▼a Print version record.
590 ▼a Added to collection customer.56279.3
650 0 ▼a Computer algorithms.
650 0 ▼a Python (Computer program language)
650 7 ▼a Computer algorithms. ▼2 fast ▼0 (OCoLC)fst00872010
650 7 ▼a Python (Computer program language) ▼2 fast ▼0 (OCoLC)fst01084736
655 4 ▼a Electronic books.
77608 ▼i Print version: ▼a Lonza, Andrea. ▼t Reinforcement Learning Algorithms with Python : Learn, Understand, and Develop Smart Algorithms for Addressing AI Challenges. ▼d Birmingham : Packt Publishing, Limited, 짤2019 ▼z 9781789131116
85640 ▼3 EBSCOhost ▼u http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=2278656
938 ▼a Askews and Holts Library Services ▼b ASKH ▼n AH36843042
938 ▼a ProQuest Ebook Central ▼b EBLB ▼n EBL5964771
938 ▼a EBSCOhost ▼b EBSC ▼n 2278656
990 ▼a 관리자
994 ▼a 92 ▼b N$T