找回密码
 To register

QQ登录

只需一步,快速开始

扫一扫,访问微社区

Titlebook: Deep Reinforcement Learning with Python; With PyTorch, Tensor Nimish Sanghi Book 20211st edition Nimish Sanghi 2021 Artificial Intelligence

[复制链接]
查看: 19442|回复: 47
发表于 2025-3-21 19:03:20 | 显示全部楼层 |阅读模式
书目名称Deep Reinforcement Learning with Python
副标题With PyTorch, Tensor
编辑Nimish Sanghi
视频video
概述Explains deep reinforcement learning implementation using TensorFlow, PyTorch and OpenAI Gym..Covers deep reinforcement implementation using CNN and deep q-networks.Explains deep-q learning and policy
图书封面Titlebook: Deep Reinforcement Learning with Python; With PyTorch, Tensor Nimish Sanghi Book 20211st edition Nimish Sanghi 2021 Artificial Intelligence
描述Deep reinforcement learning is a fast-growing discipline that is making a significant impact in fields of autonomous vehicles, robotics, healthcare, finance, and many more. This book covers deep reinforcement learning using deep-q learning and policy gradient models with coding exercise..You‘ll begin by reviewing the Markov decision processes, Bellman equations, and dynamic programming that form the core concepts and foundation of deep reinforcement learning. Next, you‘ll study model-free learning followed by function approximation using neural networks and deep learning. This is followed by various deep reinforcement learning algorithms such as deep q-networks, various flavors of actor-critic methods, and other policy-based methods. .You‘ll also look at exploration vs exploitation dilemma, a key consideration in reinforcement learning algorithms, along with Monte Carlo tree search (MCTS), which played a key role inthe success of AlphaGo. The final chapters conclude with deep reinforcement learning implementation using popular deep learning frameworks such as TensorFlow and PyTorch. In the end, you‘ll understand deep reinforcement learning along with deep q networks and policy grad
出版日期Book 20211st edition
关键词Artificial Intelligence; Deep Reinforcement Learning; PyTorch; Neural Networks; Robotics; Autonomous Vehi
版次1
doihttps://doi.org/10.1007/978-1-4842-6809-4
isbn_ebook978-1-4842-6809-4
copyrightNimish Sanghi 2021
The information of publication is updating

书目名称Deep Reinforcement Learning with Python影响因子(影响力)




书目名称Deep Reinforcement Learning with Python影响因子(影响力)学科排名




书目名称Deep Reinforcement Learning with Python网络公开度




书目名称Deep Reinforcement Learning with Python网络公开度学科排名




书目名称Deep Reinforcement Learning with Python被引频次




书目名称Deep Reinforcement Learning with Python被引频次学科排名




书目名称Deep Reinforcement Learning with Python年度引用




书目名称Deep Reinforcement Learning with Python年度引用学科排名




书目名称Deep Reinforcement Learning with Python读者反馈




书目名称Deep Reinforcement Learning with Python读者反馈学科排名




单选投票, 共有 0 人参与投票
 

0票 0%

Perfect with Aesthetics

 

0票 0%

Better Implies Difficulty

 

0票 0%

Good and Satisfactory

 

0票 0%

Adverse Performance

 

0票 0%

Disdainful Garbage

您所在的用户组没有投票权限
发表于 2025-3-21 21:57:44 | 显示全部楼层
http://image.papertrans.cn/d/image/264660.jpg
发表于 2025-3-22 02:48:00 | 显示全部楼层
https://doi.org/10.1007/978-1-4842-6809-4Artificial Intelligence; Deep Reinforcement Learning; PyTorch; Neural Networks; Robotics; Autonomous Vehi
发表于 2025-3-22 04:45:01 | 显示全部楼层
Implementing Continuous Integrationas led to many significant advances that are increasingly getting machines closer to acting the way humans do. In this book, we will start with the basics and finish up with mastering some of the most recent developments in the field. There will be a good mix of theory (with minimal mathematics) and
发表于 2025-3-22 11:21:24 | 显示全部楼层
发表于 2025-3-22 12:53:35 | 显示全部楼层
Marc Joseph Saugey Restoration,earns a policy π(.| .) that maps states to actions. The agent uses this policy to take an action . = . when in state . = .. The system transitions to the next time instant of . + 1. The environment responds to the action (. = .) by putting the agent in a new state of . = . and providing feedback to
发表于 2025-3-22 18:17:47 | 显示全部楼层
发表于 2025-3-22 21:52:39 | 显示全部楼层
https://doi.org/10.1007/978-3-642-70880-0rlo approach (MC), and finally using the temporal difference (TD) approach. In all these approaches, we always looked at problems where the state space and actions were both discrete. Only in the previous chapter toward the end did we talk about Q-learning in a continuous state space. We discretized
发表于 2025-3-23 05:01:22 | 显示全部楼层
发表于 2025-3-23 06:18:21 | 显示全部楼层
What Is the Microsoft HoloLens? a given current policy. In a second step, these estimated values were used to find a better policy by choosing the best action in a given state. These two steps were carried out in a loop again and again until no further improvement in values was observed. In this chapter, we will look at a differe
 关于派博传思  派博传思旗下网站  友情链接
派博传思介绍 公司地理位置 论文服务流程 影响因子官网 SITEMAP 大讲堂 北京大学 Oxford Uni. Harvard Uni.
发展历史沿革 期刊点评 投稿经验总结 SCIENCEGARD IMPACTFACTOR 派博系数 清华大学 Yale Uni. Stanford Uni.
|Archiver|手机版|小黑屋| 派博传思国际 ( 京公网安备110108008328) GMT+8, 2025-5-25 09:09
Copyright © 2001-2015 派博传思   京公网安备110108008328 版权所有 All rights reserved
快速回复 返回顶部 返回列表