找回密码
 To register

QQ登录

只需一步,快速开始

扫一扫,访问微社区

Titlebook: Data Orchestration in Deep Learning Accelerators; Tushar Krishna,Hyoukjun Kwon,Ananda Samajdar Book 2020 Springer Nature Switzerland AG 20

[复制链接]
查看: 7877|回复: 39
发表于 2025-3-21 17:19:43 | 显示全部楼层 |阅读模式
书目名称Data Orchestration in Deep Learning Accelerators
编辑Tushar Krishna,Hyoukjun Kwon,Ananda Samajdar
视频video
丛书名称Synthesis Lectures on Computer Architecture
图书封面Titlebook: Data Orchestration in Deep Learning Accelerators;  Tushar Krishna,Hyoukjun Kwon,Ananda Samajdar Book 2020 Springer Nature Switzerland AG 20
描述This Synthesis Lecture focuses on techniques for efficient data orchestration within DNN accelerators. The End of Moore‘s Law, coupled with the increasing growth in deep learning and other AI applications has led to the emergence of custom Deep Neural Network (DNN) accelerators for energy-efficient inference on edge devices. Modern DNNs have millions of hyper parameters and involve billions of computations; this necessitates extensive data movement from memory to on-chip processing engines. It is well known that the cost of data movement today surpasses the cost of the actual computation; therefore, DNN accelerators require careful orchestration of data across on-chip compute, network, and memory elements to minimize the number of accesses to external DRAM. The book covers DNN dataflows, data reuse, buffer hierarchies, networks-on-chip, and automated design-space exploration. It concludes with data orchestration challenges with compressed and sparse DNNs and future trends. The target audience is students, engineers, and researchers interested in designing high-performance and low-energy accelerators for DNN inference.
出版日期Book 2020
版次1
doihttps://doi.org/10.1007/978-3-031-01767-4
isbn_softcover978-3-031-00639-5
isbn_ebook978-3-031-01767-4Series ISSN 1935-3235 Series E-ISSN 1935-3243
issn_series 1935-3235
copyrightSpringer Nature Switzerland AG 2020
The information of publication is updating

书目名称Data Orchestration in Deep Learning Accelerators影响因子(影响力)




书目名称Data Orchestration in Deep Learning Accelerators影响因子(影响力)学科排名




书目名称Data Orchestration in Deep Learning Accelerators网络公开度




书目名称Data Orchestration in Deep Learning Accelerators网络公开度学科排名




书目名称Data Orchestration in Deep Learning Accelerators被引频次




书目名称Data Orchestration in Deep Learning Accelerators被引频次学科排名




书目名称Data Orchestration in Deep Learning Accelerators年度引用




书目名称Data Orchestration in Deep Learning Accelerators年度引用学科排名




书目名称Data Orchestration in Deep Learning Accelerators读者反馈




书目名称Data Orchestration in Deep Learning Accelerators读者反馈学科排名




单选投票, 共有 0 人参与投票
 

0票 0%

Perfect with Aesthetics

 

0票 0%

Better Implies Difficulty

 

0票 0%

Good and Satisfactory

 

0票 0%

Adverse Performance

 

0票 0%

Disdainful Garbage

您所在的用户组没有投票权限
发表于 2025-3-21 21:55:53 | 显示全部楼层
Dataflow and Data Reuse,to billions of computations, we cannot fit all of the computations within an accelerator, which typically has hundreds to thousands of compute units. Therefore, we need to slice the problem into smaller chunks (i.e., computation tiles) and run them in a certain order (i.e., tile scheduling). Within
发表于 2025-3-22 01:21:47 | 显示全部楼层
Buffer Hierarchies,ic accelerators have constraints and goals that differ in key ways. It is important to understand in detail how these cause accelerator architects to make different hardware choices. In this chapter, we present a framework for understanding key options, and explore tradeoffs between design effort an
发表于 2025-3-22 07:24:38 | 显示全部楼层
Networks-on-Chip, contain an array of hundreds of PEs. These accelerators aim to achieve high throughput by exploiting massive parallel computations over the PEs while keeping the cost-of-operation much lower than off-the-shelf components with the same compute budget. However, adding more compute elements in an acce
发表于 2025-3-22 11:50:47 | 显示全部楼层
发表于 2025-3-22 13:00:49 | 显示全部楼层
发表于 2025-3-22 19:12:22 | 显示全部楼层
发表于 2025-3-23 00:52:58 | 显示全部楼层
Buffer Hierarchies,ic accelerators have constraints and goals that differ in key ways. It is important to understand in detail how these cause accelerator architects to make different hardware choices. In this chapter, we present a framework for understanding key options, and explore tradeoffs between design effort and cross-project reuse.
发表于 2025-3-23 02:15:42 | 显示全部楼层
Jason Gu,Rajeeb Dey,Nabanita Adhikaryrovide a brief background on Deep Neural Networks (DNNs), which are the underlying computational mechanisms within Deep Learning applications. Our objective is not to go into the theory behind the structure and accuracy of DNNs (which readers can find in any modern textbook on Machine Learning or De
发表于 2025-3-23 06:55:08 | 显示全部楼层
and the Co-production of Men’s Healthto billions of computations, we cannot fit all of the computations within an accelerator, which typically has hundreds to thousands of compute units. Therefore, we need to slice the problem into smaller chunks (i.e., computation tiles) and run them in a certain order (i.e., tile scheduling). Within
 关于派博传思  派博传思旗下网站  友情链接
派博传思介绍 公司地理位置 论文服务流程 影响因子官网 SITEMAP 大讲堂 北京大学 Oxford Uni. Harvard Uni.
发展历史沿革 期刊点评 投稿经验总结 SCIENCEGARD IMPACTFACTOR 派博系数 清华大学 Yale Uni. Stanford Uni.
|Archiver|手机版|小黑屋| 派博传思国际 ( 京公网安备110108008328) GMT+8, 2025-6-25 21:42
Copyright © 2001-2015 派博传思   京公网安备110108008328 版权所有 All rights reserved
快速回复 返回顶部 返回列表