Fr. 150.00

Artificial Intelligence Hardware Design - Challenges and Solutions

English · Hardback

Shipping usually within 1 to 3 weeks (not available at short notice)

Description

Read more

ARTIFICIAL INTELLIGENCE HARDWARE DESIGN
 
Learn foundational and advanced topics in Neural Processing Unit design with real-world examples from leading voices in the field
 
In Artificial Intelligence Hardware Design: Challenges and Solutions, distinguished researchers and authors Drs. Albert Chun Chen Liu and Oscar Ming Kin Law deliver a rigorous and practical treatment of the design applications of specific circuits and systems for accelerating neural network processing. Beginning with a discussion and explanation of neural networks and their developmental history, the book goes on to describe parallel architectures, streaming graphs for massive parallel computation, and convolution optimization.
 
The authors offer readers an illustration of in-memory computation through Georgia Tech's Neurocube and Stanford's Tetris accelerator using the Hybrid Memory Cube, as well as near-memory architecture through the embedded eDRAM of the Institute of Computing Technology, the Chinese Academy of Science, and other institutions.
 
Readers will also find a discussion of 3D neural processing techniques to support multiple layer neural networks, as well as information like:
* A thorough introduction to neural networks and neural network development history, as well as Convolutional Neural Network (CNN) models
* Explorations of various parallel architectures, including the Intel CPU, Nvidia GPU, Google TPU, and Microsoft NPU, emphasizing hardware and software integration for performance improvement
* Discussions of streaming graph for massive parallel computation with the Blaize GSP and Graphcore IPU
* An examination of how to optimize convolution with UCLA Deep Convolutional Neural Network accelerator filter decomposition
 
Perfect for hardware and software engineers and firmware developers, Artificial Intelligence Hardware Design is an indispensable resource for anyone working with Neural Processing Units in either a hardware or software capacity.

List of contents

Author Biographies xi
 
Preface xiii
 
Acknowledgments xv
 
Table of Figures xvii
 
1 Introduction 1
 
1.1 Development History 2
 
1.2 Neural Network Models 4
 
1.3 Neural Network Classification 4
 
1.3.1 Supervised Learning 4
 
1.3.2 Semi-supervised Learning 5
 
1.3.3 Unsupervised Learning 6
 
1.4 Neural Network Framework 6
 
1.5 Neural Network Comparison 10
 
Exercise 11
 
References 12
 
2 Deep Learning 13
 
2.1 Neural Network Layer 13
 
2.1.1 Convolutional Layer 13
 
2.1.2 Activation Layer 17
 
2.1.3 Pooling Layer 18
 
2.1.4 Normalization Layer 19
 
2.1.5 Dropout Layer 20
 
2.1.6 Fully Connected Layer 20
 
2.2 Deep Learning Challenges 22
 
Exercise 22
 
References 24
 
3 Parallel Architecture 25
 
3.1 Intel Central Processing Unit (CPU) 25
 
3.1.1 Skylake Mesh Architecture 27
 
3.1.2 Intel Ultra Path Interconnect (UPI) 28
 
3.1.3 Sub Non-unified Memory Access Clustering (SNC) 29
 
3.1.4 Cache Hierarchy Changes 31
 
3.1.5 Single/Multiple Socket Parallel Processing 32
 
3.1.6 Advanced Vector Software Extension 33
 
3.1.7 Math Kernel Library for Deep Neural Network (MKL-DNN) 34
 
3.2 NVIDIA Graphics Processing Unit (GPU) 39
 
3.2.1 Tensor Core Architecture 41
 
3.2.2 Winograd Transform 44
 
3.2.3 Simultaneous Multithreading (SMT) 45
 
3.2.4 High Bandwidth Memory (HBM2) 46
 
3.2.5 NVLink2 Configuration 47
 
3.3 NVIDIA Deep Learning Accelerator (NVDLA) 49
 
3.3.1 Convolution Operation 50
 
3.3.2 Single Data Point Operation 50
 
3.3.3 Planar Data Operation 50
 
3.3.4 Multiplane Operation 50
 
3.3.5 Data Memory and Reshape Operations 51
 
3.3.6 System Configuration 51
 
3.3.7 External Interface 52
 
3.3.8 Software Design 52
 
3.4 Google Tensor Processing Unit (TPU) 53
 
3.4.1 System Architecture 53
 
3.4.2 Multiply-Accumulate (MAC) Systolic Array 55
 
3.4.3 New Brain Floating-Point Format 55
 
3.4.4 Performance Comparison 57
 
3.4.5 Cloud TPU Configuration 58
 
3.4.6 Cloud Software Architecture 60
 
3.5 Microsoft Catapult Fabric Accelerator 61
 
3.5.1 System Configuration 64
 
3.5.2 Catapult Fabric Architecture 65
 
3.5.3 Matrix-Vector Multiplier 65
 
3.5.4 Hierarchical Decode and Dispatch (HDD) 67
 
3.5.5 Sparse Matrix-Vector Multiplication 68
 
Exercise 70
 
References 71
 
4 Streaming Graph Theory 73
 
4.1 Blaize Graph Streaming Processor 73
 
4.1.1 Stream Graph Model 73
 
4.1.2 Depth First Scheduling Approach 75
 
4.1.3 Graph Streaming Processor Architecture 76
 
4.2 Graphcore Intelligence Processing Unit 79
 
4.2.1 Intelligence Processor Unit Architecture 79
 
4.2.2 Accumulating Matrix Product (AMP) Unit 79
 
4.2.3 Memory Architecture 79
 
4.2.4 Interconnect Architecture 79
 
4.2.5 Bulk Synchronous Parallel Model 81
 
Exercise 83
 
References 84
 
5 Convolution Optimization 85
 
5.1 Deep Convolutional Neural Network Accelerator 85
 
5.1.1 System Architecture 86
 
5.1.2 Filter Decomposition 87
 
5.1.3 Streaming Architecture 90
 
5.1.3.1 Filter Weights Reuse 90
 
5.1.3.2 Input Channel Reuse 92
 
5.1.4 Pooling 92
 
5.1.4.1 Average Pooling 92
 
5.1.4.2 Max Pooling 93
 
5.1.5 Conv

About the author










Albert Chun Chen Liu, PhD, is Chief Executive Officer of Kneron. He is Adjunct Associate Professor at National Tsing Hua University, National Chiao Tung University, and National Cheng Kung University. He has published over 15 IEEE papers and is an IEEE Senior Member. He is a recipient of the IBM Problem Solving Award based on the use of the EIP tool suite in 2007 and IEEE TCAS Darlington award in 2021.
Oscar Ming Kin Law, PhD, is the Director of Engineering at Kneron. He works on smart robot development and in-memory architecture for neural networks. He has over twenty years of experience in the semiconductor industry working with CPU, GPU, and mobile design. He has also published over 60 patents in various areas.

Summary

ARTIFICIAL INTELLIGENCE HARDWARE DESIGN

Learn foundational and advanced topics in Neural Processing Unit design with real-world examples from leading voices in the field

In Artificial Intelligence Hardware Design: Challenges and Solutions, distinguished researchers and authors Drs. Albert Chun Chen Liu and Oscar Ming Kin Law deliver a rigorous and practical treatment of the design applications of specific circuits and systems for accelerating neural network processing. Beginning with a discussion and explanation of neural networks and their developmental history, the book goes on to describe parallel architectures, streaming graphs for massive parallel computation, and convolution optimization.

The authors offer readers an illustration of in-memory computation through Georgia Tech's Neurocube and Stanford's Tetris accelerator using the Hybrid Memory Cube, as well as near-memory architecture through the embedded eDRAM of the Institute of Computing Technology, the Chinese Academy of Science, and other institutions.

Readers will also find a discussion of 3D neural processing techniques to support multiple layer neural networks, as well as information like:
* A thorough introduction to neural networks and neural network development history, as well as Convolutional Neural Network (CNN) models
* Explorations of various parallel architectures, including the Intel CPU, Nvidia GPU, Google TPU, and Microsoft NPU, emphasizing hardware and software integration for performance improvement
* Discussions of streaming graph for massive parallel computation with the Blaize GSP and Graphcore IPU
* An examination of how to optimize convolution with UCLA Deep Convolutional Neural Network accelerator filter decomposition

Perfect for hardware and software engineers and firmware developers, Artificial Intelligence Hardware Design is an indispensable resource for anyone working with Neural Processing Units in either a hardware or software capacity.

Customer reviews

No reviews have been written for this item yet. Write the first review and be helpful to other users when they decide on a purchase.

Write a review

Thumbs up or thumbs down? Write your own review.

For messages to CeDe.ch please use the contact form.

The input fields marked * are obligatory

By submitting this form you agree to our data privacy statement.