Anderson et al., 2019 - Google Patents
Toward energy–quality scaling in deep neural networksAnderson et al., 2019
- Document ID
- 2152725798907739060
- Author
- Anderson J
- Alkabani Y
- El-Ghazawi T
- Publication year
- Publication venue
- IEEE Design & Test
External Links
Snippet
Editor's notes: This article surveys the latest advances in neural network (NN) architectures by applying them to the task of energy-quality scaling. Results show that, while coarse scaling is possible with existing NN architectures, fine-grain scaling is needed for fog …
- 230000001537 neural 0 title abstract description 32
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/38—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
- G06F7/48—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
- G06F7/52—Multiplying; Dividing
- G06F7/523—Multiplying only
- G06F7/53—Multiplying only in parallel-parallel fashion, i.e. both operands being entered in parallel
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
- G06F17/5009—Computer-aided design using simulation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/16—Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/11—Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/14—Fourier, Walsh or analogous domain transformations, e.g. Laplace, Hilbert, Karhunen-Loeve, transforms
- G06F17/141—Discrete Fourier transforms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/76—Architectures of general purpose stored programme computers
- G06F15/78—Architectures of general purpose stored programme computers comprising a single central processing unit
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
- G06F9/46—Multiprogramming arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/16—Combinations of two or more digital computers each having at least an arithmetic unit, a programme unit and a register, e.g. for a simultaneous processing of several programmes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F1/00—Details of data-processing equipment not covered by groups G06F3/00 - G06F13/00, e.g. cooling, packaging or power supply specially adapted for computer application
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Dave et al. | Hardware acceleration of sparse and irregular tensor computations of ml models: A survey and insights | |
Mohaidat et al. | A survey on neural network hardware accelerators | |
Wang et al. | LUTNet: Learning FPGA configurations for highly efficient neural network inference | |
EP3877913A1 (en) | Training neural network accelerators using mixed precision data formats | |
Azari et al. | An energy-efficient reconfigurable LSTM accelerator for natural language processing | |
Kim et al. | A low-cost fully integer-based CNN accelerator on FPGA for real-time traffic sign recognition | |
Sun et al. | A high-performance accelerator for large-scale convolutional neural networks | |
Wang et al. | Briefly Analysis about CNN Accelerator based on FPGA | |
Vo et al. | A deep learning accelerator based on a streaming architecture for binary neural networks | |
Yan et al. | FPGAN: an FPGA accelerator for graph attention networks with software and hardware co-optimization | |
Lu et al. | THETA: A high-efficiency training accelerator for DNNs with triple-side sparsity exploration | |
Chen et al. | Exploiting on-chip heterogeneity of versal architecture for gnn inference acceleration | |
Mazouz et al. | Automated CNN back-propagation pipeline generation for FPGA online training | |
Wang et al. | High-performance mixed-low-precision cnn inference accelerator on fpga | |
Raha et al. | Efficient hardware acceleration of emerging neural networks for embedded machine learning: An industry perspective | |
Zhan et al. | Field programmable gate array‐based all‐layer accelerator with quantization neural networks for sustainable cyber‐physical systems | |
Zhang et al. | Reconfigurable and energy-efficient architecture for deploying multi-layer RNNs on FPGA | |
Kwak et al. | BiNPU: A 33.0 MOP/s/LUT Binary Neural Network Inference Processor Showing 88.26% CIFAR10 Accuracy With 1.9 Mbit On-Chip Parameters in a 28 nm FPGA | |
Anderson et al. | Toward energy–quality scaling in deep neural networks | |
Lu et al. | SparseNN: A performance-efficient accelerator for large-scale sparse neural networks | |
Li et al. | Neuromorphic processor-oriented hybrid Q-format multiplication with adaptive quantization for tiny YOLO3 | |
Liu et al. | Exploiting wireless technology for energy-efficient accelerators with multiple dataflows and precision | |
Su et al. | Processing element architecture design for deep reinforcement learning with flexible block floating point exploiting signal statistics | |
Kowsalya | RETRACTED ARTICLE: A novel cognitive Wallace compressor based multi operand adders in CNN architecture for FPGA | |
Bhardwaj et al. | Low-Complex and Low-Power n-dimensional Gram–Schmidt Orthogonalization Architecture Design Methodology |