[go: up one dir, main page]

Anderson et al., 2019 - Google Patents

Toward energy–quality scaling in deep neural networks

Anderson et al., 2019

Document ID
2152725798907739060
Author
Anderson J
Alkabani Y
El-Ghazawi T
Publication year
Publication venue
IEEE Design & Test

External Links

Snippet

Editor's notes: This article surveys the latest advances in neural network (NN) architectures by applying them to the task of energy-quality scaling. Results show that, while coarse scaling is possible with existing NN architectures, fine-grain scaling is needed for fog …
Continue reading at ieeexplore.ieee.org (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/38Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
    • G06F7/48Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
    • G06F7/52Multiplying; Dividing
    • G06F7/523Multiplying only
    • G06F7/53Multiplying only in parallel-parallel fashion, i.e. both operands being entered in parallel
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/50Computer-aided design
    • G06F17/5009Computer-aided design using simulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/11Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/14Fourier, Walsh or analogous domain transformations, e.g. Laplace, Hilbert, Karhunen-Loeve, transforms
    • G06F17/141Discrete Fourier transforms
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored programme computers
    • G06F15/78Architectures of general purpose stored programme computers comprising a single central processing unit
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for programme control, e.g. control unit
    • G06F9/06Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
    • G06F9/46Multiprogramming arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a programme unit and a register, e.g. for a simultaneous processing of several programmes
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N99/00Subject matter not provided for in other groups of this subclass
    • G06N99/005Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F1/00Details of data-processing equipment not covered by groups G06F3/00 - G06F13/00, e.g. cooling, packaging or power supply specially adapted for computer application

Similar Documents

Publication Publication Date Title
Dave et al. Hardware acceleration of sparse and irregular tensor computations of ml models: A survey and insights
Mohaidat et al. A survey on neural network hardware accelerators
Wang et al. LUTNet: Learning FPGA configurations for highly efficient neural network inference
EP3877913A1 (en) Training neural network accelerators using mixed precision data formats
Azari et al. An energy-efficient reconfigurable LSTM accelerator for natural language processing
Kim et al. A low-cost fully integer-based CNN accelerator on FPGA for real-time traffic sign recognition
Sun et al. A high-performance accelerator for large-scale convolutional neural networks
Wang et al. Briefly Analysis about CNN Accelerator based on FPGA
Vo et al. A deep learning accelerator based on a streaming architecture for binary neural networks
Yan et al. FPGAN: an FPGA accelerator for graph attention networks with software and hardware co-optimization
Lu et al. THETA: A high-efficiency training accelerator for DNNs with triple-side sparsity exploration
Chen et al. Exploiting on-chip heterogeneity of versal architecture for gnn inference acceleration
Mazouz et al. Automated CNN back-propagation pipeline generation for FPGA online training
Wang et al. High-performance mixed-low-precision cnn inference accelerator on fpga
Raha et al. Efficient hardware acceleration of emerging neural networks for embedded machine learning: An industry perspective
Zhan et al. Field programmable gate array‐based all‐layer accelerator with quantization neural networks for sustainable cyber‐physical systems
Zhang et al. Reconfigurable and energy-efficient architecture for deploying multi-layer RNNs on FPGA
Kwak et al. BiNPU: A 33.0 MOP/s/LUT Binary Neural Network Inference Processor Showing 88.26% CIFAR10 Accuracy With 1.9 Mbit On-Chip Parameters in a 28 nm FPGA
Anderson et al. Toward energy–quality scaling in deep neural networks
Lu et al. SparseNN: A performance-efficient accelerator for large-scale sparse neural networks
Li et al. Neuromorphic processor-oriented hybrid Q-format multiplication with adaptive quantization for tiny YOLO3
Liu et al. Exploiting wireless technology for energy-efficient accelerators with multiple dataflows and precision
Su et al. Processing element architecture design for deep reinforcement learning with flexible block floating point exploiting signal statistics
Kowsalya RETRACTED ARTICLE: A novel cognitive Wallace compressor based multi operand adders in CNN architecture for FPGA
Bhardwaj et al. Low-Complex and Low-Power n-dimensional Gram–Schmidt Orthogonalization Architecture Design Methodology