Acceleration for ML 論文導讀

Energy efficient parallel neuromorphic architectures with approximate arithmetic on FPGA

Motivation

To address the slow operation and high energy and resource consumption problem caused by realizing spiking neural network (SNN) using software.app

Problem

  1. software : slow operation, high energy consumption and space resources
  2. analog circuits: hard to reconfigure and intrinsically sensitive to process, voltage and temperature (PVT) Var.
  3. FPGA: most of works focus on the acceleration of SNN without considering energy consumption and efficiency of resource utilization.
  4. This work presented the parallel neuromorphic processor architectures with approximate arithmetic for SNN on FPGA.

There is no related work part in this paper.ide


In-Datacenter Performance Analysis of a Tensor Processing Unit

Motivation

This paper evaluates a custom ASIC - called a Tensor Processing Unit (TPU) to accelerates the inference phase of neural networks (NN).ui

Problem

Many NN applications have hard response time deadline. Hence, inference phase must response quickly when user do some action. While CPU and GPU are poor in response.this

All works are focus on hardware processing, such as DRAM, hardware protocol and so on.lua

相關文章
相關標籤/搜索