KNOW GPU AND ASIC

1 min readDec 22, 2024

This article helps people understand ASIC and GPU which are two major chips used in the AI world. Broadcom makes special purpose ASIC chips. Nvidia makes general purpose GPU chips.

ASIC

— designs specific algorithms or functions into hardware circuits, such as loss functions, activation functions, normalization processes

— can use reduced number precisions for speedup

— uses SRAM (faster, expensive, low capacity)

— can customize parallel and pipelineed circuits

— good at sparse matrix

— energy efficient

— can become obsolete soon due to new AI architecture and algorithms

— higher cost of designing and printing for each generation and iteration

— good at logic/computes, not for video graphics

— more like DIY projects. Programming API different

— can work at datacenter, edge AI; training and inference

GPU

— flexible, works for any algorithm

— software developers customize precisions

— use HBM memory (on board, fast, large capacity)

— no specialized circuits. 30000 cores for parallelism

— works for both sparse and dense matrices

— higher energy consumption

— lasts longer for a wider range of applications

— high technology barrier to entry

— good at video graphics

— general purpose with common API

— can work at datacenter, edge AI; training and inference

FPGA

— ignore this if you don’t want to learn the details

— logic on the chip can be changed (reprogrammed) after manufacturing

— can be used to prototype the ASIC

— inefficient for real applications

— FPGA is to Python as ASIC is to assemble language

--

--

Jonathan Yue, PhD
Jonathan Yue, PhD

Written by Jonathan Yue, PhD

Enthusiast on vector databases, AI, RAG, data science, consensus algorithms, distributed systems. Initiator and developer of the JaguarDB vector database

No responses yet