In the previous article, we discussed the different types of AI chips. Among these, ASICs play a significant role. Compared to GPUs, ASICs are faster and more power-efficient, thanks to their customized architectures designed for specific AI tasks or algorithms.
One common misconception about AI ASICs is that they’re used solely for inference applications. However, AI ASICs come in a wide range of types and functionalities. In this article, we will explore the different types of AI ASICs and how they vary from one another.
Types of AI ASICs
AI ASICs are available in a diverse range, with options for nearly every imaginable AI task or algorithm. While they cover a broad spectrum of applications, they can generally be classified into the following categories:
- Deep learning ASICs
- Inference ASICs
- Natural language processing (NLP) ASICs
- Vision processing ASICs
Let’s delve deeper into each of these types to understand their applications and capabilities.
Deep learning ASICs
Deep learning ASICs are designed for training and inference in deep neural networks. Their architecture is optimized for the computationally intensive processes involved in AI training. They ensure computational power and throughput, featuring a flexible and general-purpose design to support a wide range of deep learning models.
These chips often use higher precision, such as 32-bit floating-point operations, to accommodate extensive training computations. However, they do tend to be larger and consume more power than other AI chips.
Deep learning ASICs incorporate specialized hardware components ideal for the demanding computations required for training and executing deep neural networks.
Key components include:
- Matrix multiplication units (MMUs): specialized hardware units optimized for matrix multiplications, which is a fundamental operation in most deep-learning algorithms. These units are the primary workhorses of deep learning ASICs.
- Convolutional neural network (CNN) accelerators: tailored to enhance the convolutional operations commonly used in image and video processing tasks. These units often employ systolic arrays to achieve high-throughput processing.
- Recurrent neural network (RNN) accelerators: optimized hardware units designed for processing sequential data. They’re specifically built to handle the recurrent nature of RNNs, including architectures like long short-term memory (LSTM) networks and gated recurrent units (GRUs).
By integrating these specialized components, deep learning ASICs enable high-performance computations for training and inference in deep learning applications.
Deep learning models process massive amounts of data, requiring these ASICs to feature high-bandwidth memory subsystems. These subsystems — including on-chip caches or scratchpad memory — minimize bottlenecks and enable efficient data transfer between processing units and on-chip memory.
These chips employ dataflow architectures that execute instructions only when all of the necessary data inputs are available. This design allows data to flow continuously through the chip, supporting concurrent computations. Additionally, many deep learning ASICs feature custom instruction sets tailored to the specific needs of deep learning algorithms. Together, these components create a specialized platform for deep learning computations.
While deep learning ASICs offer high performance, optimal power consumption, and a customized design for deep learning operations, their high degree of specialization can sometimes pose challenges. The customization may make adapting to new architectures or emerging deep learning models in a rapidly evolving AI landscape difficult. This highly tailored design can result in risks of vendor lock-in, potentially complicating transitions to different hardware platforms.
Some well-known deep learning ASICs include:
- Google Tensor Processing Units (TPUs), designed for machine learning workloads and extensively used in Google’s data centers
- Graphcore IPUs, optimized for graph neural networks and widely applied in natural language processing and recommendation systems
- Habana Labs Gaudi and Goya processors, engineered for high-performance deep learning training and inference
- Cerebras Wafer Scale Engine (WSE), ideal for large-scale AI training and inference tasks
- Qualcomm AI Engine, developed for on-device AI capabilities in smartphones and other mobile devices
Inference ASICs
Inference ASICs execute trained AI models in real-time applications. Minimizing power consumption is a key focus in their design, as these chips are often used in edge or mobile devices.
Despite their low power requirements, inference ASICs must deliver high throughput to handle real-time processing demands for applications such as video surveillance, speech recognition, and autonomous vehicles. Their compact size and ability to operate efficiently with low-precision arithmetic, such as 8-bit integers, make them ideal for these scenarios. Typically, these ASICs employ simpler dataflow architectures that prioritize efficient data movement to minimize latency.
The design of these chips is centered on efficiency and real-time performance. It often include dedicated CNN accelerators for convolutional operations, essential in image and video processing tasks. Digital signal processors (DSPs) may also be integrated for speech recognition and other audio-related AI applications. In some cases, hardware is tailored to specific model architectures, which is especially common in applications like natural language processing (NLP) and object detection.
Inference ASICs include multiple levels of memory, such as on-chip caches, scratchpad memory, and off-chip DRAM, which minimize data movement between processors and memory units.
Power management is another critical component, with various techniques employed to optimize energy use. These include dynamic voltage and frequency scaling (DVFS), voltage islands, power gating, clock gating, thermal management units, body biasing, and multi-mode operations. The power management unit (PMU) may either be centralized or distributed across multiple smaller PMUs within the chip. Some inference ASICs even adopt a hybrid approach, combining centralized and distributed PMUs.
Popular inference ASICs include:
- Google Edge TPU, designed for edge computing and on-device AI processing
- Qualcomm AI Engine, which supports on-device AI capabilities in smartphones
- Intel’s Myriad X, optimized for drones and robots
- Ambarella CVflow, used for real-time video and image analysis in security cameras, drones, and other video-centric applications
Natural language processing (NLP) ASICs
NLP ASICs are designed for natural language processing tasks such as speech recognition, sentiment analysis, and machine translation. These chips are optimized to handle the unique computational requirements of NLP algorithms and incorporate specialized hardware units for common NLP operations.
Examples include:
- Recurrent neural network (RNN) accelerators: process sequential data, making them ideal for tasks like speech recognition and machine translation
- Transformer accelerators: execute the computations required by Transformer models, a widely used class of deep neural networks in NLP
- Sparse matrix multipliers: handle sparse matrix multiplications, typical in NLP tasks
Given the need to process vast amounts of textual data, NLP ASICs feature high-bandwidth memory subsystems to transfer data between processing units and memory efficiently. This often includes multi-level memory systems with large on-chip storage for seamless data access. These chips also leverage parallel processing architectures to achieve high throughput and real-time performance.
A primary design focus of NLP ASICs is sequential data processing, using specialized hardware units like RNNs, Transformers, and sparse matrix multipliers to effectively address the demands of NLP workloads.
NLP ASICs are deployed in applications such as:
- Voice assistants: enabling real-time speech recognition and natural language understanding
- Chatbots: powering interactive and context-aware conversations
- Machine translation: facilitating real-time language conversion
- Sentiment analysis: providing real-time insights from social media and customer reviews
- Information retrieval: enhancing search engines and personalized recommendation systems
Some popular NLP ASICs include Cerebras Wafer Scale Engine (WSE), Graphcore IPUs, Habana Labs Gaudi, and other specialized chips developed by various startups.
Vision processing ASICs
Vision processing ASICs are designed for computer vision tasks like object detection, image recognition, and video analysis. These chips include specialized hardware units tailored for various computer vision operations, such as:
- Pixel processors: handle basic image operations like filtering, color space conversion, and edge detection
- Convolutional neural network (CNN) accelerators: accelerate convolutional operations in deep learning-based computer vision algorithms
- Histogram processors: perform feature extraction and analysis
- Optical flow processors: facilitate motion estimation and object tracking
Vision processing often requires handling large amounts of image data simultaneously. To meet this demand, these ASICs leverage highly parallel architectures, such as array processors and SIMD (Single Instruction, Multiple Data) units, to execute multiple operations concurrently.
Some chips also use dataflow architectures to efficiently pipeline tasks. Additionally, many vision ASICs are designed to interface directly with image sensors, reducing data transfer bottlenecks and minimizing latency.
These ASICs are utilized in various applications, including:
- Surveillance systems: enhancing real-time object recognition and motion detection
- Medical imaging: supporting advanced diagnostic imaging technologies
- Autonomous vehicles: performing object detection, lane keeping, and pedestrian recognition tasks
- Smartphones: enabling advanced camera features and augmented reality applications
Some notable vision-processing ASICs include:
- Mobileye EyeQ Chips: used in advanced driver-assistance systems (ADAS) and autonomous vehicles
- Qualcomm Spectra ISP: integrated into Snapdragon processors for advanced camera features
- Ambarella CVflow: employed in security cameras, drones, and other real-time video analysis applications
- Esperanto Systems Tango Neural Processor: designed for high-performance deep learning inference in vision tasks
- Syntiant NDP (Neural Decision Processors): developed for always-on AI applications, including vision processing in smart home devices and wearables
Analog ASICs
While most AI ASICs discussed previously are digital signal ICs, an emerging segment in AI ASICs involves analog ASICs. These chips leverage analog circuits to perform computations directly on analog signals, offering unprecedented power efficiency and speed advantages. Analog ASICs are particularly well-suited for edge devices, where low power consumption is critical.
Examples include:
- Memristive devices: use resistance that changes based on the history of current flow, allowing them to simulate synapses in a neural network by storing and processing information within the memory itself (examples include memristors and resistive RAM)
- Spintronic devices: use the spin of electrons to perform computations (examples include spintronic transistors and magnetic tunnel junctions or MTJs)
- Quantum-dot cellular automata (QCA): employs arrays of quantum dots to represent and manipulate information at the quantum level
- Brain-inspired circuits: mimic the structure and functionality of biological neurons and synapses, enabling neuromorphic computing
A few applications and challenges include:
Analog ASICs hold significant potential for applications in edge devices and neuromorphic computing. However, they face several challenges that must be addressed to realize their full potential:
- Limited precision: analog computations generally have lower precision than digital counterparts, which can impact performance
- Noise sensitivity: analog circuits are highly noise-sensitive, making ensuring consistent and reliable operation difficult
- Integration with digital circuits: developing a seamless integration of analog ASICs with existing digital systems presents a significant technical hurdle
- Scalability: despite breakthroughs, scaling analog ASICs for widespread use remains a significant challenge
Analog ASICs represent a promising frontier in AI hardware, but overcoming these technical barriers will be critical to their adoption and success in real-world applications.
AI ASICs as mixed-signal circuits
Mixed-signal ASICs integrate digital and analog circuits, combining the strengths of each. For example, some neuromorphic chips use digital control logic alongside analog neuron models. These chips are instrumental in applications like sensor fusion, where data from analog sensors (such as cameras, microphones, and accelerometers) are integrated with digital processing for tasks like robotics.
Applications include:
- Sensor fusion chips: handle signal conditioning (e.g., filtering and amplification) in the analog domain to reduce noise and extract relevant features before converting the data to digital form. Once digitized, digital circuits process the data for AI tasks such as object recognition, sound classification, and multi-sensor integration.
- Neuromorphic chips: use analog circuits to replicate biological neurons’ behavior (e.g., spiking neurons) and digital circuits for control and communication. For instance, analog circuits simulate synaptic connections, while digital circuits manage network architecture and learning processes.
- Mixed-signal ADCs/DACs: combine analog and digital circuitry to perform high-resolution analog-to-digital (ADC) and digital-to-analog (DAC) conversions, which are critical for many AI applications.
- RF front-ends for AI: integrate analog RF components (e.g., amplifiers, filters, mixers) with digital signal processing for wireless communication. They’re widely used in IoT devices and self-driving cars.
- Power management units (PMUs):
- Dynamic voltage and frequency scaling (DVFS): these PMUs use analog circuits to dynamically adjust the voltage and frequency of digital processing units based on computational demands, optimizing power consumption.
- Power gating PMUs: employ analog switches to selectively deactivate inactive parts of the chip, further enhancing energy efficiency.
Designing and manufacturing mixed-signal ASICs is inherently more complex than working with purely digital or analog chips.
The integration of domains requires advanced engineering, including:
- Ensuring compatibility and seamless operation between analog and digital circuits
- Thorough testing and verification to ensure proper functionality and reliability
- Overcoming challenges in noise isolation and signal integrity to achieve consistent performance
Mixed-signal ASICs represent a significant advancement in AI hardware, enabling sophisticated processing capabilities across various applications while balancing power efficiency and functionality.
You may also like:
Filed Under: Tech Articles









Questions related to this article?
👉Ask and discuss on Electro-Tech-Online.com and EDAboard.com forums.
Tell Us What You Think!!
You must be logged in to post a comment.