Processors

Processors for Embedded Vision

THIS TECHNOLOGY CATEGORY INCLUDES ANY DEVICE THAT EXECUTES VISION ALGORITHMS OR VISION SYSTEM CONTROL SOFTWARE

This technology category includes any device that executes vision algorithms or vision system control software. The following diagram shows a typical computer vision pipeline; processors are often optimized for the compute-intensive portions of the software workload.

ev pipeline

The following examples represent distinctly different types of processor architectures for embedded vision, and each has advantages and trade-offs that depend on the workload. For this reason, many devices combine multiple processor types into a heterogeneous computing environment, often integrated into a single semiconductor component. In addition, a processor can be accelerated by dedicated hardware that improves performance on computer vision algorithms.

General-purpose CPUs

While computer vision algorithms can run on most general-purpose CPUs, desktop processors may not meet the design constraints of some systems. However, x86 processors and system boards can leverage the PC infrastructure for low-cost hardware and broadly-supported software development tools. Several Alliance Member companies also offer devices that integrate a RISC CPU core. A general-purpose CPU is best suited for heuristics, complex decision-making, network access, user interface, storage management, and overall control. A general purpose CPU may be paired with a vision-specialized device for better performance on pixel-level processing.

Graphics Processing Units

High-performance GPUs deliver massive amounts of parallel computing potential, and graphics processors can be used to accelerate the portions of the computer vision pipeline that perform parallel processing on pixel data. While General Purpose GPUs (GPGPUs) have primarily been used for high-performance computing (HPC), even mobile graphics processors and integrated graphics cores are gaining GPGPU capability—meeting the power constraints for a wider range of vision applications. In designs that require 3D processing in addition to embedded vision, a GPU will already be part of the system and can be used to assist a general-purpose CPU with many computer vision algorithms. Many examples exist of x86-based embedded systems with discrete GPGPUs.

Digital Signal Processors

DSPs are very efficient for processing streaming data, since the bus and memory architecture are optimized to process high-speed data as it traverses the system. This architecture makes DSPs an excellent solution for processing image pixel data as it streams from a sensor source. Many DSPs for vision have been enhanced with coprocessors that are optimized for processing video inputs and accelerating computer vision algorithms. The specialized nature of DSPs makes these devices inefficient for processing general-purpose software workloads, so DSPs are usually paired with a RISC processor to create a heterogeneous computing environment that offers the best of both worlds.

Field Programmable Gate Arrays (FPGAs)

Instead of incurring the high cost and long lead-times for a custom ASIC to accelerate computer vision systems, designers can implement an FPGA to offer a reprogrammable solution for hardware acceleration. With millions of programmable gates, hundreds of I/O pins, and compute performance in the trillions of multiply-accumulates/sec (tera-MACs), high-end FPGAs offer the potential for highest performance in a vision system. Unlike a CPU, which has to time-slice or multi-thread tasks as they compete for compute resources, an FPGA has the advantage of being able to simultaneously accelerate multiple portions of a computer vision pipeline. Since the parallel nature of FPGAs offers so much advantage for accelerating computer vision, many of the algorithms are available as optimized libraries from semiconductor vendors. These computer vision libraries also include preconfigured interface blocks for connecting to other vision devices, such as IP cameras.

Vision-Specific Processors and Cores

Application-specific standard products (ASSPs) are specialized, highly integrated chips tailored for specific applications or application sets. ASSPs may incorporate a CPU, or use a separate CPU chip. By virtue of their specialization, ASSPs for vision processing typically deliver superior cost- and energy-efficiency compared with other types of processing solutions. Among other techniques, ASSPs deliver this efficiency through the use of specialized coprocessors and accelerators. And, because ASSPs are by definition focused on a specific application, they are usually provided with extensive associated software. This same specialization, however, means that an ASSP designed for vision is typically not suitable for other applications. ASSPs’ unique architectures can also make programming them more difficult than with other kinds of processors; some ASSPs are not user-programmable.

“Key Requirements to Successfully Implement Generative AI in Edge Devices—Optimized Mapping to the Enhanced NPX6 Neural Processing Unit IP,” a Presentation from Synopsys

Gordon Cooper, Principal Product Manager at Synopsys, presents the “Key Requirements to Successfully Implement Generative AI in Edge Devices—Optimized Mapping to the Enhanced NPX6 Neural Processing Unit IP” tutorial at the May 2025 Embedded Vision Summit. In this talk, Cooper discusses emerging trends in generative AI for edge devices and… “Key Requirements to Successfully Implement

Read More »

AMD Unveils Vision for an Open AI Ecosystem, Detailing New Silicon, Software and Systems at Advancing AI 2025

Only AMD powers the full spectrum of AI, bringing together leadership GPUs, CPUs, networking and open software to deliver unmatched flexibility and performance Meta, OpenAI, xAI, Oracle, Microsoft, Cohere, HUMAIN, Red Hat, Astera Labs and Marvell discussed how they are partnering with AMD for AI solutions SANTA CLARA, Calif., June 12, 2025 (GLOBE NEWSWIRE) —

Read More »

AI and Computer Vision Insights at CVPR 2025

This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. Our papers, demos, workshops and tutorial continue our leadership in generative AI and learning systems At Qualcomm AI Research, we are advancing AI to make its core capabilities — perception, reasoning and action — ubiquitous across devices.

Read More »

“Bridging the Gap: Streamlining the Process of Deploying AI onto Processors,” a Presentation from SqueezeBits

Taesu Kim, Chief Technology Officer at SqueezeBits, presents the “Bridging the Gap: Streamlining the Process of Deploying AI onto Processors” tutorial at the May 2025 Embedded Vision Summit. Large language models (LLMs) often demand hand-coded conversion scripts for deployment on each distinct processor-specific software stack—a process that’s time-consuming and prone… “Bridging the Gap: Streamlining the

Read More »

“From Enterprise to Makers: Driving Vision AI Innovation at the Extreme Edge,” a Presentation from Sony Semiconductor Solutions

Amir Servi, Edge Deep Learning Product Manager at Sony Semiconductor Solutions, presents the “From Enterprise to Makers: Driving Vision AI Innovation at the Extreme Edge” tutorial at the May 2025 Embedded Vision Summit. Sony’s unique integrated sensor-processor technology is enabling ultra-efficient intelligence directly at the image source, transforming vision AI… “From Enterprise to Makers: Driving

Read More »

AI Helps Locate Dangerous Fishing Nets Lost at Sea

This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Conservationists have launched a new AI tool that can sift through petabytes of underwater imaging from anywhere in the world to identify signs of abandoned or lost fishing nets—so-called ghost nets. Each year, around 2% of the

Read More »

VeriSilicon’s Ultra-low Energy NPU Provides Over 40 TOPS for On-device LLM Inference in Mobile Applications

The energy-efficient architecture scales across AI-enabled devices, including AI phones and AI PCs Shanghai, China, June 9, 2025–VeriSilicon (688521.SH) today announced that its ultra-low energy and high-performance Neural Network Processing Unit (NPU) IP now supports on-device inference of large language models (LLMs) with AI computing performance scaling beyond 40 TOPS. This energy-efficient NPU architecture is

Read More »

VeriSilicon’s Scalable High-performance GPGPU-AI Computing IPs Empower Automotive and Edge Server AI Solutions

Provide AI acceleration with high computing density, multi-chip scaling, and 3D-stacked memory integration Shanghai, China, June 9, 2025–VeriSilicon (688521.SH) today announced the latest advancements in its high-performance and scalable GPGPU-AI computing IPs, which are now empowering next-generation automotive electronics and edge server applications. Combining programmable parallel computing with a dedicated Artificial Intelligence (AI) accelerator, these

Read More »

VeriSilicon’s AI-ISP Custom Chip Solution Enables Mass Production of Customer’s Smartphones

Providing architecture design, software-hardware co-development, and mass production support, and enhancing AI-powered imaging capabilities in smart devices Shanghai, China, June 9, 2025–VeriSilicon (688521.SH) recently announced that its AI-ISP custom chip solution has been successfully adopted in a customer’s mass-produced smartphones, reaffirming the company’s comprehensive one-stop custom silicon service capabilities in AI vision processing. VeriSilicon’s AI-ISP

Read More »

The SHD Group Releases New Edge AI Processor and Ecosystem Report

Now available for free download from Alliance Member company The SHD Group is their latest market report, Edge AI Market Analysis: Applications, Processors and Ecosystem Guide, developed in partnership with the Edge AI and Vision Alliance. The report provides a detailed exploration of the rapidly evolving edge AI landscape, covering critical insights on emerging applications,

Read More »

“Why It’s Critical to Have an Integrated Development Methodology for Edge AI,” a Presentation from Lattice Semiconductor

Sreepada Hegade, Director of ML Systems and Software at Lattice Semiconductor, presents the “Why It’s Critical to Have an Integrated Development Methodology for Edge AI” tutorial at the May 2025 Embedded Vision Summit. The deployment of neural networks near sensors brings well-known advantages such as lower latency, privacy and reduced… “Why It’s Critical to Have

Read More »

“Solving Tomorrow’s AI Problems Today with Cadence’s Newest Processor,” a Presentation from Cadence

Amol Borkar, Product Marketing Director at Cadence, presents the “Solving Tomorrow’s AI Problems Today with Cadence’s Newest Processor” tutorial at the May 2025 Embedded Vision Summit. Artificial intelligence is rapidly integrating into every aspect of technology. While the neural processing unit (NPU) often receives the majority of the spotlight as… “Solving Tomorrow’s AI Problems Today

Read More »

Andes Technology Unveils AndesAIRE AnDLA I370: A Next-generation Deep Learning Accelerator for Edge and Endpoint AI

Hsinchu, Taiwan –June 05, 2025 – Andes Technology, a leading supplier of high-performance, low-power 32/64-bit RISC-V processor cores and AI acceleration solutions, proudly announces the launch of AndesAIRE™ AnDLA™ I370, its latest Deep Learning Accelerator (DLA), engineered to bring advanced neural network performance to cost-sensitive edge and endpoint AI applications. Building on the success of

Read More »

“State-space Models vs. Transformers for Ultra-low-power Edge AI,” a Presentation from BrainChip

Tony Lewis, Chief Technology Officer at BrainChip, presents the “State-space Models vs. Transformers for Ultra-low-power Edge AI” tutorial at the May 2025 Embedded Vision Summit. At the embedded edge, choices of language model architectures have profound implications on the ability to meet demanding performance, latency and energy efficiency requirements. In… “State-space Models vs. Transformers for

Read More »

AMD Acquires Brium to Strengthen Open AI Software Ecosystem

News Highlights Brium’s world-class compiler and AI software experience will strengthen AMD’s ability to deliver highly optimized AI solutions across the entire stack Will reduce developer dependencies on specific hardware configurations and enable accelerated out of the box AI performance Brium’s domain-specific expertise will expand AMD’s market reach across industries such as healthcare, life sciences,

Read More »

Expedera’s Origin Evolution NPU IP Brings Generative AI to Edge Devices

Origin Evolution NPU IP uses Expedera’s unique packet-based architecture to achieve unprecedented NPU efficiency. Highlights Expedera launches its Origin Evolution™ NPU IP, bringing hardware acceleration to meet the computational demands of running LLMs on resource-constrained edge devices. New purpose-built hardware and software architecture runs LLMs and traditional neural networks with ultra-efficient PPA, providing fully scalable

Read More »

Automakers Can Launch AI-defined Vehicles a Full Year Sooner – Powered by Arm Zena CSS

News Highlights: Arm enables automakers to bring new vehicle models to market at least one year faster than traditional timelines Zena CSS accelerates software and silicon development, helping deliver AI capabilities faster and more efficiently Pre-verified, safety-capable platform saves an estimated 20% of engineering resources, reducing development cost and complexity Majority of OEMs will build

Read More »

Here you’ll find a wealth of practical technical insights and expert advice to help you bring AI and visual intelligence into your products without flying blind.

Contact

Address

Berkeley Design Technology, Inc.
PO Box #4446
Walnut Creek, CA 94596

Phone
Phone: +1 (925) 954-1411
Scroll to Top