Processors for Embedded Vision
THIS TECHNOLOGY CATEGORY INCLUDES ANY DEVICE THAT EXECUTES VISION ALGORITHMS OR VISION SYSTEM CONTROL SOFTWARE
This technology category includes any device that executes vision algorithms or vision system control software. The following diagram shows a typical computer vision pipeline; processors are often optimized for the compute-intensive portions of the software workload.
The following examples represent distinctly different types of processor architectures for embedded vision, and each has advantages and trade-offs that depend on the workload. For this reason, many devices combine multiple processor types into a heterogeneous computing environment, often integrated into a single semiconductor component. In addition, a processor can be accelerated by dedicated hardware that improves performance on computer vision algorithms.
General-purpose CPUs
While computer vision algorithms can run on most general-purpose CPUs, desktop processors may not meet the design constraints of some systems. However, x86 processors and system boards can leverage the PC infrastructure for low-cost hardware and broadly-supported software development tools. Several Alliance Member companies also offer devices that integrate a RISC CPU core. A general-purpose CPU is best suited for heuristics, complex decision-making, network access, user interface, storage management, and overall control. A general purpose CPU may be paired with a vision-specialized device for better performance on pixel-level processing.
Graphics Processing Units
High-performance GPUs deliver massive amounts of parallel computing potential, and graphics processors can be used to accelerate the portions of the computer vision pipeline that perform parallel processing on pixel data. While General Purpose GPUs (GPGPUs) have primarily been used for high-performance computing (HPC), even mobile graphics processors and integrated graphics cores are gaining GPGPU capability—meeting the power constraints for a wider range of vision applications. In designs that require 3D processing in addition to embedded vision, a GPU will already be part of the system and can be used to assist a general-purpose CPU with many computer vision algorithms. Many examples exist of x86-based embedded systems with discrete GPGPUs.
Digital Signal Processors
DSPs are very efficient for processing streaming data, since the bus and memory architecture are optimized to process high-speed data as it traverses the system. This architecture makes DSPs an excellent solution for processing image pixel data as it streams from a sensor source. Many DSPs for vision have been enhanced with coprocessors that are optimized for processing video inputs and accelerating computer vision algorithms. The specialized nature of DSPs makes these devices inefficient for processing general-purpose software workloads, so DSPs are usually paired with a RISC processor to create a heterogeneous computing environment that offers the best of both worlds.
Field Programmable Gate Arrays (FPGAs)
Instead of incurring the high cost and long lead-times for a custom ASIC to accelerate computer vision systems, designers can implement an FPGA to offer a reprogrammable solution for hardware acceleration. With millions of programmable gates, hundreds of I/O pins, and compute performance in the trillions of multiply-accumulates/sec (tera-MACs), high-end FPGAs offer the potential for highest performance in a vision system. Unlike a CPU, which has to time-slice or multi-thread tasks as they compete for compute resources, an FPGA has the advantage of being able to simultaneously accelerate multiple portions of a computer vision pipeline. Since the parallel nature of FPGAs offers so much advantage for accelerating computer vision, many of the algorithms are available as optimized libraries from semiconductor vendors. These computer vision libraries also include preconfigured interface blocks for connecting to other vision devices, such as IP cameras.
Vision-Specific Processors and Cores
Application-specific standard products (ASSPs) are specialized, highly integrated chips tailored for specific applications or application sets. ASSPs may incorporate a CPU, or use a separate CPU chip. By virtue of their specialization, ASSPs for vision processing typically deliver superior cost- and energy-efficiency compared with other types of processing solutions. Among other techniques, ASSPs deliver this efficiency through the use of specialized coprocessors and accelerators. And, because ASSPs are by definition focused on a specific application, they are usually provided with extensive associated software. This same specialization, however, means that an ASSP designed for vision is typically not suitable for other applications. ASSPs’ unique architectures can also make programming them more difficult than with other kinds of processors; some ASSPs are not user-programmable.
NVIDIA Advances Robot Learning and Humanoid Development With New AI and Simulation Tools
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. New Project GR00T workflows and AI world model development technologies to accelerate robot dexterity, control, manipulation and mobility. Robotics developers can greatly accelerate their work on AI-enabled robots, including humanoids, using new AI and simulation tools and
How to Optimize and Accelerate GPUs for Graphical and Machine Vision Processing
This blog post was originally published at Geisel Software’s website. It is reprinted here with the permission of Geisel Software. How to optimize and accelerate GPUS: tools, techniques, and real-world scenarios Machine learning (ML) systems analyze tremendous amounts of data to identify hidden patterns and make predictions based on those patterns. This requires a very
Join e-con Systems’ Exclusive Webinar In Collaboration With Advantech
e-con Systems is excited to invite you to an exclusive joint webinar on “Unlocking AI at the Edge: How Embedded Vision Is Helping Take Giant Strides” in partnership with Advantech. Learn how embedded vision is transforming real-time decision-making across industries. Dec 17, 2024, Tuesday – 11:00 AM – 12:00 PM (CET) or Dec 18th, 2024,
Qualcomm Technologies’ IoT Strategy: A New Approach, a New Opportunity
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. Our new blueprint for enabling our partners and end customers to bring more smarts to the edge was one of the highlights of Investor Day Everything around us is either already a lot smarter or aiming to
‘A Glimpse Into the Future of the Car’: Takeaways from Day 2 of Snapdragon Summit 2024
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. The Qualcomm Oryon CPU and the Snapdragon Elite name make its way to the automotive space, signaling its importance to Qualcomm Snapdragon Summit 2024 shifted gears on Tuesday and moved from the mobile world to the automotive
How NVIDIA Jetson AGX Orin Helps Unlock the Power of Surround-view Camera Solutions
This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. Autonomous vehicles, such as warehouse robots, rely on precise maneuvering. NVIDIA Jetson AGX Orin™-powered surround-view cameras provide a perfectly synchronized solution, allowing these robots to move freely within designated areas without requiring intensive manual intervention.
Chiplets: Revolutionizing Semiconductor Design and Manufacturing
In the rapidly evolving world of semiconductors, chiplet technology is emerging as a groundbreaking approach that addresses many of the challenges faced by traditional monolithic System-on-Chip (SoC) designs. As Moore’s Law slows down, the semiconductor industry is seeking innovative solutions to increase performance and functionality without merely increasing transistor density, as detailed in the IDTechEx
Streamlining AI Inference Performance and Deployment with NVIDIA TensorRT-LLM Chunked Prefill
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. In this blog post, we take a closer look at chunked prefill, a feature of NVIDIA TensorRT-LLM that increases GPU utilization and simplifies the deployment experience for developers. This builds on our previous post discussing how advanced
Microchip to Accelerate Real-time Edge AI with NVIDIA Holoscan
PolarFire® FPGA Ethernet Sensor Bridge provides low-power multi-sensor bridging to NVIDIA edge AI platforms CHANDLER, Ariz., November 14, 2024 — To enable developers building artificial intelligence (AI)-driven sensor processing systems, Microchip Technology (Nasdaq: MCHP) has released its PolarFire® FPGA Ethernet Sensor Bridge that works with the NVIDIA Holoscan sensor processing platform. PolarFire FPGAs enable multi-protocol
Snapdragon Summit’s AI Highlights: A Look at the Future of On-device AI
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. Qualcomm Technologies sets new standards in AI performance for its latest mobile, automotive and Qualcomm AI Hub advancements Our annual Snapdragon Summit wrapped up with exciting new announcements centered on the future of on-device artificial intelligence (AI).
Electrification and Autonomy: A Semiconductor Content Boost to $1,000 per Car by 2029
This market research report was originally published at the Yole Group’s website. It is reprinted here with the permission of the Yole Group. Automotive OEMs are moving upstream in the semiconductor supply chain with diverse strategies. OUTLINE The automotive semiconductor market is expecting a significant CAGR of 11% between 2023 and 2029 to almost $100
Advanced Semiconductor Packaging: Key Materials and Processing Trends
As semiconductor packaging technologies evolve, advanced methods like 2.5D and 3D Cu-to-Cu hybrid bonding are essential for achieving higher performance and power efficiency. However, manufacturing these technologies to meet high performance and yield standards while fulfilling client requirements is complex. Challenges include developing the right materials and innovating packaging manufacturing techniques. IDTechEx‘s “Materials and Processing
How to Accelerate Larger LLMs Locally on RTX With LM Studio
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. GPU offloading makes massive models accessible on local RTX AI PCs and workstations. Editor’s note: This post is part of the AI Decoded series, which demystifies AI by making the technology more accessible, and showcases new hardware,
Introducing the First AMD 1B Language Models: AMD OLMo
This blog post was originally published at AMD’s website. It is reprinted here with the permission of AMD. In recent years, the rapid development of artificial intelligence technology, especially the progress in large language models (LLMs), has garnered significant attention and discussion. From the emergence of ChatGPT to subsequent models like GPT-4 and Llama, these
MIPS Releases P8700, Industry’s First High-performance AI-enabled RISC-V Automotive CPU for ADAS and Autonomous Vehicles
SAN JOSE, CA – October 30, 2024 – MIPS, a leading developer of efficient and configurable IP compute cores, announced today the general availability(GA) launch of the MIPS P8700 Series RISC-V Processor. Designed to meet the low-latency, highly intensive data movement demands of the most advanced automotive applications such as ADAS and Autonomous Vehicles (AVs),
Introducing Qualcomm IoT Solutions Framework: Making It Easier to Develop and Deploy Solutions
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. The Qualcomm IoT Solutions Framework represents a comprehensive suite of developer tools, reference blueprints and a robust ecosystem of partners Qualcomm Technologies, Inc. is known for its vast collection of wireless-related intellectual property and its processor and
Renesas Brings the High Performance of Arm Cortex-M85 Processor to Cost-sensitive Applications with New RA8 Entry-line MCU Groups
RA8E1 and RA8E2 Deliver Unmatched Scalar and Vector Compute Performance with Best-in-Class Feature Set to Address Value-Oriented Markets TOKYO, Japan, November 5, 2024 ― Renesas Electronics Corporation (TSE:6723), a premier supplier of advanced semiconductor solutions, today introduced the RA8E1 and RA8E2 microcontroller (MCU) groups, extending the industry’s most powerful series of MCUs. Introduced in 2023, the
Give AI a Look: Any Industry Can Now Search and Summarize Vast Volumes of Visual Data
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Accenture, Dell Technologies and Lenovo are among the companies tapping a new NVIDIA AI Blueprint to develop visual AI agents that can boost productivity, optimize processes and create safer spaces. Enterprises and public sector organizations around the
The Global Robotaxi Market Value in 2024 Will be $174B
For more information, visit https://www.idtechex.com/en/research-report/autonomous-vehicles-market-2025-2045-robotaxis-autonomous-cars-sensors/1045. The global robotaxi vehicle market value in 2045 will be US$174 billion, growing with a 20-year CAGR of 37% between 2025 and 2045 and with a market share dominated by leaders from the US and China, such as Google’s Waymo, GM’s Cruise, WeRide, Baidu, and AutoX. IDTechEx’s Autonomous Vehicles Market
NXP Expands Edge AI Capabilities with eIQ Software Enablement
NXP adds GenAI Flow with Retrieval Augmented Generation (RAG) fine-tuning and eIQ Time Series Studio to its eIQ AI and machine learning development software to make it easier to deploy and use AI across a broad spectrum of edge processors, from small microcontrollers (MCUs) to larger and more powerful applications processors (MPUs) What’s New NXP