Vision Algorithms

Vision Algorithms for Embedded Vision

Most computer vision algorithms were developed on general-purpose computer systems with software written in a high-level language

Most computer vision algorithms were developed on general-purpose computer systems with software written in a high-level language. Some of the pixel-processing operations (ex: spatial filtering) have changed very little in the decades since they were first implemented on mainframes. With today’s broader embedded vision implementations, existing high-level algorithms may not fit within the system constraints, requiring new innovation to achieve the desired results.

Some of this innovation may involve replacing a general-purpose algorithm with a hardware-optimized equivalent. With such a broad range of processors for embedded vision, algorithm analysis will likely focus on ways to maximize pixel-level processing within system constraints.

This section refers to both general-purpose operations (ex: edge detection) and hardware-optimized versions (ex: parallel adaptive filtering in an FPGA). Many sources exist for general-purpose algorithms. The Embedded Vision Alliance is one of the best industry resources for learning about algorithms that map to specific hardware, since Alliance Members will share this information directly with the vision community.

General-purpose computer vision algorithms

Introduction To OpenCV Figure 1

One of the most-popular sources of computer vision algorithms is the OpenCV Library. OpenCV is open-source and currently written in C, with a C++ version under development. For more information, see the Alliance’s interview with OpenCV Foundation President and CEO Gary Bradski, along with other OpenCV-related materials on the Alliance website.

Hardware-optimized computer vision algorithms

Several programmable device vendors have created optimized versions of off-the-shelf computer vision libraries. NVIDIA works closely with the OpenCV community, for example, and has created algorithms that are accelerated by GPGPUs. MathWorks provides MATLAB functions/objects and Simulink blocks for many computer vision algorithms within its Vision System Toolbox, while also allowing vendors to create their own libraries of functions that are optimized for a specific programmable architecture. National Instruments offers its LabView Vision module library. And Xilinx is another example of a vendor with an optimized computer vision library that it provides to customers as Plug and Play IP cores for creating hardware-accelerated vision algorithms in an FPGA.

Other vision libraries

  • Halcon
  • Matrox Imaging Library (MIL)
  • Cognex VisionPro
  • VXL
  • CImg
  • Filters

“Understand the Multimodal World with Minimal Supervision,” a Keynote Presentation from Yong Jae Lee

Yong Jae Lee, Associate Professor in the Department of Computer Sciences at the University of Wisconsin-Madison and CEO of GivernyAI, presents the “Learning to Understand Our Multimodal World with Minimal Supervision” tutorial at the May 2024 Embedded Vision Summit. The field of computer vision is undergoing another profound change. Recently,… “Understand the Multimodal World with

Read More »

“Identifying and Mitigating Bias in AI,” a Presentation from Intel

Nikita Tiwari, AI Enabling Engineer for OEM PC Experiences in the Client Computing Group at Intel, presents the “Identifying and Mitigating Bias in AI” tutorial at the May 2024 Embedded Vision Summit. From autonomous driving to immersive shopping, and from enhanced video collaboration to graphic design, AI is placing a… “Identifying and Mitigating Bias in

Read More »

“The Fundamentals of Training AI Models for Computer Vision Applications,” a Presentation from GMAC Intelligence

Amit Mate, Founder and CEO of GMAC Intelligence, presents the “Fundamentals of Training AI Models for Computer Vision Applications” tutorial at the May 2024 Embedded Vision Summit. In this presentation, Mate introduces the essential aspects of training convolutional neural networks (CNNs). He discusses the prerequisites for training, including models, data… “The Fundamentals of Training AI

Read More »

Snapdragon Powers the Future of AI in Smart Glasses. Here’s How

This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. A Snapdragon Insider chats with Qualcomm Technologies’ Said Bakadir about the future of smart glasses and Qualcomm Technologies’ role in turning it into a critical AI tool Artificial intelligence (AI) is increasingly winding its way through our

Read More »

“An Introduction to Semantic Segmentation,” a Presentation from Au-Zone Technologies

Sébastien Taylor, Vice President of Research and Development for Au-Zone Technologies, presents the “Introduction to Semantic Segmentation” tutorial at the May 2024 Embedded Vision Summit. Vision applications often rely on object detectors, which determine the nature and location of objects in a scene. But many vision applications require a different… “An Introduction to Semantic Segmentation,”

Read More »

Upcoming Webinar Explores How AI Can Make Cameras See In the Dark

On September 10, 2024 at 9:00 am PT (noon ET), Alliance Member companies Ceva and Visionary.ai will deliver the free webinar “Can AI Make Cameras See In the Dark? Real-Time Video Enhancement.” From the event page: As cameras become ubiquitous in applications such as surveillance, mobile, drones, and automotive systems, achieving clear vision 24/7 under

Read More »

NVIDIA TensorRT Model Optimizer v0.15 Boosts Inference Performance and Expands Model Support

This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. NVIDIA has announced the latest v0.15 release of NVIDIA TensorRT Model Optimizer, a state-of-the-art quantization toolkit of model optimization techniques including quantization, sparsity, and pruning. These techniques reduce model complexity and enable downstream inference frameworks like NVIDIA

Read More »

“DNN Quantization: Theory to Practice,” a Presentation from AMD

Dwith Chenna, Member of the Technical Staff and Product Engineer for AI Inference at AMD, presents the “DNN Quantization: Theory to Practice” tutorial at the May 2024 Embedded Vision Summit. Deep neural networks, widely used in computer vision tasks, require substantial computation and memory resources, making it challenging to run… “DNN Quantization: Theory to Practice,”

Read More »

“Leveraging Neural Architecture Search for Efficient Computer Vision on the Edge,” a Presentation from NXP Semiconductors

Hiram Rayo Torres Rodriguez, Senior AI Research Engineer at NXP Semiconductors, presents the “Leveraging Neural Architecture Search for Efficient Computer Vision on the Edge” tutorial at the May 2024 Embedded Vision Summit. In most AI research today, deep neural networks (DNNs) are designed solely to improve prediction accuracy, often ignoring… “Leveraging Neural Architecture Search for

Read More »

“Introduction to Visual Simultaneous Localization and Mapping (VSLAM),” a Presentation from Cadence

Amol Borkar, Product Marketing Director, and Shrinivas Gadkari, Design Engineering Director, both of Cadence, co-present the “Introduction to Visual Simultaneous Localization and Mapping (VSLAM)” tutorial at the May 2024 Embedded Vision Summit. Simultaneous localization and mapping (SLAM) is widely used in industry and has numerous applications where camera or ego-motion… “Introduction to Visual Simultaneous Localization

Read More »

Scalable Public Safety with On-device AI: How Startup FocusAI is Filling Enterprise Security Market Gaps

This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm Enterprise security is not just big business, it’s about keeping you safe: Here’s how engineer-turned-CTO Sudhakaran Ram collaborated with us to do just that. Key Takeaways: On-device AI enables superior enterprise-grade security. Distributed computing cost-efficiently enables actionable

Read More »

EyePop.ai Demonstration of Effortless AI Integration

Andy Ballester, Co-Founder of EyePop.ai, demonstrates the company’s latest edge AI and vision technologies and products at the 2024 Embedded Vision Summit. Specifically, Ballester demonstrates the ease and accessibility of his company’s AI platform. Tailored for startups without machine learning teams, this demo showcases how to create a custom computer vision endpoint that can be

Read More »

Accelerating Transformer Neural Networks for Autonomous Driving

This blog post was originally published at Ambarella’s website. It is reprinted here with the permission of Ambarella. Autonomous driving (AD) and advanced driver assistance system (ADAS) providers are deploying more and more AI neural networks (NNs) to offer human-like driving experience. Several of the leading AD innovators have either deployed, or have a roadmap

Read More »

Here you’ll find a wealth of practical technical insights and expert advice to help you bring AI and visual intelligence into your products without flying blind.

Contact

Address

Berkeley Design Technology, Inc.
PO Box #4446
Walnut Creek, CA 94596

Phone
Phone: +1 (925) 954-1411
Scroll to Top