Resources
In-depth information about the edge AI and vision applications, technologies, products, markets and trends.
The content in this section of the website comes from Edge AI and Vision Alliance members and other industry luminaries.
All Resources

Breaking the Human Accuracy Barrier in Computer Vision Labeling
This article was originally published at 3LC’s website. It is reprinted here with the permission of 3LC. Introduction There’s been a lot of excitement lately around how foundation models (such as CLIP, SAM, Grounding DINO, etc.)

NVIDIA and Synopsys Announce Strategic Partnership to Revolutionize Engineering and Design
Key Highlights Multiyear collaboration spans NVIDIA CUDA accelerated computing, agentic and physical AI, and Omniverse digital twins to achieve simulation speed and scale previously unattainable through traditional CPU computing — opening new market opportunities across

Now Available: AMD Spartan UltraScale+ FPGA SCU35 Evaluation Kit–An Affordable Platform for Every Developer
November 25, 2025 The AMD Spartan™ UltraScale+™ FPGA SCU35 Evaluation Kit is now available for order. Built by AMD, this platform offers customers an accelerated path to production with Spartan UltraScale+ FPGAs. The kit features

The 8-Series Reimagined: Snapdragon 8 Gen 5 Delivers Premium Performance and Experiences
San Diego, California November 25, 2025 — Qualcomm Technologies, Inc. today announced the Snapdragon® 8 Gen 5 Mobile Platform, a premium offering that combines outstanding performance with cutting-edge technologies, raising the bar for flagship mobile

Microcontrollers enter a new growth cycle as the market targets US$34 billion in 2030
This market research report was originally published at the Yole Group’s website. It is reprinted here with the permission of the Yole Group. Yole Group releases its annual Status of the Microcontroller Industry report and expands

Why Edge AI Struggles Towards Production: The Deployment Problem
There is no shortage of articles about how to develop and train Edge AI models. The community has also written extensively about why it makes sense to run those models at the edge: to reduce

Nota AI Signs Technology Collaboration Agreement with Samsung Electronics for Exynos AI Optimization “Driving the Popularization of On-Device Generative AI”
Nota AI’s optimization technology integrated into Samsung Electronics’ Exynos AI Studio, enhancing efficiency in on-device AI model development Seoul, South Korea Nov.26, 2025 — Nota AI, a company specializing in AI model compression and

WAVE-N v2: Chips&Media’s Custom NPU Retains 16-bit FP for Superior Efficiency at High TOPS
Nov 26, 2025 Seoul, South Korea — Chips&Media is announcing that the next-generation customized NPU, WAVE-N v2, is ready for release, delivering higher computational power along with improved area and power efficiency. WAVE-N v2 is also

Google Announces LiteRT Qualcomm AI Engine Direct Accelerator
Google has announced a new LiteRT Qualcomm AI Engine Direct Accelerator, giving Android and embedded developers a much more direct path to Qualcomm NPUs for on-device AI and vision workloads. Built on top of Qualcomm’s
Technologies

What is a dust denoising filter in TOF camera, and how does it remove noise artifacts in vision systems?
This article was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. Time-of-Flight (ToF) cameras with IR sensors are susceptible to performance variations caused by environmental dust. This dust can create ‘dust noise’ in the output depth map, directly impacting camera accuracy and, consequently, the reliability of critical

Low-Light Image Enhancement: YUV vs RAW – What’s the Difference?
This blog post was originally published at Visidon’s website. It is reprinted here with the permission of Visidon. In the world of embedded vision—whether for mobile phones, surveillance systems, or smart edge devices—image quality in low-light conditions can make or break user experience. That’s where advanced AI-based denoising algorithms come into play. At our company, we

NVIDIA-Accelerated Mistral 3 Open Models Deliver Efficiency, Accuracy at Any Scale
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. The new Mistral 3 open model family delivers industry-leading accuracy, efficiency, and customization capabilities for developers and enterprises. Optimized from NVIDIA GB200 NVL72 to edge platforms, Mistral 3 includes: One large state-of-the-art sparse multimodal and multilingual mixture of
Applications

Ambarella’s CV3-AD655 Surround View with IMG BXM GPU: A Case Study
The CV3-AD family block diagram. This blog post was originally published at Imagination Technologies’ website. It is reprinted here with the permission of Imagination Technologies. Ambarella’s CV3-AD655 autonomous driving AI domain controller pairs energy-efficient compute with Imagination’s IMG BXM GPU to enable real-time surround-view visualisation for L2++/L3 vehicles. This case study outlines the industry shift

Overcoming the Skies: Navigating the Challenges of Drone Autonomy
This blog post was originally published at Inuitive’s website. It is reprinted here with the permission of Inuitive. From early military prototypes to today’s complex commercial operations, drones have evolved from experimental aircraft into essential tools across industries. Since the FAA issued its first commercial permit in 2006, applications have rapidly expanded—from disaster relief and

NVIDIA Advances Open Model Development for Digital and Physical AI
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. NVIDIA releases new AI tools for speech, safety and autonomous driving — including NVIDIA DRIVE Alpamayo-R1, the world’s first open industry-scale reasoning vision language action model for mobility — and a new independent benchmark recognizes the openness and
Functions

SAM3: A New Era for Open‑Vocabulary Segmentation and Edge AI
Quality training data – especially segmented visual data – is a cornerstone of building robust vision models. Meta’s recently announced Segment Anything Model 3 (SAM3) arrives as a potential game-changer in this domain. SAM3 is a unified model that can detect, segment, and even track objects in images and videos using both text and visual

TLens vs VCM Autofocus Technology
This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. In this blog, we’ll walk you through how TLens technology differs from traditional VCM autofocus, how TLens combined with e-con Systems’ Tinte ISP enhances camera performance, key advantages of TLens over mechanical autofocus systems, and applications

“Object Detection Models: Balancing Speed, Accuracy and Efficiency,” a Presentation from Union.ai
Sage Elliott, AI Engineer at Union.ai, presents the “Object Detection Models: Balancing Speed, Accuracy and Efficiency,” tutorial at the May 2025 Embedded Vision Summit. Deep learning has transformed many aspects of computer vision, including object detection, enabling accurate and efficient identification of objects in images and videos. However, choosing the… “Object Detection Models: Balancing Speed,
