Software for Embedded Vision

3LC: What is It and Who is It For?
This blog post was originally published at 3LC’s website. It is reprinted here with the permission of 3LC. AI performance isn’t just about better architectures or more compute – it’s about better data. Even perfectly labeled datasets can hold hidden inefficiencies that limit accuracy. See how teams use 3LC to refine datasets, optimize labeling strategies,

How e-con Systems’ TintE ISP IP Core Increases the Efficiency of Embedded Vision Applications
This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. e-con Systems has developed TintE™, a ready to deploy ISP IP core engineered to enhance image quality in camera systems. Built to deliver high performance on leading FPGA platforms, it accelerates real-time image processing with

Vision Language Model Prompt Engineering Guide for Image and Video Understanding
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Vision language models (VLMs) are evolving at a breakneck speed. In 2020, the first VLMs revolutionized the generative AI landscape by bringing visual understanding to large language models (LLMs) through the use of a vision encoder. These

Fine-tuning LLMs for Cost-effective GenAI Inference at Scale
This article was originally published at Tryolabs’ website. It is reprinted here with the permission of Tryolabs. Data is the new oil, fueling the AI revolution. From user-tailored shopping assistants to AI researchers, to recreating the King, the applicability of AI models knows no bounds. Yet these models are only as good as the data

SAM 2 + GPT-4o: Cascading Foundation Models via Visual Prompting (Part 2)
This article was originally published at Tenyks’ website. It is reprinted here with the permission of Tenyks. In Part 2 of our Segment Anything Model 2 (SAM 2) Series, we show how foundation models (e.g., GPT-4o, Claude Sonnet 3.5 and YOLO-World) can be used to generate visual inputs (e.g., bounding boxes) for SAM 2. Learn

Taming LLMs: Strategies and Tools for Controlling Responses
This article was originally published at Tryolabs’ website. It is reprinted here with the permission of Tryolabs. In the ever-evolving landscape of natural language processing, the advent of Large Language Models (LLMs) has ushered in a new era of possibilities and challenges. While these models showcase remarkable capabilities in generating human-like text, the potential for

New AI Model Offers Cellular-level View of Cancerous Tumors
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Researchers studying cancer unveiled a new AI model that provides cellular-level mapping and visualizations of cancer cells, which scientists hope can shed light on how—and why—certain inter-cellular relationships triggers cancers to grow. BioTuring, a San Diego-based startup,

D3 Embedded Partners with Silicon Highway to Provide Rugged Camera Solutions to Europe
Rochester, NY – February 12, 2025 – D3 Embedded today announced its partnership with Silicon Highway, a leading European distribution company specializing in embedded AI edge solutions, to accelerate the delivery of high-performance rugged cameras to the European market. This partnership will allow D3 Embedded to leverage Silicon Highway’s local expertise and knowledge of the

The Intersection of AI and Human Expertise: How Custom Solutions Enhance Collaboration
This blog post was originally published at Digica’s website. It is reprinted here with the permission of Digica. Artificial Intelligence-based solutions have become increasingly prevalent, transforming industries, businesses, and daily life. However, rather than completely replacing human expertise, the most effective approach lies in creating a synergy between human knowledge, experience and intuition alongside AI’s

SAM 2 + GPT-4o: Cascading Foundation Models via Visual Prompting (Part 1)
This article was originally published at Tenyks’ website. It is reprinted here with the permission of Tenyks. In Part 1 of this article we introduce Segment Anything Model 2 (SAM 2). Then, we walk you through how you can set it up and run inference on your own video clips. Learn more about visual prompting

DeepSeek-R1 1.5B on SiMa.ai for Less Than 10 Watts
February 18, 2025 09:00 AM Eastern Standard Time–SAN JOSE, Calif.–(BUSINESS WIRE)–SiMa.ai, the software-centric, embedded edge machine learning system-on-chip (MLSoC) company, today announced the successful implementation of DeepSeek-R1-Distill-Qwen-1.5B on its ONE Platform for Edge AI, achieving breakthrough performance within an unprecedented power envelope of under 10 watts. This implementation marks a significant advancement in efficient, secure

From Brain to Binary: Can Neuro-inspired Research Make CPUs the Future of AI Inference?
This article was originally published at Tryolabs’ website. It is reprinted here with the permission of Tryolabs. In the ever-evolving landscape of AI, the demand for powerful Large Language Models (LLMs) has surged. This has led to an unrelenting thirst for GPUs and a shortage that causes headaches for many organizations. But what if there

From Seeing to Understanding: LLMs Leveraging Computer Vision
This blog post was originally published at Tryolabs’ website. It is reprinted here with the permission of Tryolabs. From Face ID unlocking our phones to counting customers in stores, Computer Vision has already transformed how businesses operate. As Generative AI (GenAI) becomes more compelling and accessible, this tried-and-tested technology is entering a new era of

Autonomous Cars are Leveling Up: Exploring Vehicle Autonomy
When the Society of Automotive Engineers released their definitions of varying levels of automation from level 0 to level 5, it became easier to define and distinguish between the many capabilities and advancements of autonomous vehicles. Level 0 describes an older model of vehicle with no automated features, while level 5 describes a future ideal

Introducing Qualcomm Custom-built AI Models, Now Available on Qualcomm AI Hub
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. We’re thrilled to announce that five custom-built computer vision (CV) models are now available on Qualcomm AI Hub! Qualcomm Technologies’ custom-built models were developed by the Qualcomm R&D team, optimized for our platforms and designed with end-user applications

New AI SDKs and Tools Released for NVIDIA Blackwell GeForce RTX 50 Series GPUs
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. NVIDIA recently announced a new generation of PC GPUs—the GeForce RTX 50 Series—alongside new AI-powered SDKs and tools for developers. Powered by the NVIDIA Blackwell architecture, fifth-generation Tensor Cores and fourth-generation RT Cores, the GeForce RTX 50

10 Common Mistakes in Data Science Projects and How to Avoid Them
This blog post was originally published at Digica’s website. It is reprinted here with the permission of Digica. While Data Science is an established and highly esteemed profession with strong foundations in science, it is important to remember that it is still a craft and, as such, it is susceptible to errors coming from processes

RAG for Vision: Building Multimodal Computer Vision Systems
This blog post was originally published at Tenyks’ website. It is reprinted here with the permission of Tenyks. This article explores the exciting world of Visual RAG, exploring its significance and how it’s revolutionizing traditional computer vision pipelines. From understanding the basics of RAG to its specific applications in visual tasks and surveillance, we’ll examine