Processors for Embedded Vision
THIS TECHNOLOGY CATEGORY INCLUDES ANY DEVICE THAT EXECUTES VISION ALGORITHMS OR VISION SYSTEM CONTROL SOFTWARE
This technology category includes any device that executes vision algorithms or vision system control software. The following diagram shows a typical computer vision pipeline; processors are often optimized for the compute-intensive portions of the software workload.
The following examples represent distinctly different types of processor architectures for embedded vision, and each has advantages and trade-offs that depend on the workload. For this reason, many devices combine multiple processor types into a heterogeneous computing environment, often integrated into a single semiconductor component. In addition, a processor can be accelerated by dedicated hardware that improves performance on computer vision algorithms.
General-purpose CPUs
While computer vision algorithms can run on most general-purpose CPUs, desktop processors may not meet the design constraints of some systems. However, x86 processors and system boards can leverage the PC infrastructure for low-cost hardware and broadly-supported software development tools. Several Alliance Member companies also offer devices that integrate a RISC CPU core. A general-purpose CPU is best suited for heuristics, complex decision-making, network access, user interface, storage management, and overall control. A general purpose CPU may be paired with a vision-specialized device for better performance on pixel-level processing.
Graphics Processing Units
High-performance GPUs deliver massive amounts of parallel computing potential, and graphics processors can be used to accelerate the portions of the computer vision pipeline that perform parallel processing on pixel data. While General Purpose GPUs (GPGPUs) have primarily been used for high-performance computing (HPC), even mobile graphics processors and integrated graphics cores are gaining GPGPU capability—meeting the power constraints for a wider range of vision applications. In designs that require 3D processing in addition to embedded vision, a GPU will already be part of the system and can be used to assist a general-purpose CPU with many computer vision algorithms. Many examples exist of x86-based embedded systems with discrete GPGPUs.
Digital Signal Processors
DSPs are very efficient for processing streaming data, since the bus and memory architecture are optimized to process high-speed data as it traverses the system. This architecture makes DSPs an excellent solution for processing image pixel data as it streams from a sensor source. Many DSPs for vision have been enhanced with coprocessors that are optimized for processing video inputs and accelerating computer vision algorithms. The specialized nature of DSPs makes these devices inefficient for processing general-purpose software workloads, so DSPs are usually paired with a RISC processor to create a heterogeneous computing environment that offers the best of both worlds.
Field Programmable Gate Arrays (FPGAs)
Instead of incurring the high cost and long lead-times for a custom ASIC to accelerate computer vision systems, designers can implement an FPGA to offer a reprogrammable solution for hardware acceleration. With millions of programmable gates, hundreds of I/O pins, and compute performance in the trillions of multiply-accumulates/sec (tera-MACs), high-end FPGAs offer the potential for highest performance in a vision system. Unlike a CPU, which has to time-slice or multi-thread tasks as they compete for compute resources, an FPGA has the advantage of being able to simultaneously accelerate multiple portions of a computer vision pipeline. Since the parallel nature of FPGAs offers so much advantage for accelerating computer vision, many of the algorithms are available as optimized libraries from semiconductor vendors. These computer vision libraries also include preconfigured interface blocks for connecting to other vision devices, such as IP cameras.
Vision-Specific Processors and Cores
Application-specific standard products (ASSPs) are specialized, highly integrated chips tailored for specific applications or application sets. ASSPs may incorporate a CPU, or use a separate CPU chip. By virtue of their specialization, ASSPs for vision processing typically deliver superior cost- and energy-efficiency compared with other types of processing solutions. Among other techniques, ASSPs deliver this efficiency through the use of specialized coprocessors and accelerators. And, because ASSPs are by definition focused on a specific application, they are usually provided with extensive associated software. This same specialization, however, means that an ASSP designed for vision is typically not suitable for other applications. ASSPs’ unique architectures can also make programming them more difficult than with other kinds of processors; some ASSPs are not user-programmable.
Major Investments In the Processor Industry Set to Boost Data Center Applications
This market research report was originally published at the Yole Group’s website. It is reprinted here with the permission of the Yole Group. Amidst booming demand for AI accelerators, Nvidia is set to dominate the generative AI revolution. OUTLINE: The processor market generated revenue of $220 billion in 2023 and is on track to reach
Flux and Furious: New Image Generation Model Runs Fastest on RTX AI PCs and Workstations
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Black Forest Labs’ latest models generate high-quality images and are highly performant on NVIDIA RTX GPUs. Editor’s note: This post is part of the AI Decoded series, which demystifies AI by making the technology more accessible, and
Microchip Expands 64-bit Portfolio with High-performance, Post-quantum Security-enabled PIC64HX Microprocessors
The RISC-V-based MPUs support mission-critical intelligent edge applications with TSN Ethernet switching and AI capabilities CHANDLER, Ariz., October 21, 2024 — The global edge computing market is expected to grow by more than 30 percent in the next five years, serving mission-critical applications in the aerospace, defense, military, industrial and medical sectors. To meet this increasing
Exploring the Next Frontier of AI: Multimodal Systems and Real-time Interaction
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. Discover the state of the art in large multimodal models with Qualcomm AI Research In the realm of artificial intelligence (AI), the integration of senses — seeing, hearing and interacting — represents a frontier that is rapidly
The Chiplet Market Will Hit $411B by 2035, Transforming the Semiconductor Industry
For more information, visit https://www.idtechex.com/en/research-report/chiplet-technology-2025/1041. In the rapidly evolving world of semiconductors, chiplet technology is emerging as a groundbreaking approach that addresses many of the challenges faced by traditional monolithic System-on-Chip (SoC) designs. As Moore’s Law slows down, the semiconductor industry is seeking innovative solutions to increase performance and functionality without merely increasing transistor density.
Smart Glasses for the Consumer Market
There are currently about 250 companies in the head mounted wearables category and these companies in aggregate have received over $5B in funding. $700M has been invested in this category just since the beginning of the year. On the M&A front, there have already been a number of significant acquisitions in the space, notably the
Orchestrating Innovation at Scale with NVIDIA Maxine and Texel
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. The NVIDIA Maxine AI developer platform is a suite of NVIDIA NIM microservices, cloud-accelerated microservices, and SDKs that offer state-of-the-art features for enhancing real-time video and audio. NVIDIA partners use Maxine features to create better virtual interaction
“Removing Weather-related Image Degradation at the Edge,” a Presentation from Rivian
Ramit Pahwa, Machine Learning Scientist at Rivian, presents the “Removing Weather-related Image Degradation at the Edge” tutorial at the May 2024 Embedded Vision Summit. For machines that operate outdoors—such as autonomous cars and trucks—image quality degradation due to weather conditions presents a significant challenge. For example, snow, rainfall and raindrops… “Removing Weather-related Image Degradation at
The Rise of AI Drives a Ninefold Surge in Liquid Cooling Technology
AI servers, driven by Nvidia’s GB200 superchip, have experienced significant growth. The cutting-edge B200 chip, due to its high thermal design power, requires direct-to-chip cooling. Supermicro announced that it had shipped over 2000 direct-liquid-cooled AI server racks by the end of August 2024, and it has expanded its manufacturing capacity to 5000 racks per month.
AMD Unveils Leadership AI Solutions at Advancing AI 2024
AMD launches 5th Gen AMD EPYC processors, AMD Instinct MI325X accelerators, next gen networking solutions and AMD Ryzen AI PRO processors powering enterprise AI at scale Dell, Google Cloud, HPE, Lenovo, Meta, Microsoft, Oracle Cloud Infrastructure, Supermicro and AI leaders Databricks, Essential AI, Fireworks AI, Luma AI and Reka AI joined AMD to showcase expanding
Qualcomm Introduces Industrial-grade IQ Series and IoT Solutions Framework to Usher in New Era of Industrial Intelligence
Highlights: Qualcomm introduces a new portfolio for industrial IoT for the AI era, leading the transition to connected intelligent end points and empowering developers and enterprises to build the next generation of edge AI solutions across industries. The Qualcomm IoT Solutions Framework aids enterprises in solving the most daunting business challenges by tailoring platforms that
Bringing Mixed Reality to the Masses with Meta Quest 3S Powered by Snapdragon
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. In a Nutshell: Snapdragon XR Platforms power the latest commercial augmented, virtual and mixed reality devices from Meta including the just announced Quest 3S. Qualcomm Technologies, Inc. and Meta continue their strong collaboration and are powering XR
Accelerating LLMs with llama.cpp on NVIDIA RTX Systems
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. The NVIDIA RTX AI for Windows PCs platform offers a thriving ecosystem of thousands of open-source models for application developers to leverage and integrate into Windows applications. Notably, llama.cpp is one popular tool, with over 65K GitHub
“Adventures in Moving a Computer Vision Solution from Cloud to Edge,” a Presentation from MetaConsumer
Nate D’Amico, CTO and Head of Product at MetaConsumer, presents the “Adventures in Moving a Computer Vision Solution from Cloud to Edge” tutorial at the May 2024 Embedded Vision Summit. Optix is a computer vision-based AI system that measures advertising and media exposures on mobile devices for real-time marketing optimization.… “Adventures in Moving a Computer
Transforming Interconnects in AI Systems: Co-Packaged Optics’ Role
In recent years, there has been a noticeable trend in optical transceiver technology, moving toward bringing the transceiver closer to the ASIC. Traditionally, pluggable optics—optical modules inserted and removed from the front panel of a switch—have been located near the edge of the printed circuit board (PCB). These pluggable optics are widely used in data
Redefining Hybrid Meetings With AI-powered 360° Videoconferencing
This blog post was originally published at Ambarella’s website. It is reprinted here with the permission of Ambarella. The global pandemic catalyzed a boom in videoconferencing that continues to grow as companies embrace hybrid work models and seek more sustainable approaches to business communication with less travel. Now, with videoconferencing becoming a cornerstone of modern
Qualcomm Partners with Meta to Support Llama 3.2. Why This is a Big Deal for On-device AI
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. On-device artificial intelligence (AI) is critical to making your everyday AI experiences fast and security-rich. That’s why it’s such a win that Qualcomm Technologies and Meta have worked together to support the Llama 3.2 large language models (LLMs)
“Introduction to Depth Sensing,” a Presentation from Meta
Harish Venkataraman, Depth Cameras Architecture and Tech Lead at Meta, presents the “Introduction to Depth Sensing” tutorial at the May 2024 Embedded Vision Summit. We live in a three-dimensional world, and the ability to perceive in three dimensions is essential for many systems. In this talk, Venkataraman introduced the main… “Introduction to Depth Sensing,” a
Synaptics Astra AI-native IoT Compute Platform Wins 2024 EDGE Award
SAN JOSE, Calif., Oct. 01, 2024 (GLOBE NEWSWIRE) — Synaptics® Incorporated (Nasdaq: SYNA) today announced that its Synaptics Astra™ AI-native IoT compute platform won in the Machine Learning and Deep Learning category of the 2024 EDGE Awards. The annual awards from Endeavor Media celebrate outstanding innovation in product design and function for the engineering industry.
Deploying Accelerated Llama 3.2 from the Edge to the Cloud
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Expanding the open-source Meta Llama collection of models, the Llama 3.2 collection includes vision language models (VLMs), small language models (SLMs), and an updated Llama Guard model with support for vision. When paired with the NVIDIA accelerated