Processors for Embedded Vision
THIS TECHNOLOGY CATEGORY INCLUDES ANY DEVICE THAT EXECUTES VISION ALGORITHMS OR VISION SYSTEM CONTROL SOFTWARE
This technology category includes any device that executes vision algorithms or vision system control software. The following diagram shows a typical computer vision pipeline; processors are often optimized for the compute-intensive portions of the software workload.
The following examples represent distinctly different types of processor architectures for embedded vision, and each has advantages and trade-offs that depend on the workload. For this reason, many devices combine multiple processor types into a heterogeneous computing environment, often integrated into a single semiconductor component. In addition, a processor can be accelerated by dedicated hardware that improves performance on computer vision algorithms.
General-purpose CPUs
While computer vision algorithms can run on most general-purpose CPUs, desktop processors may not meet the design constraints of some systems. However, x86 processors and system boards can leverage the PC infrastructure for low-cost hardware and broadly-supported software development tools. Several Alliance Member companies also offer devices that integrate a RISC CPU core. A general-purpose CPU is best suited for heuristics, complex decision-making, network access, user interface, storage management, and overall control. A general purpose CPU may be paired with a vision-specialized device for better performance on pixel-level processing.
Graphics Processing Units
High-performance GPUs deliver massive amounts of parallel computing potential, and graphics processors can be used to accelerate the portions of the computer vision pipeline that perform parallel processing on pixel data. While General Purpose GPUs (GPGPUs) have primarily been used for high-performance computing (HPC), even mobile graphics processors and integrated graphics cores are gaining GPGPU capability—meeting the power constraints for a wider range of vision applications. In designs that require 3D processing in addition to embedded vision, a GPU will already be part of the system and can be used to assist a general-purpose CPU with many computer vision algorithms. Many examples exist of x86-based embedded systems with discrete GPGPUs.
Digital Signal Processors
DSPs are very efficient for processing streaming data, since the bus and memory architecture are optimized to process high-speed data as it traverses the system. This architecture makes DSPs an excellent solution for processing image pixel data as it streams from a sensor source. Many DSPs for vision have been enhanced with coprocessors that are optimized for processing video inputs and accelerating computer vision algorithms. The specialized nature of DSPs makes these devices inefficient for processing general-purpose software workloads, so DSPs are usually paired with a RISC processor to create a heterogeneous computing environment that offers the best of both worlds.
Field Programmable Gate Arrays (FPGAs)
Instead of incurring the high cost and long lead-times for a custom ASIC to accelerate computer vision systems, designers can implement an FPGA to offer a reprogrammable solution for hardware acceleration. With millions of programmable gates, hundreds of I/O pins, and compute performance in the trillions of multiply-accumulates/sec (tera-MACs), high-end FPGAs offer the potential for highest performance in a vision system. Unlike a CPU, which has to time-slice or multi-thread tasks as they compete for compute resources, an FPGA has the advantage of being able to simultaneously accelerate multiple portions of a computer vision pipeline. Since the parallel nature of FPGAs offers so much advantage for accelerating computer vision, many of the algorithms are available as optimized libraries from semiconductor vendors. These computer vision libraries also include preconfigured interface blocks for connecting to other vision devices, such as IP cameras.
Vision-Specific Processors and Cores
Application-specific standard products (ASSPs) are specialized, highly integrated chips tailored for specific applications or application sets. ASSPs may incorporate a CPU, or use a separate CPU chip. By virtue of their specialization, ASSPs for vision processing typically deliver superior cost- and energy-efficiency compared with other types of processing solutions. Among other techniques, ASSPs deliver this efficiency through the use of specialized coprocessors and accelerators. And, because ASSPs are by definition focused on a specific application, they are usually provided with extensive associated software. This same specialization, however, means that an ASSP designed for vision is typically not suitable for other applications. ASSPs’ unique architectures can also make programming them more difficult than with other kinds of processors; some ASSPs are not user-programmable.
BrainChip Awarded Air Force Research Laboratory Radar Development Contract
Laguna Hills, Calif. – DECEMBER 9, 2024 – BrainChip Holdings Ltd (ASX: BRN, OTCQX: BRCHF, ADR: BCHPY), the world’s first commercial producer of ultra-low power, fully digital, event-based, neuromorphic AI, today announced that it was awarded a development contract for $1.8M from Air Force Research Laboratory (AFRL) on neuromorphic radar signaling processing technologies. The AFRL contract
How RTX AI PCs Unlock AI Agents That Solve Complex Problems Autonomously With Generative AI
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. NVIDIA RTX-accelerated AnythingLLM launches Community Hub for sharing prompts, slash commands and AI agent skills. Editor’s note: This post is part of the AI Decoded series, which demystifies AI by making the technology more accessible, and showcases
Ceva-NeuPro-Nano Wins Product of the Year Award at Prestigious EE Awards Asia Event
Extending its Smart Edge IP leadership, Ceva-NeuPro-Nano™ received the Best IP/Processor of the year Ceva-NeuPro-Nano NPUs deliver optimal balance of ultra-low power and best performance in small area to efficiently execute embedded AI workloads in consumer, industrial and general-purpose AIoT products ROCKVILLE, MD., December 5, 2024 – Ceva, Inc. (NASDAQ: CEVA), the leading licensor of
LG and Ambarella Join Forces to Advance AI-driven In-cabin Vehicle Safety Solutions
LG Sets New Standard in Accuracy and Reliability for In-Cabin Solutions with Ambarella-Powered Driver Monitoring System; Plans Demo During CES 2025 SEOUL, Korea and SANTA CLARA, Calif., Dec. 4, 2024 — LG Electronics (LG), a mobility sector technology leader, and Ambarella, Inc. (NASDAQ: AMBA), an edge AI semiconductor company, today announced that LG will showcase
Trust What You See: How Truepic Authenticates Images and Videos In the Age of Deepfakes
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. Truepic is leading the charge to make sure people know the origin of what they’re seeing online Key Takeaways: The ability to trust what you see when it comes to digital content is more critical than ever.
NVIDIA Advances Robot Learning and Humanoid Development With New AI and Simulation Tools
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. New Project GR00T workflows and AI world model development technologies to accelerate robot dexterity, control, manipulation and mobility. Robotics developers can greatly accelerate their work on AI-enabled robots, including humanoids, using new AI and simulation tools and
How to Optimize and Accelerate Machine Vision Processing on GPUs
This blog post was originally published at Geisel Software’s website. It is reprinted here with the permission of Geisel Software. How to optimize and accelerate GPUS: tools, techniques, and real-world scenarios Machine learning (ML) systems analyze tremendous amounts of data to identify hidden patterns and make predictions based on those patterns. This requires a very
Join e-con Systems’ Exclusive Webinar In Collaboration With Advantech
e-con Systems is excited to invite you to an exclusive joint webinar on “Unlocking AI at the Edge: How Embedded Vision Is Helping Take Giant Strides” in partnership with Advantech. Learn how embedded vision is transforming real-time decision-making across industries. Dec 17, 2024, Tuesday – 11:00 AM – 12:00 PM (CET) or Dec 18th, 2024,
Qualcomm Technologies’ IoT Strategy: A New Approach, a New Opportunity
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. Our new blueprint for enabling our partners and end customers to bring more smarts to the edge was one of the highlights of Investor Day Everything around us is either already a lot smarter or aiming to
‘A Glimpse Into the Future of the Car’: Takeaways from Day 2 of Snapdragon Summit 2024
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. The Qualcomm Oryon CPU and the Snapdragon Elite name make its way to the automotive space, signaling its importance to Qualcomm Snapdragon Summit 2024 shifted gears on Tuesday and moved from the mobile world to the automotive
How NVIDIA Jetson AGX Orin Helps Unlock the Power of Surround-view Camera Solutions
This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. Autonomous vehicles, such as warehouse robots, rely on precise maneuvering. NVIDIA Jetson AGX Orin™-powered surround-view cameras provide a perfectly synchronized solution, allowing these robots to move freely within designated areas without requiring intensive manual intervention.
Chiplets: Revolutionizing Semiconductor Design and Manufacturing
In the rapidly evolving world of semiconductors, chiplet technology is emerging as a groundbreaking approach that addresses many of the challenges faced by traditional monolithic System-on-Chip (SoC) designs. As Moore’s Law slows down, the semiconductor industry is seeking innovative solutions to increase performance and functionality without merely increasing transistor density, as detailed in the IDTechEx
Streamlining AI Inference Performance and Deployment with NVIDIA TensorRT-LLM Chunked Prefill
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. In this blog post, we take a closer look at chunked prefill, a feature of NVIDIA TensorRT-LLM that increases GPU utilization and simplifies the deployment experience for developers. This builds on our previous post discussing how advanced
Microchip to Accelerate Real-time Edge AI with NVIDIA Holoscan
PolarFire® FPGA Ethernet Sensor Bridge provides low-power multi-sensor bridging to NVIDIA edge AI platforms CHANDLER, Ariz., November 14, 2024 — To enable developers building artificial intelligence (AI)-driven sensor processing systems, Microchip Technology (Nasdaq: MCHP) has released its PolarFire® FPGA Ethernet Sensor Bridge that works with the NVIDIA Holoscan sensor processing platform. PolarFire FPGAs enable multi-protocol
Snapdragon Summit’s AI Highlights: A Look at the Future of On-device AI
This blog post was originally published at Qualcomm’s website. It is reprinted here with the permission of Qualcomm. Qualcomm Technologies sets new standards in AI performance for its latest mobile, automotive and Qualcomm AI Hub advancements Our annual Snapdragon Summit wrapped up with exciting new announcements centered on the future of on-device artificial intelligence (AI).
Electrification and Autonomy: A Semiconductor Content Boost to $1,000 per Car by 2029
This market research report was originally published at the Yole Group’s website. It is reprinted here with the permission of the Yole Group. Automotive OEMs are moving upstream in the semiconductor supply chain with diverse strategies. OUTLINE The automotive semiconductor market is expecting a significant CAGR of 11% between 2023 and 2029 to almost $100
Advanced Semiconductor Packaging: Key Materials and Processing Trends
As semiconductor packaging technologies evolve, advanced methods like 2.5D and 3D Cu-to-Cu hybrid bonding are essential for achieving higher performance and power efficiency. However, manufacturing these technologies to meet high performance and yield standards while fulfilling client requirements is complex. Challenges include developing the right materials and innovating packaging manufacturing techniques. IDTechEx‘s “Materials and Processing
How to Accelerate Larger LLMs Locally on RTX With LM Studio
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. GPU offloading makes massive models accessible on local RTX AI PCs and workstations. Editor’s note: This post is part of the AI Decoded series, which demystifies AI by making the technology more accessible, and showcases new hardware,
Introducing the First AMD 1B Language Models: AMD OLMo
This blog post was originally published at AMD’s website. It is reprinted here with the permission of AMD. In recent years, the rapid development of artificial intelligence technology, especially the progress in large language models (LLMs), has garnered significant attention and discussion. From the emergence of ChatGPT to subsequent models like GPT-4 and Llama, these
MIPS Releases P8700, Industry’s First High-performance AI-enabled RISC-V Automotive CPU for ADAS and Autonomous Vehicles
SAN JOSE, CA – October 30, 2024 – MIPS, a leading developer of efficient and configurable IP compute cores, announced today the general availability(GA) launch of the MIPS P8700 Series RISC-V Processor. Designed to meet the low-latency, highly intensive data movement demands of the most advanced automotive applications such as ADAS and Autonomous Vehicles (AVs),