Technologies
scroll to learn more or view by subtopic
The listing below showcases the most recently published content associated with various AI and visual intelligence functions.
View all Posts

3LC: What is It and Who is It For?
This blog post was originally published at 3LC’s website. It is reprinted here with the permission of 3LC. AI performance isn’t just about better architectures or more compute – it’s about better data. Even perfectly labeled datasets can hold hidden inefficiencies that limit accuracy. See how teams use 3LC to refine datasets, optimize labeling strategies,

Vision Language Model Prompt Engineering Guide for Image and Video Understanding
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Vision language models (VLMs) are evolving at a breakneck speed. In 2020, the first VLMs revolutionized the generative AI landscape by bringing visual understanding to large language models (LLMs) through the use of a vision encoder. These

SAM 2 + GPT-4o: Cascading Foundation Models via Visual Prompting (Part 2)
This article was originally published at Tenyks’ website. It is reprinted here with the permission of Tenyks. In Part 2 of our Segment Anything Model 2 (SAM 2) Series, we show how foundation models (e.g., GPT-4o, Claude Sonnet 3.5 and YOLO-World) can be used to generate visual inputs (e.g., bounding boxes) for SAM 2. Learn

Nearly $1B Flows into Automotive Radar Startups
According to IDTechEx‘s latest report, “Automotive Radar Market 2025-2045: Robotaxis & Autonomous Cars“, newly established radar startups worldwide have raised nearly US$1.2 billion over the past 12 years; approximately US$980 million of which is predominantly directed toward the automotive sector. Through more than 40 funding rounds, these companies have driven the implementation and advancement of

New AI Model Offers Cellular-level View of Cancerous Tumors
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Researchers studying cancer unveiled a new AI model that provides cellular-level mapping and visualizations of cancer cells, which scientists hope can shed light on how—and why—certain inter-cellular relationships triggers cancers to grow. BioTuring, a San Diego-based startup,

Top-tier ADAS Systems: Exploring Automotive Radar Technology
Radars have had a place within the automotive sector for over two decades, beginning with the first use for adaptive cruise control and many other developments taking place since. IDTechEx‘s “Automotive Radar Market 2025-2045: Robotaxis & Autonomous Cars” report explores the latest developments in radar technology within the automotive sector. ADAS safety systems ADAS (advanced

SAM 2 + GPT-4o: Cascading Foundation Models via Visual Prompting (Part 1)
This article was originally published at Tenyks’ website. It is reprinted here with the permission of Tenyks. In Part 1 of this article we introduce Segment Anything Model 2 (SAM 2). Then, we walk you through how you can set it up and run inference on your own video clips. Learn more about visual prompting

Empowering Civil Construction with AI-driven Spatial Perception
This blog post was originally published at Au-Zone Technologies’ website. It is reprinted here with the permission of Au-Zone Technologies. Transforming Safety, Efficiency, and Automation in Construction Ecosystems In the rapidly evolving field of civil construction, AI-based spatial perception technologies are reshaping the way machinery operates in dynamic and unpredictable environments. These systems enable advanced

Seeing In the Dark: Infrared for Automotive
Infrared sensors are becoming more popular in vehicles for advanced driver assistance systems (ADAS), in-cabin sensing, and driver monitoring systems (DMS), largely due to advancements in vehicle safety and awareness. IDTechEx‘s latest report, “Infrared (IR) Cameras for Automotive 2025-2035: Technologies, Opportunities, Forecasts“, explores the types of infrared sensors that are commonly used, and forecasts for

Multimodal Large Language Models: Transforming Computer Vision
This blog post was originally published at Tenyks’ website. It is reprinted here with the permission of Tenyks. This article introduces multimodal large language models (MLLMs) [1], their applications using challenging prompts, and the top models reshaping computer vision as we speak. What is a multimodal large language model (MLLM)? In layman terms, a multimodal