Using AI to Better Understand the Ocean

This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA.

Humans know more about deep space than we know about Earth’s deepest oceans. But scientists have plans to change that—with the help of AI.

“We have better maps of Mars than we do of our own exclusive economic zone,” said Nick Rotker, chief BlueTech strategist at MITRE, a US government-sponsored nonprofit research organization. “Around 70% of the Earth is covered in water and we’ve explored virtually none of it.”

Speaking at GTC to a packed ballroom, which included Robert Ballard, the famed oceanographer and Titanic discoverer, Rotker and two colleagues outlined a plan for developing detailed maps of the Earth’s oceans, sketching out a collaborative strategy involving the scientific, academic, and enterprise communities.

For more information, see Exploring Earth’s Oceans: Using Digital Twins to Drive Digital Ocean Collaboration on NVIDIA On-Demand.

Rotker explained how MITRE’s efforts rest, at least in part, on pursuing a three-computer strategy to create high-fidelity digital twins of the world’s oceans.

The first computer is powered by the NVIDIA DGX platform. It uses vast amounts of atmospheric and ocean-related data to train models, including data about salinity levels, current patterns, ice-formation, and the ocean’s floor topography. The myriad multimodal sources for that data comes from LiDAR data, the SOFAR network of spotter buoys, satellites, as well as different atmospheric inputs.

The second computer is actually a fleet of autonomous robots designed to explore the ocean, which are integrated with the third computer, a collection of Virginia-based computers running NVIDIA Omniverse.

These three computers work together to train Vespa-sized, unmanned, underwater robots designed for autonomous observation.

MITRE’s Unmanned Underwater Autonomous Vehicle Digital Twin

Rotker’s team is training these robots in a MITRE BlueTech in Bedford, Massachusetts, where, tethered to communication lines, the autonomous robots share information they’re gathering back to the computers in Virginia.

Those computers, which are running NVIDIA Omniverse, create a digital twin of what the robot observes in real time and is used to help improve the robot’s underwater autonomy. The autonomous robots are packed with sensors, running edge compute with NVIDIA Jetson AGX Xavier cards.

When the robots are sufficiently trained to autonomously navigate zero-shot underwater situations, Rotker and his team will release them into the ocean, where they will gather more data that can provide new insights about the ocean.

Initial use cases for new ocean-related data include improving humans’ understanding of fishery dynamics, more accurate modeling of extreme weather events and how they could impact oceans and communities located near oceans, and identifying threats to coral reef health.

Figure 1. Training autonomous underwater vehicles in the MITRE BlueTech Lab

Long-term, MITRE plans to create integrated, autonomous platforms that can unlock an even deeper understanding of the world’s oceans.

“There are almost innumerable variables connected with the ocean which are driven by physics,” said Connor Lewellyn, MITRE’s chief environmental engineer. “We’re using new compute to better train weather models to drive a fuller predictive ocean model.

MITRE plans to collaborate with SOFAR and others to scale ocean exploration and advance a universal standard—a kind of USB port for underwater data—that researchers, academics, and businesses can plug into to better understand oceans.

Related resources

Elias Wolfberg
Executive Communications team, NVIDIA

Here you’ll find a wealth of practical technical insights and expert advice to help you bring AI and visual intelligence into your products without flying blind.

Contact

Address

Berkeley Design Technology, Inc.
PO Box #4446
Walnut Creek, CA 94596

Phone
Phone: +1 (925) 954-1411
Scroll to Top