Energy News
ROBO SPACE
Teaching robots to map large environments

Teaching robots to map large environments

by Adam Zewe for MIT News
Boston MA (SPX) Nov 06, 2025

A new approach developed at MIT could help a search-and-rescue robot navigate an unpredictable environment by rapidly generating an accurate map of its surroundings.

The artificial intelligence-driven system incrementally creates and aligns smaller submaps of the scene, which it stitches together to reconstruct a full 3D map, like of an office cubicle, while estimating the robot's position in real-time.

A robot searching for workers trapped in a partially collapsed mine shaft must rapidly generate a map of the scene and identify its location within that scene as it navigates the treacherous terrain.

Researchers have recently started building powerful machine-learning models to perform this complex task using only images from the robot's onboard cameras, but even the best models can only process a few images at a time. In a real-world disaster where every second counts, a search-and-rescue robot would need to quickly traverse large areas and process thousands of images to complete its mission.

To overcome this problem, MIT researchers drew on ideas from both recent artificial intelligence vision models and classical computer vision to develop a new system that can process an arbitrary number of images. Their system accurately generates 3D maps of complicated scenes like a crowded office corridor in a matter of seconds.

The AI-driven system incrementally creates and aligns smaller submaps of the scene, which it stitches together to reconstruct a full 3D map while estimating the robot's position in real-time.

Unlike many other approaches, their technique does not require calibrated cameras or an expert to tune a complex system implementation. The simpler nature of their approach, coupled with the speed and quality of the 3D reconstructions, would make it easier to scale up for real-world applications.

Beyond helping search-and-rescue robots navigate, this method could be used to make extended reality applications for wearable devices like VR headsets or enable industrial robots to quickly find and move goods inside a warehouse.

"For robots to accomplish increasingly complex tasks, they need much more complex map representations of the world around them. But at the same time, we don't want to make it harder to implement these maps in practice. We've shown that it is possible to generate an accurate 3D reconstruction in a matter of seconds with a tool that works out of the box," says Dominic Maggio, an MIT graduate student and lead author of a paper on this method.

Maggio is joined on the paper by postdoc Hyungtae Lim and senior author Luca Carlone, associate professor in MIT's Department of Aeronautics and Astronautics (AeroAstro), principal investigator in the Laboratory for Information and Decision Systems (LIDS), and director of the MIT SPARK Laboratory. The research will be presented at the Conference on Neural Information Processing Systems.

Mapping out a solution

For years, researchers have been grappling with an essential element of robotic navigation called simultaneous localization and mapping (SLAM). In SLAM, a robot recreates a map of its environment while orienting itself within the space.

Traditional optimization methods for this task tend to fail in challenging scenes, or they require the robot's onboard cameras to be calibrated beforehand. To avoid these pitfalls, researchers train machine-learning models to learn this task from data.

While they are simpler to implement, even the best models can only process about 60 camera images at a time, making them infeasible for applications where a robot needs to move quickly through a varied environment while processing thousands of images.

To solve this problem, the MIT researchers designed a system that generates smaller submaps of the scene instead of the entire map. Their method "glues" these submaps together into one overall 3D reconstruction. The model is still only processing a few images at a time, but the system can recreate larger scenes much faster by stitching smaller submaps together.

"This seemed like a very simple solution, but when I first tried it, I was surprised that it didn't work that well," Maggio says.

Searching for an explanation, he dug into computer vision research papers from the 1980s and 1990s. Through this analysis, Maggio realized that errors in the way the machine-learning models process images made aligning submaps a more complex problem.

Traditional methods align submaps by applying rotations and translations until they line up. But these new models can introduce some ambiguity into the submaps, which makes them harder to align. For instance, a 3D submap of a one side of a room might have walls that are slightly bent or stretched. Simply rotating and translating these deformed submaps to align them doesn't work.

"We need to make sure all the submaps are deformed in a consistent way so we can align them well with each other," Carlone explains.

A more flexible approach

Borrowing ideas from classical computer vision, the researchers developed a more flexible, mathematical technique that can represent all the deformations in these submaps. By applying mathematical transformations to each submap, this more flexible method can align them in a way that addresses the ambiguity.

Based on input images, the system outputs a 3D reconstruction of the scene and estimates of the camera locations, which the robot would use to localize itself in the space.

"Once Dominic had the intuition to bridge these two worlds - learning-based approaches and traditional optimization methods - the implementation was fairly straightforward," Carlone says. "Coming up with something this effective and simple has potential for a lot of applications.

Their system performed faster with less reconstruction error than other methods, without requiring special cameras or additional tools to process data. The researchers generated close-to-real-time 3D reconstructions of complex scenes like the inside of the MIT Chapel using only short videos captured on a cell phone.

The average error in these 3D reconstructions was less than 5 centimeters.

In the future, the researchers want to make their method more reliable for especially complicated scenes and work toward implementing it on real robots in challenging settings.

"Knowing about traditional geometry pays off. If you understand deeply what is going on in the model, you can get much better results and make things much more scalable," Carlone says.

This work is supported, in part, by the U.S. National Science Foundation, U.S. Office of Naval Research, and the National Research Foundation of Korea. Carlone, currently on sabbatical as an Amazon Scholar, completed this work before he joined Amazon.

Research Report:VGGT-SLAM: Dense RGB SLAM Optimized on the SL(4) Manifold

Related Links
Massachusetts Institute of Technology
All about the robots on Earth and beyond!

Subscribe Free To Our Daily Newsletters
Tweet

RELATED CONTENT
The following news reports may link to other Space Media Network websites.
ROBO SPACE
Robotic exosuit designed to assist astronaut movement tested in simulated lunar mission
London, UK (SPX) Nov 05, 2025
A soft robotic exosuit created at the University of Bristol is engineered to boost astronaut mobility and reduce muscle fatigue while allowing natural movement. The lightweight garment, which incorporates fabric-based artificial muscles, is designed to be worn underneath a standard spacesuit. The exosuit features two-layer artificial muscles, with an outer nylon component and an inner thermoplastic layer that inflates to provide motion support. Kevlar components are used for the waistband and knee ... read more

ROBO SPACE
Wits expands earth science with new observatory and CORES center

China increases lead in global remote sensing research as US share slips

OlmoEarth AI Platform Released to Expand Access to Planetary Data and Insights

Reflectivity of ocean clouds drops as air pollution falls and global temperatures climb

ROBO SPACE
Centimeter-level RTK positioning now available for IoT deployments

Nanometer precision ranging demonstrated across 113 kilometers sets new benchmark for space measurement

PntGuard delivers maritime resilience against navigation signal interference

Next-generation visual navigation startup Vermeer secures major funding milestone

ROBO SPACE
Ethiopia's invasive prosopis tree chokes livelihoods and land

Amazonian forests altered by human actions show broad changes in diversity and evolutionary patterns

Amazon poised to host toughest climate talks in years

World leaders launch fund to save forests, get first $5 bn

ROBO SPACE
Illinois team creates aviation fuel from food waste with circular economy benefits

Industrial microbe enables conversion of carbon monoxide to ethanol

Revolutionary microbe enables resilient renewable energy from food waste

Finnish carbon-neutral ferry aims to set global benchmark for shipping

ROBO SPACE
China emissions flat in third quarter as solar surges: study

PolyU team advances tandem solar cell efficiency and reliability targets

Enhanced solar water splitting achieved with MoS2 GaN nanorod heterostructures

Graphene solar cells promise long-lasting self-powered sensor networks

ROBO SPACE
S.Africa seeks to save birds from wind turbine risks

Vertical wind turbines may soon power UK railways using tunnel airflow

Danish wind giant Orsted to cut workforce by a quarter

French-German duo wins mega offshore wind energy project

ROBO SPACE
Earth cannot 'sustain' intensive fossil fuel use, Lula tells COP30

China's power paradox: record renewables, continued coal

US government aims to open more public lands to coal mining

China coal power surges even as renewables hit record high

ROBO SPACE
China's 'Singles Day' shopping fest loses its shine for weary consumers

Daughter of 'underground' pastor urges China for his release

Unruffled by Trump, Chinese parents chase 'American dream' for kids

China dreams of football glory at last... in gaming

Subscribe Free To Our Daily Newsletters




The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.