Robots Track Moving Objects With Unprecedented Precision
February 20, 2019 | MITEstimated reading time: 5 minutes

System uses RFID tags to home in on targets; could benefit robotic manufacturing, collaborative drones, and other applications.
A novel system developed at MIT uses RFID tags to help robots home in on moving objects with unprecedented speed and accuracy. The system could enable greater collaboration and precision by robots working on packaging and assembly, and by swarms of drones carrying out search-and-rescue missions.
In a paper being presented next week at the USENIX Symposium on Networked Systems Design and Implementation, the researchers show that robots using the system can locate tagged objects within 7.5 milliseconds, on average, and with an error of less than a centimeter.
In the system, called TurboTrack, an RFID (radio-frequency identification) tag can be applied to any object. A reader sends a wireless signal that reflects off the RFID tag and other nearby objects, and rebounds to the reader. An algorithm sifts through all the reflected signals to find the RFID tag’s response. Final computations then leverage the RFID tag’s movement — even though this usually decreases precision — to improve its localization accuracy.
The researchers say the system could replace computer vision for some robotic tasks. As with its human counterpart, computer vision is limited by what it can see, and it can fail to notice objects in cluttered environments. Radio frequency signals have no such restrictions: They can identify targets without visualization, within clutter and through walls.
To validate the system, the researchers attached one RFID tag to a cap and another to a bottle. A robotic arm located the cap and placed it onto the bottle, held by another robotic arm. In another demonstration, the researchers tracked RFID-equipped nanodrones during docking, maneuvering, and flying. In both tasks, the system was as accurate and fast as traditional computer-vision systems, while working in scenarios where computer vision fails, the researchers report.
“If you use RF signals for tasks typically done using computer vision, not only do you enable robots to do human things, but you can also enable them to do superhuman things,” says Fadel Adib, an assistant professor and principal investigator in the MIT Media Lab, and founding director of the Signal Kinetics Research Group. “And you can do it in a scalable way, because these RFID tags are only 3 cents each.”
In manufacturing, the system could enable robot arms to be more precise and versatile in, say, picking up, assembling, and packaging items along an assembly line. Another promising application is using handheld “nanodrones” for search and rescue missions. Nanodrones currently use computer vision and methods to stitch together captured images for localization purposes. These drones often get confused in chaotic areas, lose each other behind walls, and can’t uniquely identify each other. This all limits their ability to, say, spread out over an area and collaborate to search for a missing person. Using the researchers’ system, nanodrones in swarms could better locate each other, for greater control and collaboration.
“You could enable a swarm of nanodrones to form in certain ways, fly into cluttered environments, and even environments hidden from sight, with great precision,” says first author Zhihong Luo, a graduate student in the Signal Kinetics Research Group.
The other Media Lab co-authors on the paper are visiting student Qiping Zhang, postdoc Yunfei Ma, and Research Assistant Manish Singh.
Page 1 of 2
Suggested Items
Driving Innovation: Direct Imaging vs. Conventional Exposure
07/01/2025 | Simon Khesin -- Column: Driving InnovationMy first camera used Kodak film. I even experimented with developing photos in the bathroom, though I usually dropped the film off at a Kodak center and received the prints two weeks later, only to discover that some images were out of focus or poorly framed. Today, every smartphone contains a high-quality camera capable of producing stunning images instantly.
Specially Developed for Laser Plastic Welding from LPKF
06/25/2025 | LPKFLPKF introduces TherMoPro, a thermographic analysis system specifically developed for laser plastic welding that transforms thermal data into concrete actionable insights. Through automated capture, evaluation, and interpretation of surface temperature patterns immediately after welding, the system provides unprecedented process transparency that correlates with product joining quality and long-term product stability.
United Electronics Corporation Advances Manufacturing Capabilities with Schmoll MDI-ST Imaging Equipment
06/24/2025 | United Electronics CorporationUnited Electronics Corporation has successfully installed the advanced Schmoll MDI-ST (XL) imaging equipment at their advanced printed circuit board facility. This significant technology investment represents a continued commitment to delivering superior products and maintaining their position as an industry leader in precision PCB manufacturing.
IBM, RIKEN Unveil First IBM Quantum System Two Outside of the U.S.
06/24/2025 | IBMIBM and RIKEN, a national research laboratory in Japan, today unveiled the first IBM Quantum System Two ever to be deployed outside of the United States and beyond an IBM Quantum Data Center.
Excellon Installs COBRA Hybrid Laser at Innovative Circuits
06/23/2025 | ExcellonExcellon is pleased to announce the successful installation of a second COBRA Hybrid Laser System at Innovative Circuits, located in Alpharetta, Georgia. The Excellon COBRA Hybrid Laser System uniquely combines both UV and CO₂ (IR) laser sources on a single platform—making it ideal for high-density prototype and production printed circuit boards (PCBs).