Amplifying Small Motions in Large Motions
June 18, 2015 | MITEstimated reading time: 3 minutes

For several years now, the research groups of MIT professors of computer science and engineering William Freeman and Frédo Durand have been investigating techniques for amplifying movements captured by video but indiscernible to the human eye. Versions of their algorithms can make the human pulse visible and even recover intelligible speech from the vibrations of objects filmed through soundproof glass.
Earlier this month, at the Computer Vision and Pattern Recognition conference, Freeman, Durand, and colleagues at the Qatar Computing Research Institute (QCRI) presented a new version of the algorithm that can amplify small motions even when they’re contained within objects executing large motions. So, for instance, it could make visible the precise sequence of muscle contractions in the arms of a baseball player swinging the bat, or in the legs of a soccer player taking a corner kick.
“The previous version of the algorithm assumed everything was small in the video,” Durand says. “Now we want to be able magnify small motions that are hidden within large motions. The basic idea is to try to cancel the large motion and go back to the previous situation.”
Canceling the large motion means determining which pixels of successive frames of video belong to a moving object and which belong to the background. As Durand explains, that problem becomes particularly acute at the object’s boundaries.
If a digital camera captures an image of, say, a red object against a blue background, some of its photosensors will register red light, and some will register blue. But the sensors corresponding to the object’s boundaries may in fact receive light from both foreground and background, so they’ll register varying shades of purple.
See the MIT researchers’ algorithms amplify movements of a shaky gate captured on video. Courtesy of the researchers
Ordinarily, an algorithm separating foreground from background could probably get away with keeping those borderline pixels: A human viewer probably wouldn’t notice a tiny fringe of purple around a red object. But the purpose of the MIT researchers’ motion amplification algorithm is precisely to detect variations invisible to the naked eye. Changes of color at an object’s boundaries could be interpreted as motions requiring magnification.
So Durand, Freeman, and Mohamed Elgharib and Mohamed Hefeeda of QCRI instead assign each boundary pixel a weight, corresponding to the likelihood that it belongs to the foreground object. In the example of the red object against a blue background, that weight would simply depend on whether the shade of purple is bluer or redder. Then, on the basis of the pixels’ weights, the algorithm randomly discards some and keeps others. On average, it will make the right decision, and it will disrupt any patterns of color change that could be mistaken for motion.
The problem of identifying the same object from frame to frame, Durand says, is related to the problem of image stabilization, which attempts to remove camera jitter from video. Identifying the motion of a single object, however, is more difficult than determining the motion of the image as a whole.
The MIT and QCRI researchers make a few assumptions to render the problem more tractable. First, they assume a correlation between the direction and rate of motion of adjacent pixels. Second, they assume “smoothness” — that the direction and rate of motion will be consistent over time. Finally, they assume that pixels’ trajectories across frames can be captured by linear mathematical relationships, which enables their algorithm to analyze pixels individually.
Then, rather than looking for correlations between one frame and the next, their algorithm considers five frames at a time, using consistencies across frames to resolve ambiguities between adjacent frames.
Once the algorithm has identified the pixels correlating to a single moving object, it corrects for the object’s motion and performs the same motion magnification procedure that previous versions did. Finally, it reinserts the magnified motions back into the original video stream.
Suggested Items
Rules of Thumb: Design007 Magazine, November 2024
11/11/2024 | I-Connect007 Editorial TeamRules of thumb are everywhere, but there may be hundreds of rules of thumb for PCB design. They’re built on design formulas, fabricators’ limitations, and tribal knowledge. And unfortunately, some longtime rules of thumb should be avoided at all costs. How do we separate the wheat from the chaff, so to speak?
Connect the Dots: Best Practices for Prototyping
09/21/2023 | Matt Stevenson -- Column: Connect the DotsPCB prototyping is a critical juncture during an electronic device’s journey from concept to reality. Regardless of a project’s complexity, the process of transforming a design into a working board is often enlightening in terms of how a design can be improved before a PCB is ready for full production.
The Drive Toward UHDI and Substrates
09/20/2023 | I-Connect007 Editorial TeamPanasonic’s Darren Hitchcock spoke with the I-Connect007 Editorial Team on the complexities of moving toward ultra HDI manufacturing. As we learn in this conversation, the number of shifting constraints relative to traditional PCB fabrication is quite large and can sometimes conflict with each other.
Asia/Pacific AI Spending Surge to Reach a Projected $78 Billion by 2027
09/19/2023 | IDCAsia/Pacific spending on Artificial Intelligence (AI) ), including software, services, and hardware for AI-centric systems will grow to $78.4 billion in 2027, according to International Data Corporation's latest Worldwide Artificial Intelligence Spending Guide.
Intel to Sell Minority Stake in IMS Nanofabrication Business to TSMC
09/13/2023 | IntelIntel Corporation announced that it has agreed to sell an approximately 10% stake in the IMS Nanofabrication business to TSMC. TSMC’s investment values IMS at approximately $4.3 billion, consistent with the valuation of the recent stake sale to Bain Capital Special Situations.