UW Security Researchers Show That Google’s AI Tool for Video Searching Can Be Easily Deceived
April 4, 2017 | University of WashingtonEstimated reading time: 2 minutes
University of Washington researchers have shown that Google’s new tool that uses machine learning to automatically analyze and label video content can be deceived by inserting a photograph periodically and at a very low rate into videos. After they inserted an image of a car into a video about animals, for instance, the system returned results suggesting the video was about an Audi.
Google recently released its Cloud Video Intelligence API to help developers build applications that can automatically recognize objects and search for content within videos. Automated video annotation would be a breakthrough technology, helping law enforcement efficiently search surveillance videos, sports fans instantly find the moment a goal was scored or video hosting sites weed out inappropriate content.
Google launched a demonstration website that allows anyone to select a video for annotation. The API quickly identifies the key objects within the video, detects scene changes and provides shot labels of the video events over time. The API website says the system can be used to “separate signal from noise, by retrieving relevant information at the video, shot or per frame” level.
In a new research paper, the UW electrical engineers and security researchers, including doctoral students Hossein Hosseini and Baicen Xiao and professor Radha Poovendran, demonstrated that the API can be deceived by slightly manipulating the videos. They showed one can subtly modify the video by inserting an image into it, so that the system returns only the labels related to the inserted image.
The same research team recently showed that Google’s machine-learning-based platform designed to identify and weed out comments from internet trolls can be easily deceived by typos, misspelling offensive words or adding unnecessary punctuation.
“Machine learning systems are generally designed to yield the best performance in benign settings. But in real-world applications, these systems are susceptible to intelligent subversion or attacks,” said senior author Radha Poovendran, chair of the UW electrical engineering department and director of the Network Security Lab. “Designing systems that are robust and resilient to adversaries is critical as we move forward in adopting the AI products in everyday applications.”
As an example, a screenshot of the API’s output is shown below for a sample video named “animals.mp4,” which is provided by the API website. Google’s tool does indeed accurately identify the video labels.
The researchers then inserted the following image of an Audi car into the video once every two seconds. The modification is hardly visible, since the image is added once every 50 video frames, for a frame rate of 25.
The following figure shows a screenshot of the API’s output for the manipulated video. As seen below, the Google tool believes with high confidence that the manipulated video is all about the car.
“Such vulnerability of the video annotation system seriously undermines its usability in real-world applications,” said lead author and UW electrical engineering doctoral student Hossein Hosseini. “It’s important to design the system such that it works equally well in adversarial scenarios.”
“Our Network Security Lab research typically works on the foundations and science of cybersecurity,” said Poovendran, the lead principal investigator of a recently awarded MURI grant, where adversarial machine learning is a significant component. “But our focus also includes developing robust and resilient systems for machine learning and reasoning systems that need to operate in adversarial environments for a wide range of applications.”
The research is funded by the National Science Foundation, Office of Naval Research and Army Research Office.
Suggested Items
The EEcosystem and Dr. Eric Bogatin Launch Free Masterclass for Electronics Engineers
05/01/2025 | The EEcosystemThe EEcosystem, a podcast media and education brand serving professional electronics engineers, is proud to announce the launch of a new online learning platform: The EEcosystem Electronics Masterclass. The platform debuts with Transmission Lines 101, a free course created in partnership with world-renowned signal integrity expert Dr. Eric Bogatin. The course will be available starting May 1, 2025.
Kasuo Electronics Launches Advanced Testing Laboratory to Strengthen Global Supply Chain Quality Assurance
04/29/2025 | BUSINESS WIREKasuo Electronics Co., Ltd, a globally recognized trader of electronic components, has officially operationalized its state-of-the-art testing laboratory.
OSI Systems Receives Orders for $50 Million to Support the Deployment of Cargo and Vehicle Inspection Solutions
04/29/2025 | BUSINESS WIREOSI Systems, Inc. announced that its Security division received multiple awards totaling approximately $50 million from an existing U.S. customer to support the ongoing deployment of its industry leading Rapiscan Cargo and Vehicle Non-Intrusive Inspection technology.
BAE Systems Selected to Enhance Ground System for Space Force Missile Warning Satellites
04/29/2025 | BAE SystemsBAE Systems has been selected by U.S. Space Force Space Systems Command to provide a new satellite command and control (C2) system for its Future Operationally Resilient Ground Evolution (FORGE) program.
Elektrobit, Metoak Forge Strategic Partnership to Establish New Benchmark for Intelligent Driving Safety Ecosystem
04/29/2025 | ElektrobitElektrobit announced a comprehensive strategic partnership with Metoak, a leading Chinese provider of intelligent driving solutions.