Home / Science / UW security researchers show that Google’s AI tool for video searching can be easily deceived

UW security researchers show that Google’s AI tool for video searching can be easily deceived

The UW electrical engineering research team includes Baicen Xiao, Radha Poovendran and Hossein Hosseini.

Click for a full size image



UW security researchers show that Google’s AI tool for video searching can be easily deceived

Jennifer Langston

University of Washington researchers have shown that Google’s new tool that uses machine learning to automatically analyze and label video content can be deceived by inserting a photograph periodically and at a very low rate into videos. After they inserted an image of a car into a video about animals, for instance, the system returned results suggesting the video was about an Audi.

Google recently released its Cloud Video Intelligence API to help developers build applications that can automatically recognize objects and search for content within videos. Automated video annotation would be a breakthrough technology, helping law enforcement efficiently search surveillance videos, sports fans instantly find the moment a goal was scored or video hosting sites weed out inappropriate content.

Google launched a demonstration website that allows anyone to select a video for annotation. The API quickly identifies the key objects within the video, detects scene changes and provides shot labels of the video events over time. The API website says the system can be used to “separate signal from noise, by retrieving relevant information at the video, shot or per frame” level.

In a new research paper, the UW electrical engineers and security researchers, including doctoral students Hossein Hosseini and Baicen Xiao and professor Radha Poovendran, demonstrated that the API can be deceived by slightly manipulating the videos. They showed one can subtly modify the video by inserting an image into it, so that the system returns only the labels related to the inserted image.

The same research team recently showed that Google’s machine-learning-based platform designed to identify and weed out comments from internet trolls can be easily deceived by typos, misspelling offensive words or adding unnecessary punctuation.

“Machine learning systems are generally designed to yield the best performance in benign settings. But in real-world applications, these systems are susceptible to intelligent subversion or attacks,” said senior author Radha Poovendran, chair of the UW electrical engineering department and director of the Network Security Lab. “Designing systems that are robust and resilient to adversaries is critical as we move forward in adopting the AI products in everyday applications.”

As an example, a screenshot of the API’s output is shown below for a sample video named “animals.mp4,” which is provided by the API website. Google’s tool does indeed accurately identify the video labels.

Click for a full size image


The researchers then inserted the following image of an Audi car into the video once every two seconds. The modification is hardly visible, since the image is added once every 50 video frames, for a frame rate of 25.

Click for a full size image


The following figure shows a screenshot of the API’s output for the manipulated video. As seen below, the Google tool believes with high confidence that the manipulated video is all about the car.

Click for a full size image


“Such vulnerability of the video annotation system seriously undermines its usability in real-world applications,” said lead author and UW electrical engineering doctoral student Hossein Hosseini. “It’s important to design the system such that it works equally well in adversarial scenarios.”

“Our Network Security Lab research typically works on the foundations and science of cybersecurity,” said Poovendran, the lead principal investigator of a recently awarded MURI grant, where adversarial machine learning is a significant component. “But our focus also includes developing robust and resilient systems for machine learning and reasoning systems that need to operate in adversarial environments for a wide range of applications.”

The research is funded by the National Science Foundation, Office of Naval Research and Army Research Office.

University of Washington – Seattle, WA

Telegram Channel

About Mohammad Daeizadeh

  • تمامی فایل ها قبل از قرار گیری در سایت تست شده اند.لطفا در صورت بروز هرگونه مشکل از طریق نظرات مارا مطلع سازید.
  • پسورد تمامی فایل های موجود در سایت www.parsseh.com می باشد.(تمامی حروف را می بایست کوچک وارد کنید)
  • Password = www.parsseh.com
  • لطفا نظرات خود را به صورت فارسی بنویسید در صورت تایپ بصورت فینگلیش نظر شما پاک خواهد شد

Leave a Reply

Your email address will not be published. Required fields are marked *


This site uses Akismet to reduce spam. Learn how your comment data is processed.