Quantcast
Channel: Artificial Intelligence
Viewing all articles
Browse latest Browse all 1375

DeepMind created a YouTube database of humans doing stuff that can help AIs to understand us (GOOG)

$
0
0

James Bardgett checks the quality of beer before filling up kegs and barrels to be dispatched at the Wild Beer Co brewery at Lower Westcombe Farm on February 11, 2016 near Evercreech, England. Over recent years there has been a surge in the popularity of craft beers, which are brewed by smaller independent breweries, such as the Wild Beer Co in Somerset which specialises in beers that have an emphasis on quality, flavour and brewing methods. (Photo by )

Google DeepMind has created a database of hundreds of thousands of YouTube URLs that can help artificial intelligence (AI) agents to identify human actions such as drinking beer, riding a mechanical bull, and bench pressing.

The London-based AI lab, which was acquired by Google in 2014 for a reported £400 million, developed the "Kinetics" dataset for the ActivityNet Challenge and published a paper detailing the work in May. The full paper can be read here.

The database is comprised of some 300,000 already-published "realistic" and "challenging" YouTube URLs that are no more than 10 seconds in length. There are 400 URLs for each of the 400 actions. Each URL directs to a video of a human doing one of 400 actions and has been tagged accordingly by someone via the Amazon Mechanical Turk platform, which allows people to earn money for doing small computer tasks.

"The actions are human focused and cover a broad range of classes including human-object interactions such as playing instruments, as well as human-human interactions such as shaking hands," DeepMind's authors wrote.

DeepMind fed the dataset into a number of off-the-shelf AIs and successfully taught them how recognise certain tasks. Interestingly, DeepMind had more luck teaching AIs how to recognise actions such as bowling, tennis, and trapezing than it did teaching them how to spot yawning, headbutting, and faceplanting. Google DeepMind human actions

Earlier reports suggested that DeepMind had used clips of Homer Simpson performing various actions but this was not the case.

A Google DeepMind spokesperson told Business Insider that data classification is essential to machine learning research.

"AI systems are now very good at recognising objects in images, but still have trouble making sense of videos," said the spokesperson. "One of the main reasons for this is that the research community has so far lacked a large, high-quality video dataset, such as the one we now provide.

"We hope that the Kinetics dataset will help the machine learning community to advance models for video understanding, making a whole range of new research opportunities possible."

A separate paper, published shortly after the first one, shows how DeepMind then used its own algorithms on the Kinetics dataset with even better results.

"We have shown that the performance of deep learning architectures can be substantially improved by first training on Kinetics, and then training and evaluating on standard action classification benchmarks," the DeepMind spokesperson added.

Join the conversation about this story »

NOW WATCH: Here's everything Apple is rumored to be launching in 2017


Viewing all articles
Browse latest Browse all 1375

Trending Articles