Auto Check

  • Subscribe to our RSS feed.
  • Twitter
  • StumbleUpon
  • Reddit
  • Facebook
  • Digg

Tuesday, 26 June 2012

Using large-scale brain simulations for machine learning and A.I.

Posted on 10:41 by Unknown
You probably use machine learning technology dozens of times a day without knowing it—it’s a way of training computers on real-world data, and it enables high-quality speech recognition, practical computer vision, email spam blocking and even self-driving cars. But it’s far from perfect—you’ve probably chuckled at poorly transcribed text, a bad translation or a misidentified image. We believe machine learning could be far more accurate, and that smarter computers could make everyday tasks much easier. So our research team has been working on some new approaches to large-scale machine learning.

Today’s machine learning technology takes significant work to adapt to new uses. For example, say we’re trying to build a system that can distinguish between pictures of cars and motorcycles. In the standard machine learning approach, we first have to collect tens of thousands of pictures that have already been labeled as “car” or “motorcycle”—what we call labeled data—to train the system. But labeling takes a lot of work, and there’s comparatively little labeled data out there.

Fortunately, recent research on self-taught learning (PDF) and deep learning suggests we might be able to rely instead on unlabeled data—such as random images fetched off the web or out of YouTube videos. These algorithms work by building artificial neural networks, which loosely simulate neuronal (i.e., the brain’s) learning processes.

Neural networks are very computationally costly, so to date, most networks used in machine learning have used only 1 to 10 million connections. But we suspected that by training much larger networks, we might achieve significantly better accuracy. So we developed a distributed computing infrastructure for training large-scale neural networks. Then, we took an artificial neural network and spread the computation across 16,000 of our CPU cores (in our data centers), and trained models with more than 1 billion connections.

We then ran experiments that asked, informally: If we think of our neural network as simulating a very small-scale “newborn brain,” and show it YouTube video for a week, what will it learn? Our hypothesis was that it would learn to recognize common objects in those videos. Indeed, to our amusement, one of our artificial neurons learned to respond strongly to pictures of... cats. Remember that this network had never been told what a cat was, nor was it given even a single image labeled as a cat. Instead, it “discovered” what a cat looked like by itself from only unlabeled YouTube stills. That’s what we mean by self-taught learning.

One of the neurons in the artificial neural network, trained from still frames from unlabeled YouTube videos, learned to detect cats.

Using this large-scale neural network, we also significantly improved the state of the art on a standard image classification test—in fact, we saw a 70 percent relative improvement in accuracy. We achieved that by taking advantage of the vast amounts of unlabeled data available on the web, and using it to augment a much more limited set of labeled data. This is something we’re really focused on—how to develop machine learning systems that scale well, so that we can take advantage of vast sets of unlabeled training data.

We’re reporting on these experiments, led by Quoc Le, at ICML this week. You can get more details in our Google+ post or read the full paper (PDF).

We’re actively working on scaling our systems to train even larger models. To give you a sense of what we mean by “larger”—while there’s no accepted way to compare artificial neural networks to biological brains, as a very rough comparison an adult human brain has around 100 trillion connections. So we still have lots of room to grow.

And this isn’t just about images—we’re actively working with other groups within Google on applying this artificial neural network approach to other areas such as speech recognition and natural language modeling. Someday this could make the tools you use every day work better, faster and smarter.

Posted by Jeff Dean, Google Fellow and Andrew Ng, Visiting Faculty
Email ThisBlogThis!Share to XShare to Facebook
Posted in education and research | No comments
Newer Post Older Post Home

0 comments:

Post a Comment

Subscribe to: Post Comments (Atom)

Popular Posts

  • Hulu Plus now works with Chromecast
    Hulu has added Chromecast support to their Hulu Plus app—just in time for the fall television season. Now you can easily enjoy your favori...
  • Providing a springboard for women entrepreneurs in India
    Meghana Musunuri was a typical female entrepreneur in India. Born and brought up in Medak , she received a good education and spent time ab...
  • A look inside our 2011 diversity report
    We work hard to ensure that our commitment to diversity is built into everything we do—from hiring our employees and building our company cu...
  • Software downloads in Syria
    Free expression is a fundamental human right and a core value of our company—but sometimes there are limits to where we can make our product...
  • Celebrating teachers on National Teacher Day
    One of the best parts of my job working on the Google Education team has been hearing inspiring stories time and again of great teachers who...
  • Shiver me timbers, the 2012 D4G Winner is....
    After 114,000 submissions and millions of your votes, second grader Dylan Hoffman of Caledonia, Wisc. is this year’s U.S. Doodle 4 Google N...
  • Supporting Innovation in African News
    Cross-posted from the European Public Policy Blog We’re eager to see journalism flourish in the digital age, in all forms and on all contine...
  • Google+ Hangouts On Air: broadcast your conversation to the world
    Last year we introduced Hangouts On Air to a limited number of broadcasters, enabling them to go live with friends and fans, for all the wo...
  • New research shows smartphone growth is global
    Last October, we launched Our Mobile Planet , a resource enabling anyone to visualize the ways smartphones are transforming how people conne...
  • Local—now with a dash of Zagat and a sprinkle of Google+
    Finding the best places to go is an essential part of our lives, as are the people and resources that help us make those decisions. In fact,...

Categories

  • accessibility
  • acquisition
  • ads
  • Africa
  • Android
  • apps
  • Asia
  • books + book search
  • chrome
  • chrome + chrome os
  • commerce
  • computing history
  • crisis response
  • Cultural Institute
  • culture
  • developers
  • display advertising
  • diversity
  • doodles
  • education
  • education and research
  • energy
  • enterprise
  • entrepreneurs at Google
  • entrepreneurship
  • Europe
  • events
  • faster web
  • free expression
  • g2g
  • giving
  • Google Apps highlights
  • google ideas
  • google play
  • google.org
  • google+
  • googleplus
  • googlers and culture
  • government transparency
  • green
  • innovation
  • ipv6
  • journalism and news
  • Latin America
  • local
  • maps and earth
  • mobile
  • online safety
  • open source
  • personalization
  • photos
  • policy and issues
  • politics
  • privacy
  • privacy and security
  • publishers
  • scholarships
  • search
  • search stories
  • search trends
  • security
  • security and safety tips
  • small business
  • transparency
  • youtube and video

Blog Archive

  • ►  2013 (190)
    • ►  December (11)
    • ►  November (13)
    • ►  October (15)
    • ►  September (12)
    • ►  August (10)
    • ►  July (13)
    • ►  June (28)
    • ►  May (16)
    • ►  April (21)
    • ►  March (18)
    • ►  February (19)
    • ►  January (14)
  • ▼  2012 (269)
    • ►  December (25)
    • ►  November (20)
    • ►  October (18)
    • ►  September (16)
    • ►  August (19)
    • ►  July (20)
    • ▼  June (28)
      • Celebrating Pride 2012
      • Chrome & Apps @ Google I/O: Your web, everywhere
      • Project Glass demo: Hangouts IN Air
      • +1
      • Android @ I/O: the playground is open
      • Using large-scale brain simulations for machine le...
      • Become a Google power searcher
      • Supporting entrepreneurship in France at Le Camping
      • Tune in to I/O Live at 9:30 a.m. PDT on June 27
      • In schools, all you need is web
      • A tribute to Turing, the father of modern computing
      • Project Re: Brief, the documentary
      • Fueling great nonprofits with technology
      • The Endangered Languages Project: Supporting langu...
      • Street View says "aloha" from Hawaii
      • Safe Browsing—protecting web users for five years ...
      • The Surui Cultural Map
      • Energy efficiency in the cloud
      • More transparency into government requests
      • Ads Integrity Alliance: Working together to fight ...
      • Find out what people are searching for with the up...
      • Connecting shoppers and great stores online
      • AdWords, meet AdMob
      • The never-ending quest for the perfect map
      • 15 Google Science Fair Finalists and the Science i...
      • World IPv6 Launch: Keeping the Internet growing
      • Google + Quickoffice = get more done anytime, anyw...
      • Reimagining the future of buying and selling ads o...
    • ►  May (30)
    • ►  April (19)
    • ►  March (27)
    • ►  February (23)
    • ►  January (24)
  • ►  2011 (41)
    • ►  December (33)
    • ►  November (8)
Powered by Blogger.

About Me

Unknown
View my complete profile