What is Computer Science - unplugged activity

0:01   What do you want to be when you grow up Olivia? Umm, an astronaut! Do you happen to know what
0:11   a computer programmer is? Yeahh, umm, no. Umm, what what? I'm not really sure how to
0:18   explain it. Computer programming is pretty simple. It's a set of instructions, like a
0:22   recipe. You have to follow them step by step to get the end result you want. Computer science
0:27   is a way to impact the world. It can be music videos, it can be games, detect whether or
0:34   not someone is related to someone else. Find you know, people's friends. You can do all
0:38   sorts of other crazy things that actually save lives. You do have to have a drive I
0:43   think. It is to me like a paintbrush. I think great programming is not all that dissimilar
0:51   from great art. When I finally learned a little bit of programming, that blank wall resolved
0:56   into a bunch of doors and you open them and of course then you find behind them is another
1:01   hallway filled with a bunch of doors. Programming is fun and easy. You can do anything your
1:07   mind wants to do. Finally you start to open enough doors the light comes in. To me a finished
1:13   program is like a structure filled with light. All the corners are illuminated. The number
1:18   of people that you can touch and interact with is something the world has never seen
1:25   before. Our first lesson in this series is all about what computer science is, what a
1:31   computer scientist does and how you can be more responsible in your use of technology.
1:37   It's a very important lesson but it is a little text-heavy. At the end, you get to make your
1:42   very own customized encoding using your initials. It's a fun activity and it's very empowering
1:49   because binary is one of those things that feels very technical but once you understand
1:54   it, it's like you speak a secret language.
Transcripción : Youtube

"Code Stars" - Short Film

0:01   "Everybody in this country should learn how to program a computer...
0:04   because it teaches you how to think." - Steve Jobs
0:09   What do you want to be when you grow up? Um... an astronaut.
0:12   I want to be a fashion designer. A basketball player. I want to be an actor. A doctor. A
0:17   teacher. A chef. An artist. What do you wanna be when you grow up? A mermaid!
0:24   (interviewer) Do you know what a computer programmer is? (student) Yeah--umm... no. No. Uhh, no.
0:32   I think it's something that has code and it's able to decode a mystery. I think that they...
0:41   umm, wait what? (interviewer) Computer programmer? (student) No.
0:48   Nowadays, just about everything requires some form of programming. So what is it?
0:55   Programming is basically explaining to a computer what you want it to do for you. When you're programming, you're teaching possibly
1:02   the stupidest thing in the universe, a computer, how to do something. Programming is one of
1:08   the only things in the world that you can do where you can sit down and just make something
1:13   completely new from scratch-whatever you want. It's really not unlike playing an instrument
1:18   or playing a sport. It starts out being very intimidating, but you kind of get the hang
1:26   of it over time. Coding is something that can be learned and I know it can be intimidating,
1:31   and a lot of things are intimidating, but what isn't? A lot of the coding people do
1:38   is actually fairly simple. It's more about the process of breaking down problems than
1:46   coming up with complicated algorithms as people traditionally think about it. Well if it's
1:51   fairly simple, why aren't there more of us doing it? Over the next 10 years there will
1:57   be 1.4 million jobs in computer science and only about 400,000 grads qualify for those
2:02   jobs. That's a shortage of a million people! So how do you start?
2:16   I was obsessed with maps when I was a kid, and cities specifically, so I taught myself how to program. I had a
2:25   very clear goal of what I wanted to do which was to see a map of the city on my screen
2:31   and play with it. Put things on the map, move things around the map, see what was happening
2:36   in the city. How it worked, how it lived, how it breathed. The best early thing was
2:41   actually using software to decide when the classes in my school would meet. And that
2:47   put me in a position to decide which girls were in my class.
2:54   The first program I wrote asked things like, "What's your favorite color?" or "How old are you?" I first learned how
3:00   to make a green circle and a red square appear on the screen. The first time I actually had
3:06   something come up and say "Hello world!" I made a computer do that? It was astonishing.
3:11   When I finally learned a little bit of programming, that blank wall resolved into a bunch of doors.
3:17   And you open them and finally you start to open enough doors that the light comes in.
3:21   And to me, a finished program is like a structure filled with light. All the corners are illuminated
3:27   and you understand the structure of it. It's a really serene feeling to have completed that.
3:41   It took me some time to realize that creating things with your hands or creating
3:45   code, creating programs is just a different way to express creativity. I think right now
3:52   there's a big emergence of the culture of making. People who make their own scarves
3:59   and hats, people who write their own apps. Now it's just limited by your imagination.
4:03   And sort of what kinds of ideas, what kind of understanding can you build into a computer
4:10   to do these things that were previously impossible.
4:18   All great things are built in teams when you
4:22   collaborate with other smart people. You're testing your ideas, you're stimulating each
4:28   other, that's what makes us successful. It's not some flash of brilliance from somebody
4:33   who codes 24 hours a day for 3 weeks. The magic happens when we're all on the same page
4:41   collaborating and building something together. There's a much greater need in the world for
4:46   engineers and people who can write code than there will ever be supply. And so we all live
4:54   these very charmed lives. To get the very best people we try to make the office as awesome as possible.
5:21   We have a fantastic chef. Free food. Breakfast, lunch, and dinner. Free laundry.
5:28   Snacks. Even places to play and video games and scooters. There's all these kind of interesting
5:35   things around the office.
5:38   Places where people can play or relax, or go to think, or play music, or be creative.
5:46   I went on the Bureau of Labor Statistics for the United States
5:49   and there's about a third of the pie that's all the things that you would expect. They're
5:54   working for the government, they're working in typical technology jobs, but then the rest
6:00   of the pie--the majority of the pie--just split down into these little tiny slices of
6:04   every industry imaginable. And what it is, is computers are everywhere! Do you want to
6:09   work in agriculture? Do you want to work in entertainment? Do you want to work in manufacturing?
6:15   It's just all over.
6:28   Here we are, 2013, and we all depend on technology to communicate,
6:33   to bank. Information. And none of us know how to read and write code.
6:43   So you guys, what else? Who else has an idea of what we can change with our program?
6:48   What else can we do?
6:51   What I saw my students take away from using Scratch and programming in our classroom
6:57   is that they're willing to push through problems. It really builds critical thinking. It builds
7:02   problem solving. And it's something that they can then apply to math in the classroom. Or
7:08   their reading skills. We integrated science with this programming and I saw my scores
7:15   go up 30%. When I was in school I was in this after school club called the Whiz Kids and
7:23   when people found out they laughed at me. You know all these things. And I'm like, man
7:27   I don't care. I think it's cool. You know I'm learning a lot and some of my friends
7:33   have jobs. It's important for these kids. It should be mandatory. To be a citizen on
7:40   this planet, to read and write code.
7:47   I just think you have to start small. That's one of the biggest misconceptions about
7:52   computer science and programming overall is that you
7:55   have to learn this big body of information before you can do anything. You don't have
8:00   to be a genius to know how to code. You need to be determined. Addition, subtraction, that's
8:06   about it. You should probably know your multiplication tables. You don't have to be a genius to code.
8:11   Do you have to be a genius to read? Do you have to be a genius to do math? No. I think
8:18   if someone had told me that software is really about humanity. That it's really about helping
8:26   people by using computer technology, it would have changed my outlook a lot earlier. Whether
8:31   you're trying to make a lot of money or whether you just want to change the world, computer
8:34   programming is an incredibly empowering skill to learn. To be able to actually come up with
8:38   an idea and then see it in your hands and then press a button and have it be in millions
8:44   of people's hands, I think we're the first generation in the world that's really had
8:48   that kind of experience. The programmers of tomorrow are the wizards of the future. You're
8:53   going to look like you have magic powers compared to everybody else. I think it's amazing. I
8:57   think it's the closest thing we have to a superpower. Great coders are today's rockstars.
9:03   That's it.
9:06   To start learning a superpower go to Code.org.
Transcripción : Youtube

Machine Learning: Google's Vision - Google I/O 2016

0:00   TOM SIMONITE: Hi.
0:10   Good morning.
0:11   Welcome to day three of Google I/O,
0:14   and what should be a fun conversation about machine
0:16   learning and artificial intelligence.
0:18   My name is Tom Simonite.
0:19   I'm San Francisco bureau chief for MIT Technology Review.
0:23   And like all of you, I've been hearing a lot recently
0:26   about the growing power of machine learning.
0:28   We've seen some striking results come out
0:30   of academic and industrial research labs,
0:33   and they've moved very quickly into the hands of developers,
0:36   who have been using them to make new products and services
0:39   and companies.
0:40   I'm joined by three people this morning
0:42   who can tell us about how this new technology
0:45   and the capabilities it brings are coming out into the world.
0:48   They are Aparna Chennapragada, who
0:51   is the director of product management
0:53   and worked on the Google Now mobile assistant,
0:56   Jeff Dean, who leads the Google Brain research group here
1:00   in Mountain View, and John Giannandrea,
1:02   who is head of search and machine intelligence at Google.
1:06   Thanks for joining me, all of you.
1:08   We're going to talk for about 30 minutes,
1:10   and then there will be time for questions from the floor.
1:15   John, why don't we start with you?
1:16   You could set the scene for us.
1:19   Artificial intelligence and machine learning
1:21   are not brand new concepts.
1:23   They've been around for a long time,
1:24   but we're suddenly hearing a lot more about them.
1:27   Large companies and small companies
1:28   are investing more in this technology,
1:30   and there's a lot of excitement.
1:31   You can even get a large number of people
1:33   to come to a talk about this thing early in the morning.
1:37   So what's going on?
1:39   Tell these people why they're here.
1:41   JOHN GIANNANDREA: What's going on?
1:41   Yeah, thanks, Tom.
1:42   I mean, I think in the last few years,
1:44   we've seen extraordinary results in fields that hadn't really
1:48   moved the needle for many years, like speech recognition
1:51   and image understanding.
1:52   The error rates are just falling dramatically,
1:55   mostly because of advances in deep neural networks,
1:58   so-called deep learning.
2:00   I think these techniques are not new.
2:03   People have been using neural networks for many, many years.
2:06   But a combination of events over the last few years
2:09   has made them much more effective,
2:11   and caused us to invest a lot in getting them
2:14   into the hands of developers.
2:17   People talk about it in terms of AI winters,
2:19   and things like this.
2:20   I think we're kind of an AI spring right now.
2:23   We're just seeing remarkable progress
2:25   across a huge number of fields.
2:26   TOM SIMONITE: OK.
2:27   And now, how long have you worked
2:28   in artificial intelligence, John?
2:30   JOHN GIANNANDREA: Well, we started
2:31   investing heavily in this at Google about four years ago.
2:33   I mean, we've been working in these fields,
2:35   like speech recognition, for over a decade.
2:38   But we kind of got serious about our investments
2:40   about four years ago, and getting organized
2:44   to do things that ultimately resulted
2:46   in the release of things like TensorFlow, which
2:48   Jeff's team's worked on.
2:49   TOM SIMONITE: OK.
2:49   And we'll talk more about that later, I'm sure.
2:52   Aparna, give us a perspective from the view of someone
2:56   who builds products.
2:57   So John says this technology has suddenly
2:59   become more powerful and accurate and useful.
3:03   Does that open up new horizons for you,
3:05   when you're thinking about what you can build?
3:06   APARNA CHENNAPRAGADA: Yeah, absolutely.
3:08   I think for me, these are great as a technology.
3:12   But as a means to an end, they're
3:13   powerful tool kits to help solve real problems, right?
3:17   And for us, as building products, and for you guys,
3:20   too, there's two ways that machine learning
3:22   changes the game.
3:24   One is that it can turbo charge existing use cases-- that
3:27   is, existing problems like speech recognition--
3:30   by dramatically changing some technical components
3:33   that power the product.
3:34   If you're building a voice enabled assistant, the word
3:37   error rate that John was talking about, as soon as it dropped,
3:40   we actually saw the usage go up.
3:42   So the product gets more usable as machine learning improves
3:46   the underlying engine.
3:47   Same thing with translation.
3:48   As translation gets better, Google Translate,
3:51   it scales to 100-plus languages.
3:54   And photos is a great example.
3:55   You've heard Sundar talk about it, too,
3:57   that as soon as you have better image understanding,
4:00   the photo labeling gets better, and therefore, I
4:02   can organize my photos.
4:03   So it's a means to an end.
4:04   That's one way, certainly, that we have seen.
4:06   But I think the second way that's, personally, far more
4:09   exciting to see is where it can unlock new product use cases.
4:14   So turbocharging existing use cases is one thing,
4:17   but where can you kind of see problems
4:19   that really weren't thought of as AI or data problems?
4:22   And thanks to mobile, here-- 3 billion phones-- a lot
4:26   of the real world problems are turning into AI problems,
4:29   right?
4:29   Transportation, health, and so on.
4:31   That's pretty exciting, too.
4:32   TOM SIMONITE: OK.
4:33   And so is one consequence of this
4:35   that we can make computers less annoying, do you think?
4:38   I mean, that would be nice.
4:40   We'd all had these experiences where
4:41   you have a very clear idea of what it is you're trying to do,
4:44   but it feels like the software is doing
4:46   everything it can to stop you.
4:47   Maybe that's a form of artificial intelligence, too.
4:50   I don't know.
4:50   But can you make more seamless experiences
4:53   that just make life easier?
4:55   APARNA CHENNAPRAGADA: Yeah.
4:56   And I think in this case, again, one of the things
4:59   to think about is, how do you make sure-- especially
5:01   as you build products-- how do you
5:03   make sure your interface scales with the intelligence?
5:06   The UI needs to be proportional to AI.
5:09   I cannot believe I said some pseudo formula in front of Jeff
5:12   Dean.
5:14   But I think that's really important,
5:15   to make sure that the UI scales with the AI.
5:18   TOM SIMONITE: OK.
5:19   And Jeff, for people like Aparna,
5:23   building products, to do that, we
5:26   need this kind of translation step
5:27   which your group is working on.
5:29   So Google Brain is a research group.
5:30   Works in some very fundamental questions in its field.
5:33   But you also build this infrastructure,
5:36   which you're kind of inventing from scratch, that makes
5:38   it possible to use this stuff.
5:41   JEFF DEAN: Yeah.
5:42   I mean, I think, obviously, in order
5:44   to make progress on these kinds of problems,
5:46   it's really important to be able to try lots of experiments
5:50   and do that as quickly as you can.
5:52   There's a very fundamental difference
5:55   between having an experiment take a few hours,
5:58   versus something that takes six weeks.
5:59   It's just a very different model of doing science.
6:03   And so, one of the things we work on
6:06   is trying to build really scalable systems that are also
6:10   flexible and easy to express new kinds of machine learning
6:13   ideas.
6:14   So that's how TensorFlow came about.
6:16   It's sort of our internal research vehicle,
6:19   but also robust enough to take something you've done and done
6:23   lots of experiments on, and then, when you get something
6:25   that works well, to take that and move it into a production
6:28   environment, run things on phones or in data
6:31   centers, on RTPUs, that we announced a couple days ago.
6:36   And that seamless transition from research
6:39   to putting things into real products
6:41   is what we're all about.
6:43   TOM SIMONITE: OK.
6:44   And so, TensorFlow is this very flexible package.
6:48   It's very valuable to Google.
6:49   You're building a lot of things on top of it.
6:51   But you're giving it away for free.
6:52   Have you thought this through?
6:54   Isn't this something you should be keeping closely held?
6:56   JEFF DEAN: Yeah.
6:57   There was actually a little bit of debate internally.
7:00   But I think we decided to open source it,
7:02   and it's got a nice Apache 2.0 license which basically
7:05   means you can take it and do pretty much whatever
7:07   you want with it.
7:09   And the reason we did that is several fold.
7:14   One is, we think it's a really good way of making research
7:18   ideas and machine learning propagate more quickly
7:20   throughout the community.
7:22   People can publish something they've done,
7:26   and people can pick up that thing
7:27   and reproduce those people's results or build on them.
7:30   And if you look on GitHub, there's
7:33   like 1,500 repositories, now, that mention TensorFlow,
7:36   and only five of them are from Google.
7:38   And so, it's people doing all kinds of stuff with TensorFlow.
7:41   And I think that free exchange of ideas and accelerating
7:43   of that is one of the main reasons we did that.
7:47   TOM SIMONITE: OK.
7:47   And where is this going?
7:49   So I imagine, right now, that TensorFlow is mostly
7:52   used by people who are quite familiar with machine learning.
7:55   But ultimately, the way I hear people
7:59   talk about machine learning, it's
8:00   just going to be used by everyone everywhere.
8:03   So can developers who don't have much
8:05   of a background in this stuff pick it up yet?
8:07   Is that possible?
8:08   JEFF DEAN: Yeah.
8:09   So I think, actually, there's a whole set
8:12   of ways in which people can take advantage of machine learning.
8:15   One is, as a fundamental machine learning researcher,
8:18   you want to develop new algorithms.
8:19   And that's going to be a relatively small fraction
8:21   of people in the world.
8:23   But as new algorithms and models are developed
8:26   to solve particular problems, those models
8:29   can be applied in lots of different kinds of things.
8:31   If you look at the use of machine learning
8:35   in the diabetic retinopathy stuff
8:36   that Sundar mentioned a couple days ago,
8:39   that's a very similar problem to a lot of other problems
8:41   where you're trying to look at an image
8:43   and detect some part of it that's unusual.
8:45   We have a similar problem of finding text
8:48   in Street View images so that we can read the text.
8:51   And that looks pretty similar to a model
8:54   to detect diseased parts of an eye, just different training
8:58   data, but the same model.
8:59   So I think the broader set of models
9:02   will be accessible to more and more people.
9:05   And then there's even an easier way,
9:07   where you don't really need much machine learning knowledge
9:09   at all, and that is to use pre-trained APIs.
9:12   Essentially, you can use our Cloud Vision API
9:15   or our Speech APIs very simply.
9:17   You just give us an image, and we give you back good stuff.
9:20   And as part of the TensorFlow flow open source,
9:22   we also released, for example, an inception model that
9:26   does image classification that's the same model as underlies
9:29   Google Photos.
9:30   TOM SIMONITE: OK.
9:31   So will it be possible for someone-- maybe they're
9:33   an experienced builder of apps, but don't know much about
9:37   machine learning-- they could just
9:39   have an idea and kind of use these building blocks to put it
9:42   together?
9:42   JEFF DEAN: Yeah.
9:42   Actually, I think one of the reasons TensorFlow has taken
9:45   off, is the tutorials in TensorFlow are actually
9:47   quite good at illustrating six or seven important kinds
9:53   of models in machine learning, and showing people
9:55   how they work, stepping through both the machine learning
9:58   that's going on underneath, and also how you express them
10:01   in TensorFlow.
10:01   That's been pretty well received.
10:03   TOM SIMONITE: OK.
10:04   And Aparna, I think we've seen in the past
10:06   that when a new platform of mode of interaction comes forward,
10:10   we have to experiment with it for some time
10:13   before we figure out what works, right?
10:16   And sometimes, when we look back,
10:17   we might think, oh, those first generation
10:19   mobile apps were kind of clunky, and maybe not so smart.
10:23   How are we going with that process
10:25   here, where we're starting to have to understand
10:28   what types of interaction work?
10:30   APARNA CHENNAPRAGADA: Yeah.
10:32   And I think it's one of the things that's not intuitive
10:34   when you start out, you rush out into a new area,
10:36   like we've all done.
10:38   So one experience, for example, when
10:39   we started working on Google Now, one thing we realized
10:42   is, it's really important to make sure
10:44   that, depending on the product domain, some of these black box
10:49   systems, you need to pay attention
10:51   to what we call internally as the wow to WTH ratio.
10:55   That is, as soon as you kind of say,
10:57   hey, there are some delightful magical moments, right?
11:00   But then, if you kind of get it wrong,
11:02   there's a high cost to the user.
11:04   So to give you an example, in Google Search,
11:06   let's say you search for, I don't know, Justin Timberlake,
11:09   and we got a slightly less relevant answer.
11:12   Not a big deal, right?
11:13   But then, if the assistant told you to sit in the car,
11:16   go drive to the airport, and you missed
11:18   your flight, what the hell?
11:21   So I think it's really important to get that ratio right,
11:23   especially in the early stages of this new platform.
11:27   The other thing we noticed also is
11:29   that explainability or interpretability really builds
11:33   trust in many of these cases.
11:35   So you want to be careful about looking
11:37   at which parts of the problem you use machine learning
11:41   and you drop this into.
11:43   You want to look at problems that are easy for machines
11:46   and hard for humans, the repetitive things,
11:48   and then make sure that those are the problems that you
11:50   throw machine learning against.
11:52   But you don't want to be unpredictable and inscrutable.
11:56   TOM SIMONITE: And one mode of interaction that everyone seems
11:59   to be very excited about, now, is this idea
12:01   of conversational interface.
12:02   So we saw the introduction on Wednesday of Google Assistant,
12:06   but lots of other companies are building these things, too.
12:10   Do we know that definitely works?
12:13   What do we know about how you design
12:15   a conversational interface, or what the limitations
12:17   and strengths are?
12:19   APARNA CHENNAPRAGADA: I think, again, at a broad level,
12:21   you want to make sure that you can have this trust.
12:24   So [INAUDIBLE] domains make it easy.
12:26   So it's very hard to make a very horizontal system
12:29   work that works for anything.
12:31   But I'm actually pretty excited at the progress.
12:33   We just launched-- open sourced-- the sentence parser,
12:36   Parsey Mcparseface.
12:37   I just wanted to say that name.
12:41   But it's really exciting, because then you say,
12:43   OK, you're starting to see the beginning of conversational,
12:46   or at least a natural language sentence understanding,
12:49   and then you have building blocks that build on top of it.
12:52   TOM SIMONITE: OK.
12:52   And John, with your search hat on for a second,
12:56   we heard on Wednesday that, I think, 20% of US searches
13:01   are now done by voice.
13:02   So people have clearly got comfortable with this,
13:04   and you've managed to provide something
13:06   that they want to use.
13:09   Is the Assistant interface to search
13:12   going to grow in a similar way, do you think?
13:14   Is it going to take over a big chunk of people's search
13:17   queries?
13:18   JOHN GIANNANDREA: Yeah.
13:19   We think of the Assistant as a fundamentally different product
13:22   than search, and I think it's going
13:24   to be used in a different way.
13:25   But we've been working on what we
13:26   call voice search for many, many years,
13:28   and we have this evidence that people
13:30   like it and are using it.
13:32   And I would say our key differentiator, there, is just
13:36   the depth of search, and the number of questions
13:38   we can answer, and the kinds of complexities
13:40   that we can deal with.
13:43   I think language and dialogue is the big unsolved problem
13:46   in computer science.
13:48   So imagine you're reading an article
13:50   and then writing a shorter version of it.
13:52   That's currently beyond the state of the art.
13:54   I think the important thing about the open source release
13:56   we did of the parser is it's using TensorFlow as well.
14:02   So in the same way as Jeff explained,
14:03   the functionality of this in Google Photos for finding
14:06   your photos is actually available open source,
14:08   and people can actually play with it
14:09   and run a cloud version of it.
14:11   We feel the same way about natural language understanding,
14:13   and we have many more years of investment
14:15   to make in getting to really natural dialogue systems,
14:19   where you can say anything you want,
14:20   and we have a good shot of understanding it.
14:23   So for us, this is a journey.
14:25   Clearly, we have a fairly usable product in voice search today.
14:29   And the Assistant, we hope, when we launch
14:31   later this year, people will similarly
14:33   like to use it and find it useful.
14:36   TOM SIMONITE: OK.
14:36   Do you need a different monetization model
14:39   for the Assistant dialogue?
14:40   Is that something--
14:42   JOHN GIANNANDREA: We're really focused, right now,
14:42   on building something that users like to use.
14:45   I think Google has a long history
14:46   of trying to build things that people find useful.
14:49   And if they find them useful, and they use them at scale,
14:52   then we'll figure out a way to actually have a business
14:54   to support that.
14:56   TOM SIMONITE: OK.
14:57   So you mentioned that there are still
14:58   a lot of open research questions here,
14:59   so maybe we could talk about that a little bit.
15:03   As you described, there have been
15:05   some very striking improvements in machine learning recently,
15:08   but there's a lot that can't be done.
15:09   I mean, if I go to my daughter's preschool,
15:11   I would see young children learning and using
15:14   language in ways that your software can't match right now.
15:17   So can you give us a summary of the territory that's
15:21   still to be explored?
15:22   JOHN GIANNANDREA: Yeah.
15:23   There's a lot still to be done.
15:25   I think there's a couple of areas
15:28   which researchers around the world
15:30   are furiously trying to attack.
15:32   So one is learning from smaller numbers of examples.
15:35   Today, the learning systems that we have,
15:37   including deep neural networks, typically
15:39   require really large numbers of examples.
15:41   Which is why, as Jeff was describing,
15:43   they can take a long time to train,
15:44   and the experiment time can be slow.
15:48   So it's great that we can give systems
15:51   hundreds of thousands or millions of labeled examples,
15:53   but clearly, small children don't need to do that.
15:56   They can learn from very small numbers of examples.
15:58   So that's an open problem.
16:00   I think another very important problem in machine learning
16:02   is what the researchers call transfer learning, which
16:05   is learning something in one domain,
16:07   and then being able to apply it in another.
16:09   Right now, you have to build a system
16:11   to learn one particular task, and then that's not
16:13   transferable to another task.
16:14   So for example, the AlphaGo system that
16:17   won the Go Championship in Korea,
16:20   that system can't, a priori, play chess or tic tac toe.
16:24   So that's a big, big open problem
16:26   in machine learning that lots of people are interested in.
16:28   TOM SIMONITE: OK.
16:29   And Jeff, this is kind of on your group, to some extent,
16:33   isn't it?
16:34   You need to figure this out.
16:35   Are there particular avenues or recent results
16:38   that you would highlight that seem to be promising?
16:41   JEFF DEAN: Yeah.
16:42   I think we're making, actually, pretty significant progress
16:46   in doing a better job of language understanding.
16:48   I think, if you look at where computer vision was three
16:53   or four or five years ago, it was
16:54   kind of just starting to show signs of life,
16:57   in terms of really making progress.
16:58   And I think we're starting to see the same thing in language
17:02   understanding kinds of models, translation, parsing, question
17:06   answering kinds of things.
17:08   In terms of open problems, I think unsupervised
17:12   learning, being able to learn from observations
17:14   of the world that are not labeled,
17:15   and then occasionally getting a few labeled examples that
17:18   tell you, these are important things about the world
17:21   to pay attention to, that's really
17:23   one of the key open challenges in machine learning.
17:27   And one more, I would add, is, right now,
17:31   what you need a lot of machine learning expertise for
17:34   is to kind of device the right model structure
17:36   for a particular kind of problem.
17:38   For an image problem, I should use convolutional neural nets,
17:41   or for language problems, I should use this particular kind
17:44   of recurrent neural net.
17:46   And I think one of the things that
17:48   would be really powerful and amazing
17:50   is if the system itself could device the right structure
17:54   for the data it's observing.
17:57   So learning model structure concurrently
17:59   with trying to solve some set of tasks, I think,
18:02   would be a really great open research problem.
18:05   TOM SIMONITE: OK.
18:05   So instead of you having to design the system
18:08   and then setting it loose to learn,
18:11   the learning system would build itself, to some extent?
18:13   JEFF DEAN: Right.
18:14   Right now, you kind of define the scaffolding of the model,
18:17   and then you fiddle with parameters
18:18   as part of the learning process, but you don't sort of
18:21   introduce new kinds of connections
18:22   in the model structure itself.
18:24   TOM SIMONITE: Right.
18:25   OK.
18:25   And unsupervised learning, just giving it that label,
18:29   it makes it sound like one unitary problem, which
18:31   may not be true.
18:32   But will big progress on that come
18:36   from one flash of insight and a new algorithm,
18:41   or will it be-- I don't know-- a longer slog?
18:46   JEFF DEAN: Yeah.
18:47   If I knew, that would be [INAUDIBLE].
18:50   I have a feeling that it's not going to be, like,
18:53   100 different things.
18:54   I feel like there's a few key insights
18:57   that new kinds of learning algorithms
19:00   could pick up on as to what aspects
19:03   of the world the model is observing are important.
19:06   And knowing which things are important
19:08   is one of the key things about unsupervised learning.
19:11   TOM SIMONITE: OK.
19:12   Aparna, so what Jeff's team kind of works out, eventually,
19:18   should come through into your hands,
19:19   and you could build stuff with it.
19:21   Is there something that you would really
19:23   like him to invent tomorrow, so you can start building
19:26   stuff with it the day after?
19:28   APARNA CHENNAPRAGADA: Auto generate emails.
19:30   No, I'm kidding.
19:32   I do think, actually, what's interesting is, you've heard
19:35   these building blocks, right?
19:36   So machine perception, computer vision, wasn't a thing,
19:40   and now it's actually reliable.
19:41   Language understanding, it's getting there.
19:44   Translation is getting there.
19:45   To me, the next other building block you can make machines do
19:49   is hand-eye coordination.
19:51   So you've seen the robot arms video
19:53   that Sundar talked about and showed at the keynote,
19:56   but imagine if you could kind of have these rote tasks that
20:00   are harder, tedious for humans, but if you
20:03   had reliable hand-eye coordination built in, that's
20:07   in a learned system versus a controlled system code
20:09   that you usually write, and it's very brittle,
20:11   suddenly, it opens up a lot more opportunities.
20:13   Just off the top of my head, why isn't there
20:16   anything for, like, elderly care?
20:18   Like, you are an 80-year-old woman with a bad back,
20:21   and you're picking up things.
20:23   Why isn't there something there?
20:24   Or even something as mundane with natural language
20:27   understanding, right?
20:28   I have a seven-year-old.
20:29   I'm a mom of a 7-year-old.
20:31   Why isn't there something for, I don't know,
20:33   math homework, with natural language understanding?
20:36   JOHN GIANNANDREA: So I think one of things
20:38   we've learned in the last few years
20:39   is that things that are hard for people
20:42   to do, we can teach computers to do,
20:44   and things that are easy for us to do
20:45   are still the hard problems for computers.
20:47   TOM SIMONITE: Right.
20:48   OK.
20:49   And does that mean we're still missing some big new field
20:56   we need to invent?
20:57   Because most of the things we've been talking about so far
20:59   have been built on top of this deep learning
21:01   and neural network.
21:02   JOHN GIANNANDREA: I think robotics work is interesting,
21:04   because it gives the computer system an embodiment
21:08   in the world, right?
21:10   So learning from tactile environments
21:13   is a new kind of learning, as opposed to just looking
21:16   at unsupervised or supervised.
21:17   Just reading text is a particular environment.
21:21   Perception, looking at images, looking at audio,
21:23   trying to understand what this song is,
21:25   that's another kind of problem.
21:27   I think interacting with the real world
21:29   is a whole other kind of problem.
21:30   TOM SIMONITE: Right.
21:30   OK.
21:31   That's interesting.
21:33   Maybe this is a good time to talk a little bit more
21:35   about DeepMind.
21:35   I know that they are very interested in this idea
21:38   of embodiment, the idea you have to submerge this learning
21:43   agent in a world that it can learn from.
21:45   Can you explain how they're approaching this?
21:47   JOHN GIANNANDREA: Yeah, sure.
21:48   I mean, DeepMind is another research group
21:49   that we have at Google, and we work closely with them
21:52   all the time.
21:53   They are particularly interested in learning from simulations.
21:57   So they've done a lot of work with video games
21:59   and simulations of physical environments,
22:01   and that's one of the research directions that they have.
22:04   It's been very productive.
22:06   TOM SIMONITE: OK.
22:08   Is it just games?
22:09   Are they moving into different types of simulation?
22:12   JOHN GIANNANDREA: Well, there's a very fine line
22:14   between a video game-- a three-dimensional video game--
22:17   and a physics simulation already environment, right?
22:20   I mean, some video games are, in fact,
22:22   full simulations of worlds, so there's not really
22:26   a bright line there.
22:27   TOM SIMONITE: OK.
22:27   And do DeepMind work on robotics?
22:29   They don't, I didn't think.
22:30   JOHN GIANNANDREA: They're doing a bunch of work
22:32   in a bunch of different fields, some of which
22:33   gets published, some of which is not.
22:35   TOM SIMONITE: OK.
22:36   And the robot arms that we saw in the keynote on Wednesday,
22:40   are they within your group, Jeff?
22:41   JEFF DEAN: Yes.
22:42   TOM SIMONITE: OK.
22:42   So can you tell us about that project?
22:44   JEFF DEAN: Sure.
22:44   So that was a collaboration between our group
22:46   and the robotics teams in Google X. Actually, what happened was,
22:52   one of our researchers discovered
22:53   that the robotics team, actually,
22:55   had 20 unused arms sitting in a closet somewhere.
22:59   They were a model that was going to be discontinued
23:01   and not actually used.
23:02   So we're like, hey, we should set these up in a room.
23:06   And basically, just the idea of having
23:10   a little bit larger scale robotics test environment
23:12   than just one arm, which is what you typically
23:14   have in a physical robotics lab, would
23:18   make it possible to do a bit more exploratory research.
23:22   So one of the first things we did with that was just
23:24   have the robots learn to pick up objects.
23:27   And one of the nice properties that has,
23:29   it's a completely supervised problem.
23:32   The robot can try to grab something,
23:34   and if it closes its griper all the way, it failed.
23:36   And if it didn't close it all the way,
23:38   and it picked something up, it succeeded.
23:40   And so it's learning from raw camera pixel inputs
23:44   directly to torque motor controls.
23:45   And there's just a neural net there
23:47   that's trained to pick things up based on the observations it's
23:51   making of things as it approaches a particular object.
23:55   TOM SIMONITE: And is that quite a slow process?
23:57   I mean, that fact that you have multiple arms going
23:59   at once made me think that, maybe, you
24:02   were trying to maximize your throughput, or something.
24:04   JEFF DEAN: Right.
24:05   So if you have 20 arms, you get 20 times as much experience.
24:08   And if you think about how small kids learn to pick stuff up,
24:11   it takes them maybe a year, or something,
24:13   to go from being able to move their arm to really be
24:17   able to grasp simple objects.
24:19   And by parallelizing this across more arms,
24:22   you can pool the experience of the robotic arms a bit.
24:24   TOM SIMONITE: I see.
24:25   OK.
24:27   JEFF DEAN: And they need less sleep.
24:29   TOM SIMONITE: Right.
24:31   John, at the start of the session,
24:32   you referred to this concept of AI winter,
24:35   and you said you thought it was spring.
24:39   When do we know that it's summer?
24:43   JOHN GIANNANDREA: Summer follows spring.
24:45   I mean, there's still a lot of unsolved problems.
24:47   I think problems around dialogue and language
24:49   are the ones that I'm particularly interested in.
24:52   And so, until we can teach a computer to really read,
24:56   I don't think we can declare that it's summer.
24:59   I mean, if you can imagine a computer's really reading
25:02   and internalizing a document.
25:04   So it's interesting.
25:05   So translation is reading a paragraph in one language
25:08   and writing it in another language.
25:10   In order to do that really, really well,
25:12   you have to be able to paraphrase.
25:13   You have to be able to reorder words, and so on and so
25:15   forth So imagine translating something
25:17   from English to English.
25:18   So you read a paragraph, and you write a different paragraph.
25:21   If we could do that, I think I would declare summer.
25:25   TOM SIMONITE: OK.
25:26   Reading is-- well, there are different levels of reading,
25:30   aren't there?
25:31   Do you know--
25:33   JOHN GIANNANDREA: If you can paraphrase, then you really--
25:35   TOM SIMONITE: Then you think that-- if you
25:36   could reach that level.
25:37   JOHN GIANNANDREA: And actually understood--
25:37   TOM SIMONITE: Then you've got some argument.
25:39   JOHN GIANNANDREA: And to a certain extent,
25:40   today, our translation systems, which
25:42   are not perfect by any means, are getting better.
25:45   They do do some of that.
25:46   They do do some paraphrasing.
25:48   They do do some re-ordering.
25:49   They do do a remarkable amount of language understanding.
25:52   So I'm hopeful researchers around the world
25:54   will get there.
25:55   And it's very important to us that our natural language
25:57   APIs become part of our cloud platform,
25:59   and that people can experiment with it, and help.
26:02   JEFF DEAN: One thing I would say is,
26:04   I don't think there's going to be
26:05   this abrupt line between spring and summer, right?
26:08   There's going to be developments that push the state of the art
26:11   forward in lots of different areas in kind
26:13   of this smooth gradient of capabilities.
26:16   And at some point, something becomes
26:18   possible that didn't used to be possible,
26:21   and people kind of move the goalposts
26:23   of what they think of as really, truly hard problems.
26:28   APARNA CHENNAPRAGADA: The classic joke, right?
26:30   It's only AI until it starts working,
26:32   and then it's computer science.
26:34   JEFF DEAN: Like, if you'd asked me four years ago,
26:36   could a computer write a sentence
26:38   given an image as input?
26:40   And I would have said, I don't think they
26:42   can do that for a little while.
26:43   And they can actually do that today,
26:44   and that's kind of a good example of something
26:46   that has made a lot of progress in the last few years.
26:48   And now you sort of say, OK, that's in our tool
26:51   chest of capabilities.
26:53   TOM SIMONITE: OK.
26:53   But if we're not that great at predicting
26:56   how the progress goes, does that mean we can't see winter,
27:00   if it comes back?
27:04   JOHN GIANNANDREA: If we stop seeing progress,
27:06   then I think we could question what the future's going
27:09   to look like.
27:10   But today, the rate of-- I think researchers in the field
27:14   are excited about this, and maybe the field
27:16   is a little bit over-hyped because of the rate of progress
27:18   we're seeing.
27:19   Because something like speech recognition,
27:21   which didn't work for my wife five years ago,
27:23   and now works flawlessly, because image identification
27:29   is now working better than human raters for many fields.
27:32   So there's these narrow fields for which algorithms are not
27:36   superhuman in their capabilities.
27:37   So we're seeing tremendous progress.
27:39   And so it's very exciting for people working in this field.
27:42   TOM SIMONITE: OK.
27:43   Great.
27:44   I should just note that, in a couple of minutes,
27:46   we will open up the floor for questions.
27:48   There are microphones here and here in the main seating area,
27:52   and there's one microphone up in the press area, which
27:55   I can't see right now, but hopefully you
27:57   can figure out where it is.
28:01   Sundar Pichai, CEO of Google, has spoken a lot recently
28:04   about how he thinks we're moving from a world which
28:06   is mobile-first to AI-first.
28:11   I'm interested to hear what you think that means.
28:13   Maybe, Aparna, you could speak to that.
28:16   APARNA CHENNAPRAGADA: I interpret
28:18   it a couple different ways.
28:19   One is, if you look at how mobile's changed,
28:21   how you experience computing, it's
28:25   not happened at one level of the stack, right?
28:28   It's at the interface level, it's
28:29   at the information level, and infrastructure.
28:31   And I think that's the same thing that's
28:33   going to happen with AI and any of these machine learning
28:36   techniques, which is, you'll have infrastructure layer
28:39   improvements.
28:39   You saw the announcement about TPU.
28:41   You'll have a bunch of algorithms and models
28:44   improvements at the intelligence and information layer,
28:47   and there will be interface changes.
28:48   So the best UI is probably no UI.
28:51   TOM SIMONITE: Right.
28:52   OK.
28:53   John, what does AI-first mean to you?
28:57   JOHN GIANNANDREA: I think it means
28:58   that this assistant kind of layer is available to you
29:01   wherever you are.
29:02   Whether you're in your car, or whether it's
29:05   ambient in your house, or whether you're
29:07   using your mobile device or laptop,
29:10   that there is this smart assistance
29:12   that you find very quietly useful to you all the time.
29:17   Kind of how Google search is for most people today.
29:19   I think most people would not want search engines taken away
29:23   from them, right?
29:24   So I think that being that useful to people,
29:26   so that people take it for granted,
29:27   and then it's ambient across all your devices,
29:29   is what AI-first means to me.
29:31   TOM SIMONITE: And we're in the early stages of this,
29:33   do you think?
29:34   JOHN GIANNANDREA: Yeah.
29:35   It's a journey, I think.
29:36   It's a multi-year journey
29:37   TOM SIMONITE: OK.
29:38   Great.
29:39   So thanks for a fascinating conversation.
29:41   Now, we'll let someone else ask the questions for a little bit.
29:44   I will alternate between the press mic and the mics
29:49   down here at the front.
29:51   Please keep your questions short,
29:53   so we can get through more of them,
29:54   and make sure they're questions, not statements.
29:58   We will start with the press mic, wherever it is.
30:13   MALE SPEAKER: There's nobody there.
30:14   TOM SIMONITE: I really doubt the press has no questions.
30:18   What's happening?
30:18   Why don't we start with the developer mic
30:20   right here on the right?
30:23   AUDIENCE: I have a philosophical question about prejudice.
30:28   People tend to have prejudice.
30:31   Do you think this is a step stone
30:33   that we need to take in artificial intelligence,
30:36   and how would society accept that?
30:40   JOHN GIANNANDREA: I'm not sure I understand the question.
30:43   Some people have prejudice, and?
30:46   AUDIENCE: Some people have the tendency
30:49   to have prejudice, which might lead to behaviors
30:53   such as discrimination.
30:56   TOM SIMONITE: So the question is,
30:57   will the systems that the people build have biases?
31:00   JOHN GIANNANDREA: Oh, I see.
31:01   I see.
31:02   Will people's prejudices creep into machine learning systems?
31:05   I think that is a risk.
31:07   I think it all depends on the training data that we choose.
31:10   We've already seen some issues with this kind of problem.
31:13   So I think it all depends on carefully
31:14   selecting training data, particularly
31:16   for supervised systems.
31:19   TOM SIMONITE: OK.
31:21   Is the press mic working, at this point?
31:23   SEAN HOLLISTER: Hi.
31:24   I'm Sean Hollister, up here in the press mic.
31:26   TOM SIMONITE: Great.
31:27   Go for it.
31:28   SEAN HOLLISTER: Hi, there.
31:29   I wanted to ask about the role of privacy in machine learning.
31:33   You need a lot of data to make these observations
31:38   and to help people with machine learning.
31:41   I give all my photos to Google Photos,
31:44   and I wonder what happens to them afterwards.
31:47   What allows Google to see what they
31:49   are, and is that ever shared in any way with anyone else?
31:53   Personally, I don't care very much about that.
31:55   I'm not worried my photos are going
31:57   to get out to other folks, but where do they go?
32:00   What do you do with them?
32:01   And to what degree are they protected?
32:04   JEFF DEAN: Do you want to take that one?
32:06   APARNA CHENNAPRAGADA: I think this
32:07   is one of the most important things
32:09   that we look at across products.
32:12   So even with photos, or Google Now,
32:14   or voice, and all of these things.
32:16   There's actually two principles we codify into building this.
32:20   One is, there's a very explicit--
32:22   it's a very transparent contract between the user
32:25   and the product that is, you basically know what benefits
32:29   you're getting with the data, and the data
32:31   is there to help you.
32:32   That's one principle.
32:34   But the second is, by default, it's an opt-in experience.
32:39   You're in the driver's seat.
32:40   In some sense, let's say, you're saying,
32:42   hey, I do want to get traffic information when
32:45   I'm on Shoreline, because it's clogged up to Shoreline
32:48   Amphitheater, you, of course, need the system
32:50   to know where your location is.
32:51   Because you don't want to know how the traffic is in Napa.
32:55   So having that contract be transparent, but also
32:58   an opt-in, I think it really addresses the equation.
33:04   But I think the other thing to add in here
33:06   is also that, by definition, all of these are for your eyes
33:11   only, right?
33:12   In terms of, like, all your data is yours, and that's an axiom.
33:16   JOHN GIANNANDREA: And to answer his question,
33:18   we would never share his photos.
33:19   We train models based on other photos that are not yours,
33:24   and then the machine looks at your photos,
33:26   and it can label it, but we would never
33:27   share your private photo there.
33:29   SEAN HOLLISTER: To what degree is advertising
33:31   anonymously-targeted at folks like me,
33:34   based on the contents of things I upload,
33:37   little inferences you make in the meta data?
33:40   Is any of that going to advertisers in any way,
33:44   even in aggregate, hey, this is a person who
33:47   seems to like dogs?
33:50   JOHN GIANNANDREA: For your photos?
33:51   No.
33:52   Absolutely not.
33:52   APARNA CHENNAPRAGADA: No.
33:53   TOM SIMONITE: OK.
33:53   Let's go to this mic right here.
33:55   AUDIENCE: My questions is for Aparna, about,
33:58   what is the thought process behind creating a new product?
34:02   Because there are so many things that these guys are creating.
34:05   So how do you go from-- because it's kind of obvious right
34:08   now to see if you have my emails,
34:10   and you know that I'm traveling tomorrow to New York,
34:14   it's kind of simple to do that on my calendar
34:16   and create an event.
34:17   How do you go from robotic arms, trying
34:21   to understand how to get things, to an actual product?
34:25   The question is, what is the thought process behind it?
34:27   APARNA CHENNAPRAGADA: Yeah.
34:27   I'll give you the short version of it.
34:29   And, obviously, there's a longer version of it.
34:32   Wait for the medium post.
34:33   But I think the short version of it
34:35   is, to echo one thing JG said, you
34:38   want to pick problems that are easy for machines
34:41   and hard for humans.
34:42   So AI plus machine learning is not
34:45   going to turn a non-problem into a real problem
34:47   that people need solving.
34:49   It's like, you can take Christopher Nolan and Ben
34:53   Affleck, and you can still end up with Batman Versus Superman.
34:56   So you want to make sure that the problem you're solving
34:59   is a real one.
35:00   Many of our failures, even internally
35:02   and external, like frenzy around bots and AI,
35:06   is when you kid yourself that the problem needs solving.
35:09   And the second one, the second quick insight there,
35:12   is that you also want to build an iterative model.
35:15   That is, you want to kind of start small, and say, hey,
35:18   travel needs some assistance.
35:19   What are the top five things that people need help with?
35:22   And see which of these things can scale.
35:25   JEFF DEAN: I would add one thing to that,
35:26   which is, often, we're doing research
35:29   on a particular kind of problem.
35:31   And then, when we have something we think is useful,
35:34   we'll share that internally, as presentations or whatever,
35:37   and maybe highlight a few places where
35:39   we think this kind of technology could be used.
35:42   And that's sort of a good way to inform the product designers
35:45   about what kinds of things are now possible that
35:49   didn't used to be possible.
35:50   TOM SIMONITE: OK.
35:51   Let's have another question from the press section up there.
35:54   AUDIENCE: Yeah.
35:54   There's a lot of talk, lately, about sort of a fear of AI.
35:59   Elon Musk likened it to summoning the demon.
36:04   Whether that's overblown or not, whether it's
36:07   perception versus reality, there seems
36:10   to be a lot of mistrust or fear of going
36:13   too far in this direction.
36:15   How much stock you put into that?
36:18   And how do you win the trust of the public, when
36:22   you show experiments like the robot arm thing
36:24   on the keynote, which was really cool, but sort
36:26   of simultaneously creepy at the same time?
36:29   JOHN GIANNANDREA: So I get this question a lot.
36:31   I think there's this notion that's
36:34   been in the press for the last couple of years
36:36   about so-called super intelligence,
36:38   that somehow AI will beget more AI,
36:40   and then it will be exponential.
36:42   I think researchers in the field don't put much stock in that.
36:46   I don't think we think it's a real concern yet.
36:48   In fact, I think we're a long way away
36:49   from it being a concern.
36:51   There are some researchers who actually
36:53   think about these ethical problems,
36:55   and think about AI safety, and we
36:56   think that's really important.
36:57   And we work on this stuff with them,
37:00   and we support that kind of work.
37:01   But I think it's a concern that is decades and decades away.
37:06   It's also conflated with the fact
37:08   that people look at things like robots learning
37:10   to pick things up, and that's somehow
37:12   inherently scary to people.
37:14   I think it's our job, when we bring products
37:16   to market, to do it in a thoughtful way
37:19   that people find genuinely useful.
37:21   So a good example I would give you is, in Google products,
37:26   when you're looking for a place, like a coffee shop
37:28   or something, we'll show you when it's busy.
37:30   And that's the product of fairly advanced machine learning
37:34   that takes aggregate signals in a privacy-preserving way
37:36   and says, yeah, this coffee shop is really
37:38   busy on a Saturday morning.
37:39   That doesn't seem scary to me, right?
37:41   That doesn't seem anything like a bad thing
37:46   to bring into the world.
37:47   So I think there's a bit of a disconnect between the somewhat
37:50   extended hype, and the actual use of this technology
37:52   in everyday products.
37:54   TOM SIMONITE: OK.
37:54   Next question.
37:55   AUDIENCE: Thank you.
37:56   So given Google's source of revenue
37:58   and the high use of ad blockers, is there
38:02   any possibility of using machine learning
38:04   to maybe ensure that the appropriate ads are served?
38:07   Or if there's multiple versions of the same ad,
38:10   that the ad that would apply most to me
38:12   would be served to me, and to a different user,
38:14   a different version, and things like that?
38:16   Is that on the roadmap?
38:17   JEFF DEAN: Yeah.
38:18   I think, in general, there's a lot
38:20   of potential applications of machine
38:21   learning to advertising.
38:24   Google has actually been using machine
38:25   learning in our advertising system for more than a decade.
38:29   And I think one of the things about deciding
38:34   what ads to show to users is, you
38:35   want them to be relevant and useful to that user.
38:38   And it's better to not show an ad at all,
38:40   if you don't have something that seems plausibly relevant.
38:44   And that's always been Google's advertising philosophy.
38:47   And other websites on the web don't necessarily quite
38:51   have the same balance, in that respect.
38:53   But I do think there's plenty of opportunity to continue
38:56   to improve advertising systems and make them better,
38:59   so that you see less ads, but they're actually more useful.
39:03   TOM SIMONITE: OK.
39:03   Next question from at the top.
39:05   JACK CLARK: Jack Clark with Bloomberg News.
39:08   So how do you differentiate to the user
39:12   between a sponsored advert, and one that is provided by your AI
39:17   naturally?
39:18   How do I know that the burger joint you're suggesting
39:21   is like a paid-for link, or is it a genuine link?
39:26   JEFF DEAN: So in our user interfaces,
39:27   we always clearly delimit advertisements.
39:30   And in general, all ads that we show
39:33   are selected algorithmically by our systems.
39:36   They're not like, you can just give us an ad,
39:38   and we will always show it to someone.
39:40   We always decide what is the likelihood
39:43   that this ad is going to be useful to someone,
39:45   before we decide to show that advertiser's ad.
39:48   JACK CLARK: Does this extend to stuff like Google Home, where
39:51   it will say, this is a sponsored restaurant
39:53   we're going to send you to.
39:57   JEFF DEAN: I don't know that product.
39:58   JOHN GIANNANDREA: I mean, we haven't
40:00   launched Google Home yet.
40:01   So a lot of these product decisions are still to be made.
40:05   I think we do, as a general rule,
40:08   clearly identify when something is sponsored
40:10   versus when it's organic.
40:13   TOM SIMONITE: OK.
40:13   Next question here.
40:15   AUDIENCE: Hi.
40:15   This is a question for Jeff Dean.
40:19   I'm very much intrigued by the Google Brain project
40:22   that you're doing.
40:22   Very cool t-shirt.
40:25   The question is, what is the road map of that,
40:28   and how does it relate to the point of singularity?
40:32   JEFF DEAN: Aha.
40:34   So the road map of-- this is sort of the project code name
40:41   for the team that I work on.
40:43   Basically, the team was developed
40:45   to investigate the use of advanced methods
40:49   in machine learning to solve difficult problems in AI.
40:54   And we're continuing to work on pushing the state
40:57   of the art in that area.
40:59   And I think that means working in lots of different areas,
41:01   building the right kinds of hardware with TPUs,
41:04   building the right systems infrastructure with things
41:07   like TensorFlow.
41:08   Solving the right research problems
41:10   that are not connected to products,
41:14   and then figuring out ways in which machine learning can
41:17   be used to advance different kinds of fields,
41:22   as we solve different problems along the road.
41:25   I'm not a big believer in the singularity.
41:27   I think all exponentials look like exponentials
41:30   at the beginning, but then they run out of stuff.
41:34   TOM SIMONITE: OK.
41:35   Thanks for the question.
41:36   Back to the pressbox.
41:38   STEVEN MAX PATTERSON: Hi.
41:39   Steven Max Patterson, IDG.
41:41   I was looking at Google Home and Google Assistant,
41:45   and it looks like it's really a platform.
41:50   And it's a composite of other platforms,
41:53   like the Knowledge Graph, Google Cloud Speech, Google machine
41:58   learning, the Awareness API.
42:00   Is this a feature that other consumer device manufacturers
42:06   could include, and is that the intent and direction of Google,
42:09   is to make this a platform?
42:13   JOHN GIANNANDREA: It's definitely
42:14   the case that most of our machine learning APIs
42:18   are migrating to the cloud platform, which enables people
42:21   to use, for example, our speech capabilities in other products.
42:25   I think the Google Assistant is intended to be, actually,
42:27   a holistic product delivered from Google.
42:29   That makes sense.
42:30   But it may make sense to syndicate
42:32   that to other manufacturers at some point.
42:34   We don't have any plans to do that today.
42:36   But in general, we're trying to be
42:37   as open as we can with the component pieces
42:39   that you just mentioned, and make
42:41   them available as Cloud APIs, and in many cases,
42:43   as open source solutions as well.
42:45   JEFF DEAN: Right.
42:46   I think one of the things about that
42:47   is, making those individual pieces available
42:49   enables everyone in the world to take advantage of some
42:53   of the machine learning research we've done,
42:55   and be able to do things like label images,
42:57   or do speech recognition really well.
42:59   And then they can go off and build
43:00   really cool, amazing things that aren't necessarily
43:03   the kinds of things we're working on.
43:05   JOHN GIANNANDREA: Yeah, and many companies are doing this today.
43:07   They're using our translate APIs.
43:08   They're using our Cloud Speech APIs today.
43:12   TOM SIMONITE: Right.
43:13   We have time for one last quick question from this mic here.
43:15   AUDIENCE: Hi.
43:16   I'm [INAUDIBLE].
43:18   John, you said that you would declare summer
43:22   if, in language understanding, it
43:25   would be able to translate from one paragraph in English
43:30   to another paragraph in English.
43:32   Don't you think that making that possible requires
43:35   really complete understanding of the world, and everything
43:40   that's going on, just to catch the emotional level that
43:44   is in the paragraph, or even the physical understanding
43:47   of the world around us?
43:50   JOHN GIANNANDREA: Yeah, I do.
43:52   I use that example because it is really, really hard.
43:55   So I don't think we're going to be done for many, many years.
43:58   I think there's a lot of work to do.
44:00   We built the Google Knowledge Graph, in part,
44:02   to answer that question, so that we actually
44:03   had some semantic understanding of at least
44:05   the things in the world, and some of the relationships
44:07   between them.
44:08   But yeah, it's a very hard problem.
44:09   And I used that example because it's
44:11   pretty clear we won't be done for a long time.
44:13   TOM SIMONITE: OK.
44:14   Sorry, there's no time for other questions.
44:16   Thanks for the question.
44:17   A good forward-looking note to end on.
44:20   We'll see how it works out over the coming years.
44:23   Thank you for joining me, all of you on stage,
44:25   and thanks for the questions and coming for the session.
44:29   [MUSIC PLAYING]
Transcripción : Youtube