Placeholder Image

字幕列表 影片播放

  • JEFF DEAN: I'm really excited to be here.

  • I think it was almost four years ago to the day

  • that we were about 20 people sitting in a small conference

  • room in one of the Google buildings.

  • We've woken up early because we wanted to kind of time

  • this for an early East Coast launch where

  • we were turning on the TensorFlow.org website

  • and releasing the first version of TensorFlow

  • as an open source project.

  • And I'm really, really excited to see what it's become.

  • It's just remarkable to see the growth and all

  • the different kinds of ways in which people have used

  • this system for all kinds of interesting things

  • around the world.

  • So one thing that's interesting is the growth

  • in the use of TensorFlow also kind

  • of mirrors the growth in interest in machine learning

  • and machine learning research generally around the world.

  • So this is a graph showing the number of machine

  • learning archive papers that have been posted

  • over the last 10 years or so.

  • And you can see it's growing quite, quite rapidly, much more

  • quickly than you might expect.

  • And that lower red line is kind of the nice doubling

  • every couple of years growth rate, exponential growth

  • rate we got used to in computing power, due to Moore's law

  • for so many years.

  • That's now kind of slowed down.

  • But you can see that the machine learning research community

  • is generating research ideas at faster than that rate, which

  • is pretty remarkable.

  • We've replaced computational growth with growth of ideas,

  • and we'll see those both together will be important.

  • And really, the excitement about machine learning

  • is because we can now do things we couldn't do before, right?

  • As little as five or six years ago, computers really

  • couldn't see that well.

  • And starting in about 2012, 2013,

  • we started to have people use deep neural networks to try

  • to tackle computer vision problems, image

  • classification, object detection, things like that.

  • And so now, using deep learning and deep neural networks,

  • you can feed in the raw pixels of an image

  • and fairly reliably get a prediction of what kind

  • of object is in that image.

  • Feed in the pixels there.

  • Red, green, and blue values in a bunch of different

  • coordinates, and you get out the prediction leopard.

  • This works for speech as well.

  • You can feed an audio wave forms,

  • and by training on lots of audio wave forms and transcripts

  • of what's being said in those wave forms,

  • we can actually take a completely new recording

  • and tell you what is being said amid a transcript.

  • Bonjour, comment allez-vous?

  • You can even combine these ideas and have

  • models that take in pixels, and instead of just predicting

  • classifications of what are in the object,

  • it can actually write a short sentence, a short caption,

  • that a human might write about the image--

  • a cheetah lying on top of a car.

  • That's one of my vacation photos, which was kind of cool.

  • And so just to show the progress in computer vision, in 2011,

  • Stanford hosts an ImageNet contest every year

  • to see how well computer vision systems can

  • predict one of 1,000 categories in a full color image.

  • And you get about a million images to train on,

  • and then you get a bunch of test images

  • your model has never seen before.

  • And you need to make a prediction.

  • In 2011, the winning entrant got 26% error, right?

  • So you can kind of make out what that is.

  • But it's pretty hard to tell.

  • We know from human experiment that human error

  • of a well-trained human, someone who's

  • practiced at this particular task

  • and really understands 1,000 categories,

  • gets about 5% error.

  • So this is not a trivial task.

  • And in 2016, the winning entrant got 3% error.

  • So just look at that tremendous progress

  • in the ability of computers to resolve and understand

  • computer imagery and have computer vision

  • that actually works.

  • This is remarkably important in the world,

  • because now we have systems that can perceive

  • the world around us and we can do all kinds of really

  • interesting things about.

  • We've seen similar progress in speech recognition and language

  • translation and things like that.

  • So for the rest of the talk, I'd like to kind of structure it

  • around this nice list of 14 challenges

  • that the US National Academy of Engineering

  • put out and felt like these were important things

  • for the science and engineering communities

  • to work on for the next 100 years.

  • They put this out in 2008 and came up

  • with this list of 14 things after some deliberation.

  • And I think you'll agree that these

  • are sort of pretty good large challenging problems,

  • that if we actually make progress

  • on them, that we'll actually have

  • a lot of progress in the world.

  • We'll be healthier.

  • We'll be able to learn things better.

  • We'll be able to develop better medicines.

  • We'll have all kinds of interesting energy solutions.

  • So I'm going to talk about a few of these.

  • And the first one I'll talk about

  • is restoring and improving urban infrastructure.

  • So we're on the cusp of the sort of widespread commercialization

  • of a really interesting new technology that's

  • going to really change how we think about transportation.

  • And that is autonomous vehicles.

  • And this is a problem that has been worked on

  • for quite a while, but it's now starting

  • to look like it's actually completely

  • possible and commercially viable to produce these things.

  • And a lot of the reason is that we now

  • have computer vision and machine learning techniques

  • that can take in sort of raw forms of data

  • that the sensors on these cars collect.

  • So they have the spinning LIDARs on the top that

  • give them 3D point cloud data.

  • They have cameras in lots of different directions.

  • They have radar in the front bumper and the rear bumper.

  • And they can really take all this raw information in,

  • and with a deep neural network, fuse

  • it all together to build a high level understanding of what

  • is going on around the car.

  • Or is it another car to my side, there's a pedestrian

  • up here to the left, there's a light post over there.

  • I don't really need to worry about that moving.

  • And really help to understand the environment in which

  • they're operating and then what actions can

  • they take in the world that are both legal, safe,

  • obey all the traffic laws, and get them from A to B.

  • And this is not some distant far-off dream.

  • Alphabet's Waymo subsidiary has actually

  • been running tests in Phoenix, Arizona.

  • Normally when they run tests, they

  • have a safety driver in the front seat,

  • ready to take over if the car does

  • something kind of unexpected.

  • But for the last year or so, they've

  • been running tests in Phoenix with real passengers

  • in the backseat and no safety drivers in the front seat,

  • running around suburban Phoenix.

  • So suburban Phoenix is a slightly easier training ground

  • than, say, downtown Manhattan or San Francisco.

  • But it's still something that is like not really far off.

  • It's something that's actually happening.

  • And this is really possible because

  • of things like machine learning and the use

  • of TensorFlow in these systems.

  • Another one that I'm really, really excited

  • about is advance health informatics.

  • This is a really broad area, and I

  • think there's lots and lots of ways

  • that machine learning and the use of health data

  • can be used to make better health care

  • decisions for people.

  • So I'll talk about one of them.

  • And really, I think the potential here

  • is that we can use machine learning

  • to bring the wisdom of experts through a machine learning

  • model anywhere in the world.

  • And that's really a huge, huge opportunity.

  • So let's look at this through one problem

  • we've been working on for a while, which

  • is diabetic retinopathy.

  • So diabetic retinopathy is the fastest growing cause

  • of preventable blindness in the world.

  • And screening every year, if you're at risk for this,

  • and if you have diabetes or early sort of symptoms that

  • make it likely you might develop diabetes, you should really get

  • screened every year.

  • So there's 400 million people around the world that

  • should be screened every year.

  • But the screening is really specialized.

  • Doctors can't do it.

  • You really need ophthalmologist level of training

  • in order to do this effectively.

  • And the impact of the shortage is significant.

  • So in India, for example, there's

  • a shortage of 127,000 eye doctors

  • to do this sort of screening.

  • And as a result, 45% of patients who

  • are diagnosed to this disease actually

  • have suffered either full or partial vision loss

  • before they're actually diagnosed and then treated.

  • And this is completely tragic because this disease,

  • if you catch it in time, is completely treatable.

  • There's a very simple 99% effective treatment

  • that we just need to make sure that the right people get

  • treated at the right time.

  • So what can you do?

  • So, it turns out diabetic retinopathy screening is also

  • a computer vision problem, and the progress

  • we've made on general computer vision problems

  • where you want to take a picture and tell if that's

  • a leopard or an aircraft carrier or a car

  • actually also works for diabetic retinopathy.

  • So you can take a retinal image, which

  • is what the screening camera, sort of the raw data that

  • comes off the screening camera, and try

  • to feed that into a model that predicts 1, 2, 3, 4, or 5.

  • That's how these things are graded,

  • 1 being no diabetic retinopathy, 5 being proliferative,

  • and the other numbers being in between.

  • So it turns out you can get a collection of data

  • of retinal images and have ophthalmologists label them.

  • Turns out if you ask two ophthalmologists

  • to label the same image, they agree

  • with each other 60% of the time on the number 1, 2, 3, 4, or 5.

  • But perhaps slightly scarier if you

  • ask the same ophthalmologist to grade the same image

  • a few hours apart, they agree with themselves 65%

  • of the time.

  • But you can fix this by actually getting each image labeled

  • by a lot of ophthalmologists, so you'll

  • get it labeled by seven ophthalmologists.

  • If five of them say it's a 2, and two of them say it's a 3,

  • it's probably more like a 2 than a 3.

  • Eventually, you have a nice, high quality

  • data set you can train on.

  • Like many machine learning problems,

  • high quality data is the right raw ingredient.

  • But then you can apply, basically,

  • an off-the-shelf computer vision model trained on this data set.

  • And now you can get a model that is

  • on par or perhaps slightly better than the average board

  • certified ophthalmologist in the US, which is pretty amazing.

  • It turns out you can actually do better than that.

  • And if you get the data labeled by retinal specialists, people

  • who have more training in retinal disease

  • and change the protocol by which you label things,

  • you get three retinal specialists

  • to look at an image, discuss it amongst themselves,

  • and come up with what's called a sort of coordinated assessment

  • and one number.

  • Then you can train a model and now

  • be on par with retinal specialists, which

  • is kind of the gold standard of care in this area.

  • And that's something you can now take and distribute widely

  • around the world.

  • So one issue particularly with health care kinds of problems

  • is you want explainable models.

  • You want to be able to explain to a clinician

  • why is this person, why do we think this person has

  • moderate diabetic retinopathy.

  • So you can take a retinal image like this,

  • and one of the things that really helps

  • is if you can show in the model's assessment

  • why this is a 2 and not a 3.

  • And by highlighting parts of the input data,

  • you can actually make this more understandable for clinicians

  • and enable them to really sort of get behind the assessment

  • that the model is making.

  • And we've seen this in other areas as well.

  • There's been a lot of work on explainability,

  • so I think the notion that deep neural networks are

  • sort of complete black boxes is a bit overdone.

  • There's actually a bunch of good techniques

  • that are being developed and more

  • all the time that will improve this.

  • So a bunch of advances depend on being able to understand text.

  • And we've had a lot of really good improvements

  • in the last few years on language understanding.

  • So this is a bit of a story of research

  • and how research builds on other research.

  • So in 2017, a collection of Google researchers and interns

  • came up with a new kind of model for text called the Transformer

  • model.

  • So unlike recurrent models where you

  • have kind of a sequential process where

  • you absorb one word or one token at a time

  • and update some internal state and then go on

  • to the next token, the Transformer model

  • enables you to process a whole bunch of text, all at once

  • in parallel, making it much more computationally efficient,

  • and then to use attention on previous texts

  • to really focus on if I'm trying to predict what

  • the next word is, what are other parts of the context

  • to the left that are relevant to predicting that?

  • So that paper was quite successful

  • and showed really good results on language translation tasks

  • with a lot less compute.

  • So the blue score there and the first two columns

  • for English to German and English to French, higher

  • is better.

  • And then the compute cost of these models

  • shows that this is getting sort of state of the art

  • results at that time, with 10 to 100x less compute

  • than other approaches.

  • Then in 2018, another team of Google researchers

  • built on the idea of Transformers.

  • So everything you see there in a blue oval

  • is a Transformer module, and they

  • came up with this approach called Bidirectional Encoding

  • Representations from Transformers, or BERT.

  • It's a little bit shorter and more catchy.

  • So BERT has this really nice property

  • that, in addition to using context to the left,

  • it uses context all around the language, sort

  • of the surrounding text, in order

  • to make predictions about text.

  • And the way it works is you start

  • with a self-supervised objective.

  • So the one really nice thing about this

  • is there's lots and lots of text in the world.

  • So if you can figure out a way to use that text

  • to train a model to be able to understand text better,

  • that would be great.

  • So we're going to take this text,

  • and in the BERT training objective,

  • to make it self-supervised, we're going to drop about 15%

  • of the words.

  • And this is actually pretty hard,

  • but the model is then going to try to fill in the blanks,

  • essentially.

  • Try to predict what are the missing

  • words that were dropped.

  • And because we actually have the original words,

  • we now know if the model is correct in its guesses

  • about what goes in the box.

  • And by processing trillions of words of text like this,

  • you actually get a very good understanding

  • of contextual cues in language and how

  • to actually fill in the blanks in a really intelligent way.

  • And so that's essentially the training objective for BERT.

  • You take text, you drop 15% of it,

  • and then you try to predict those missing words.

  • And one key thing that works really well is that step one.

  • You can pre-train a model on lots and lots of text,

  • using this fill-in-the-blank self-supervised objective

  • function.

  • And then step two, you can then take a language task

  • you really care about.

  • Like maybe you want to predict, is this a five-star review

  • or a one-star review for some hotel,

  • but you don't have very much labeled

  • text for that actual task.

  • You might have 10,000 reviews and know

  • the star count of each review.

  • But you can then finetune the model,

  • starting with the model trained in step one

  • on trillions of words of text and now use

  • your paltry 10,000 examples for the text task

  • you really care about.

  • And that works extremely well.

  • So in particular, BERT gave state-of-the-art results across

  • a broad range of different text understanding benchmarks

  • in this GLUE benchmark suite, which was pretty cool.

  • And people have been using BERT now

  • in this way to improve all kinds of different things

  • all across the language understanding and NLP space.

  • So one of the grand challenges was engineer the tools

  • of scientific discovery.

  • And I think it's pretty clear machine learning is actually

  • going to be an important component of making advances

  • in a lot of these other grand challenge areas,

  • things like autonomous vehicles or other kinds of things.

  • And it's been really satisfying to see what we'd hoped

  • would happen when we released TensorFlow as an open source

  • project has actually kind of come to pass,

  • as we were hoping, in that lots of people

  • would sort of pick up TensorFlow,

  • use it for all kinds of things.

  • People would improve the core system.

  • They would use it for tasks we would never imagine.

  • And that's been quite satisfying.

  • So people have done all kinds of things.

  • Some of these are uses inside of Google.

  • Some are outside in academic institutions.

  • Some are scientists working on conserving whales

  • or understanding ancient scripts,

  • many kinds of things, which is pretty neat.

  • The breadth of uses is really amazing.

  • These are the 20 winners of the Google.org AI Impact Challenge,

  • where people could submit proposals

  • for how they might use machine learning

  • and AI to really tackle a local challenge they

  • saw in their communities.

  • And they have all kinds of things,

  • ranging from trying to predict better ambulance dispatching

  • to identifying sort of illegal logging using

  • speech recognition or audio processing.

  • Pretty neat.

  • And many of them are using TensorFlow.

  • So one of the things we're pretty excited about

  • is AutoML, which is this idea of automating

  • some of the process by which machine

  • learning experts sit down and sort of make decisions to solve

  • machine learning problems.

  • So currently, you have a machine learning expert sit down,

  • they take data, they have computation.

  • They run a bunch of experiments.

  • They kind of stir it all together.

  • And eventually, you get a solution

  • to a problem you actually care about.

  • One of the things we'd like to be able to do,

  • though, is see if we could eliminate a lot of the need

  • for the human machine learning expert to run these experiments

  • and instead, automate the experimental process by which

  • a machine learning expert comes by a high quality

  • solution for a problem you care about.

  • So lots and lots of organizations

  • around the world have machine learning problems,

  • but many, many of them don't even

  • realize they have a machine learning problem,

  • let alone have people in their organization

  • that can tackle the problem.

  • So one of the earliest pieces of work

  • our researchers did in the space was something called

  • neural architecture search.

  • So when you sit down and design a neural network

  • to tackle a particular task, you make

  • a lot of decisions about shapes of this or that,

  • and should it be used 3 by 3 filters at layer 17 or 5

  • by 5, all kinds of things like this.

  • It turns out you can automate this process

  • by having a model generating model

  • and train the model generating model based on feedback

  • about how well the models that it generates

  • work on the problem you care about.

  • So the way this will work, we're going

  • to generate a bunch of models.

  • Those are just descriptions of different neural network

  • architectures.

  • We're going to train each of those for a few hours,

  • and then we're going to see how well they work.

  • And then use the accuracy of those models

  • as a reinforcement learning signal for the model generating

  • model, to steer it away from models

  • that didn't work very well and towards models

  • that worked better.

  • And we're going to repeat many, many times.

  • And over time, we're going to get better and better

  • by steering the search to the parts of the space of models

  • that worked well.

  • And so it comes up with models that look a little strange,

  • admittedly.

  • A human probably would not sit down and wire up

  • a sort of machine learning, computer vision model exactly

  • that way.

  • But they're pretty effective.

  • So if you look at this graph, this

  • shows kind of the best human machine learning experts,

  • computer vision experts, machine learning researchers

  • in the world, producing a whole bunch of different kinds

  • of models in the last four or five years,

  • things like ResNet 50, DenseNet-201,

  • Inception-ResNet, all kinds of things.

  • That black dotted line is kind of the frontier

  • of human machine learning expert model

  • quality on the y-axis and computational cost

  • on the x-axis.

  • So what you see is as you go out the x-axis,

  • you tend to get more accuracy because you're applying

  • more computational cost.

  • But what you see is the blue dotted line

  • is AutoML-based solutions, systems where we've

  • done this automated experimentation instead

  • of pre-designing any particular architecture.

  • And you see that it's better both at the high end, where

  • you care about the most accurate model you can get,

  • regardless of computational cost,

  • but it's also accurate at the low end, where

  • you care about a really lightweight model that

  • might run on a phone or something like that.

  • And in 2019, we've actually been able to improve that

  • significantly.

  • This is a set of models called Efficient Net

  • and it has a very kind of a slider

  • about you can trade off computational cost

  • and accuracy.

  • But they're all way better than human sort

  • of guided experimentation on the black dotted line there.

  • And this is true for image recognition, for [INAUDIBLE]..

  • It's true for object detection.

  • So the red line there is AutoML.

  • The other things are not.

  • It's true for language translation.

  • So the black line there is various kinds of Transformers.

  • The red line is we gave the basic components

  • of Transformers to an AutoML system

  • and allowed it to fiddle with it and come up

  • with something better.

  • It's true for computer vision models

  • used in autonomous vehicles.

  • So this was a collaboration between Waymo and Google

  • Research.

  • We were able to come up with models that were significantly

  • lower latency for the same quality,

  • or they could trade it off and get significantly lower error

  • rate at the same latency.

  • It actually works for tabular data.

  • So if you have lots of customer records,

  • and you want to predict which customers

  • are going to be spending $1,000 with your business next month,

  • you can use AutoML to come up with a high quality

  • model for that kind of problem.

  • OK.

  • So what do we want?

  • I think we want the following properties

  • in a machine learning model.

  • So one is we tend to train separate models

  • for each different problem we care about.

  • And I think this is a bit misguided.

  • Like, really, we want one model that does a lot of things

  • so that it can build on the knowledge in how it does

  • thousands or millions of different things,

  • so that when the million and first thing comes along,

  • it can actually use its expertise from all

  • the other things it knows how to do to know how to get

  • into a good state for the new problem

  • with relatively little data and relatively

  • little computational cost.

  • So these are some nice properties.

  • I have kind of a cartoon diagram of something

  • I think might make sense.

  • So imagine we have a model like this where it's very sparsely

  • activated, so different pieces of the model

  • have different kinds of expertise.

  • And they're called upon when it makes sense,

  • but they're mostly idle, so it's relatively computationally

  • [INAUDIBLE] power efficient.

  • But it can do many things.

  • And now, each component here is some piece

  • of machine learning model with different kinds of state,

  • parameters in the model, and different operations.

  • And a new task comes along.

  • Now you can imagine something like neural architecture search

  • becoming-- squint at it just right

  • and now turn it into neural pathway search.

  • We're going to look for components

  • that are really good for this new task we care about,

  • and maybe we'll search and find that this path

  • through the model actually gets us

  • into a pretty good state for this new task.

  • Because maybe it goes through components

  • that are trained on related tasks already.

  • And now maybe we want that model to be

  • more accurate for the purple task,

  • so we can add a bit more computational capacity,

  • add a new component, start to use

  • that component for this new task, continue training it,

  • and now, that new component can also

  • be used for solving other related tasks.

  • And each component itself might be

  • running some sort of interesting architectural search inside it.

  • So I think something like that is the direction we should

  • be exploring as a community.

  • It's not what we're doing today, but I

  • think it could be a pretty interesting direction.

  • OK, and finally, I'd like to touch on thoughtful use of AI

  • in society.

  • As we've seen more and more uses of machine learning

  • in our products and around the world,

  • it's really, really important to be

  • thinking carefully about how we want

  • to apply these technologies.

  • Like any technology, these systems

  • can be used for amazing things or things

  • we might find a little detrimental in various ways.

  • And so we've come up with a set of principles by which we think

  • about applying sort of machine learning and AI

  • to our products.

  • And we've made these public about a year and a half

  • ago as a way of sort of sharing our thought process

  • with the rest of the world.

  • And I particularly like these.

  • I'll point out many of these are sort of areas of research

  • that are not fully understood yet,

  • but we aim to apply the best in the state of the art methods,

  • for example, for reducing bias in machine learning models,

  • but also continue to do research and advance the state

  • of the art in these areas.

  • And so this is just kind of a taste of different kinds

  • of work we're doing in this area--

  • how do we do machine learning with more privacy,

  • using things like federated learning?

  • How do we make models more interpretable

  • so that a clinician can understand

  • the predictions it's making on diabetic retinopathy

  • sort of examples?

  • How do we make machine learning more fair?

  • OK, and with that, I hope I've convinced

  • you that deep neural nets and machine learning--

  • you're already here, so maybe you're

  • already convinced of this-- but are

  • helping make sort of significant advances

  • in a lot of hard computer science problems, computer

  • vision, speech recognition, language understanding.

  • General use of machine learning is going

  • to push the world forward.

  • So thank you very much, and I appreciate you all being here.

  • [APPLAUSE]

  • MEGAN KACHOLIA: Hey, everyone.

  • Good morning.

  • Just want to say, first of all, welcome.

  • Today, I want to talk a little bit

  • about TensorFlow 2.0 and some of the new updates

  • that we have that are going to make your experience

  • with TensorFlow even better.

  • But before I dive into a lot of those details,

  • I want to start off by thanking you, everyone here, everyone

  • on the livestream, everyone who's

  • been contributing to TensorFlow, all of you who

  • make up the community.

  • TensorFlow was open source to help accelerate the AI

  • field for everyone.

  • You've used it in your experiments.

  • You've deployed in your businesses.

  • You've made some amazing different applications

  • that we're so excited to showcase and talk about, some

  • that we get to see a bit here today,

  • which is one of my favorite parts about conferences

  • like this.

  • And you've done so much more.

  • And all of this has helped make TensorFlow what it is today.

  • It's the most popular ML ecosystem in the world.

  • And honestly, that would not happen

  • without the community being excited and embracing and using

  • this and giving back.

  • So on behalf of the entire TensorFlow team,

  • I really just first want to say thank you

  • because it's so amazing to see how TensorFlow is used.

  • That's one of the greatest things

  • I get to see about my job, is the applications

  • and the way folks are using TensorFlow.

  • I want to take a step back and talk a little bit about some

  • of the different user groups and how we see

  • them making use of TensorFlow.

  • TensorFlow was being used across a wide range of experiments

  • and applications.

  • So here, calling out researchers, data scientists

  • and developers, and there's other groups kind of in-between

  • as well.

  • Researchers use it because it's flexible.

  • It's flexible enough to experiment with and push

  • the state-of-the-art deep learning.

  • You heard this even just a few minutes ago,

  • with folks from Twitter talking about how they're

  • able to use TensorFlow and expand on top of it

  • in order to do some of the amazing things

  • that they want to make use of on their own platform.

  • And at Google, we see examples of this when researchers

  • are creating advanced models like Excel NAT

  • and some of the other things that Jeff

  • referenced in his talk earlier.

  • Taking a step forward, looking at data scientists,

  • data scientists and enterprise engineers

  • have said they rely on TensorFlow

  • for performance and scale in training and production

  • environments.

  • That's one of the big things about TensorFlow

  • that we've always emphasized and looked at from the beginning.

  • How can we make sure this can scale to large production use

  • cases?

  • For example, Quantify and BlackRock

  • use TensorFlow to test and deploy

  • BERT in real world NLP instances,

  • such as text tokenization, as well as classification.

  • Hopping one step forward, looking a bit at application

  • developers, application developers

  • use TensorFlow because it's easy to learn ML on the platforms

  • that they care about.

  • Arduino wants to make ML simple on microcontrollers,

  • so they rely on TensorFlow pre-trained models

  • and TensorFlow Lite Micro for deployment.

  • Each of these groups is a critical part

  • of the TensorFlow ecosystem.

  • And this is why we really wanted to make sure that TensorFlow

  • 2.0 works for everyone.

  • We announced the alpha at our Dev Summit earlier this year.

  • And over the past few months, the team

  • has been working very hard to incorporate early feedback.

  • Again, thank you to the community

  • for giving us that early feedback,

  • so we can make sure we're developing something

  • that works well for you.

  • And we've been working to resolve bugs and issues

  • and things like that.

  • And just last month in September,

  • we were excited to announce the final general release

  • for TensorFlow 2.0.

  • You might be familiar with TensorFlow's architecture,

  • which has always supported the ML lifecycle from training

  • through deployment.

  • Again, one of the things we've emphasized

  • since the beginning when TensorFlow was initially open

  • sourced a few years ago.

  • But I want to emphasize how TensorFlow 2.0 makes

  • this workflow even easier and more intuitive.

  • First, we invested in Keras, an easy-to-use package

  • in TensorFlow, making it the default high level API.

  • Many developers love Keras because it's

  • easy to use and understand.

  • Again, you heard this already mentioned a little bit earlier,

  • and hopefully, we'll hear more about it

  • throughout the next few days.

  • By tightly integrating Keras into 2.0,

  • we can make Keras work even better with primitives

  • like TF data.

  • We can do performance optimizations behind the scenes

  • and run distributed training.

  • Again, we really wanted 2.0 to focus on usability.

  • How can we make it easier for developers?

  • How can we make it easier for users to get what they need out

  • of TensorFlow?

  • For instance, Lose It, a customized weight loss app,

  • said they use tf.keras for designing their network.

  • By leveraging [INAUDIBLE] strategy distribution in 2.0,

  • they were able to utilize the full power of their GPUs.

  • It's feedback like this that we love to hear,

  • and again, it's very important for us

  • to know how the community is making use of things, how

  • the community is using 2.0, the things they want to see,

  • so that we can make sure we're developing the right framework

  • and also make sure you can contribute back.

  • When you need a bit more control to create advanced algorithms,

  • 2.0 comes fully loaded with eager execution,

  • making it familiar for Python developers.

  • This is especially useful when you're stepping through, doing

  • debugging, making sure you can really understand step

  • by step what's happening.

  • This also means there's less coding

  • required when training your model,

  • all without having to use session.run.

  • Again, usability is a focus.

  • To demonstrate the power of training models with 2.0,

  • I'll show you how you can train a state-of-the-art NLP model

  • in 10 lines of code, using the Transformers NLP library

  • by Hugging Face-- again, a community contribution.

  • This popular package hosts some of the most advanced NLP

  • models available today, like BERT, GPT, Transformer-XL,

  • XLNet, and now supports TensorFlow 2.0.

  • So let's take a look.

  • Here, kind of just looking through the code,

  • you can see how you can use 2.0 to train

  • Hugging Face's DistilBERT model for text classification.

  • You can see just simply load the tokenizer, model,

  • and the data set.

  • Then prepare the data set and use tf.keras compile and fit

  • APIs.

  • And with a few lines of code, I can now train my model.

  • And with just a few more lines, we

  • can use the train model for tasks

  • such as text classification using eager execution.

  • Again, it's examples like this where

  • we can see how the community takes something and is

  • able to do something very exciting and amazing by making

  • use of the platform and the ecosystem

  • that TensorFlow is providing.

  • But building and training a model

  • is only one part of TensorFlow 2.0.

  • You need the performance to match.

  • That's why we worked hard to continue to improve performance

  • with TensorFlow 2.0.

  • It delivers up to 3x faster training performance

  • using mixed precision on NVIDIA Volta and Turing GPUs

  • in a few lines of code with models like ResNet-50 and BERT.

  • As we continue to double down on 2.0 in the future,

  • performance will remain a focus with more models

  • and with hardware accelerators.

  • For example, in 2.1, so the next upcoming TensorFlow release,

  • you can expect TPU and TPU pod support, along

  • with mixed precision for GPUs.

  • So performance is something that we're

  • keeping a focus on as well, while also making

  • sure usability really stands to the forefront.

  • But there's a lot more to the ecosystem.

  • So beyond model building and performance,

  • there are many other pieces that help round

  • out the TensorFlow ecosystem.

  • Add-ons and extensions are a very important piece here,

  • which is why we wanted to make sure that they're also

  • compatible with TensorFlow 2.0.

  • So you can use popular libraries,

  • like some other ones called out here,

  • whether it's TensorFlow Probability, TF Agents, or TF

  • Text.

  • We've also introduced a host of new libraries

  • to help researchers and ML practitioners

  • in more useful ways.

  • So for example, neural structure learning

  • helps to train neural networks with structured signals.

  • And the new Fairness Indicators add-on

  • enables regular computation and visualization

  • of fairness metrics.

  • And these are just the types of things

  • that you can see kind of as part of the TensorFlow

  • ecosystem, these add-ons that, again, can help you make

  • sure you're able to do the things you

  • need to do not with your models, but kind of beyond just that.

  • Another valuable aspect of the TensorFlow ecosystem

  • is being able to analyze your ML experiments in detail.

  • So this is showing TensorBoard.

  • TensorBoard is TensorFlow's visualization toolkit,

  • which is what helps you accomplish this.

  • It's a popular tool among researchers

  • and ML practitioners for tracking metrics,

  • visualizing model graphs and parameters, and much more.

  • It's very interesting that we've seen users enjoy TensorBoard

  • so much, they'll even take screenshots

  • of their experiments and then use those screenshots

  • to be able to share with others what

  • they're doing with TensorFlow.

  • This type of sharing and collaboration in the ML

  • community is something we really want

  • to encourage with TensorFlow.

  • Again, there's so much that can happen

  • by enabling the community to do good things.

  • That's why I'm excited to share the preview of TensorBoard.dev,

  • a new, free, managed TensorBoard experience that lets you upload

  • and share your ML experiment results with anyone.

  • You'll now be able to host and track your ML experiments

  • and share them publicly.

  • No setup required.

  • Simply upload your logs, and then share the URL,

  • so that others can see the experiments

  • and see the things that you are doing with TensorFlow.

  • As a preview, we're starting off with the [INAUDIBLE] dashboard,

  • but over time, we'll be adding a lot more

  • functionality to make the sharing experience even better.

  • But if you're not looking to build models from scratch

  • and want to reduce some computational cost,

  • TensorFlow was always made pre-trained models

  • available through TensorFlow Hub.

  • And today, we're excited to share an improved experience

  • of TensorFlow Hub that's much more intuitive, where

  • you can find a comprehensive repository

  • of pre-trained models in the TensorFlow ecosystem.

  • This means you can find models like BERT

  • and others related to image, text, video,

  • and more that are ready to use with TensorFlow Lite

  • and TensorFlow.js.

  • Again, we wanted to make sure the experience here

  • was vastly improved to make it easier

  • for you to find what you need in order to more quickly get

  • to the task at hand.

  • And since TensorFlow is driven by all of you,

  • TensorFlow Hub is hosting more pre-trained models

  • from the community.

  • You'll be able to find curated models by DeepMind, Google,

  • Microsoft's AI for Earth, and NVIDIA ready to use today

  • with many more to come.

  • We want to make sure that TensorFlow Hub is a great place

  • to find some of these excellent pre-trained models.

  • And again, there's so much the community is doing.

  • We want to be able to showcase those models as well.

  • TensorFlow 2.0 also highlights TensorFlow's core strengths

  • and areas of focus, which is being

  • able to go from model building, experimentation,

  • through to production, no matter what platform you work on.

  • You can deploy end-to-end ML pipelines

  • with TensorFlow Extended or TFX.

  • You can use your models on mobile and embedded devices

  • with TensorFlow Lite for on device inference,

  • and you can train and run models in the browser or Node.js

  • with TensorFlow.js.

  • You'll learn more about what's new in TensorFlow

  • in production during the keynote sessions tomorrow.

  • You can learn more about these updates

  • by going to tensorflow.org where you'll also

  • find the latest documentation, examples,

  • and tutorials for 2.0.

  • Again, we want to make sure it's easy for the community

  • to see what's happening, what's new,

  • and enable you to just do what you need to do with TensorFlow.

  • We've been thrilled to see the positive response to 2.0,

  • and we hope you continue to share your feedback.

  • Thank you, and I hope you enjoy the rest of TF World.

  • [APPLAUSE]

  • FREDERICK REISS: Hello, everyone.

  • I'm Fred Reiss.

  • I work for IBM.

  • I've been working for IBM since 2006.

  • And I've been contributing to TensorFlow Core since 2017.

  • But my primary job at IBM is to serve as tech lead for CODAIT.

  • That's the Center for Open Source Data and AI

  • Technologies.

  • We are an open source lab located in downtown San

  • Francisco, and we work on open source technologies

  • that are foundational to AI.

  • And we have on staff 44 full-time developers

  • who work only on open source software.

  • And that's a lot of developers, a lot of open source

  • developers.

  • Or is it?

  • Well, if you look across IBM at all of the IBM-ers

  • who are active contributors to open source, in that they have

  • committed code to GitHub in the last 30 days,

  • you'll find that there are almost 1,200 IBM-ers

  • in that category.

  • So our 44 developers are actually a very small slice

  • of a very large pie.

  • Oh, and those numbers, they don't include Red Hat.

  • When we closed that acquisition earlier this year,

  • we more than doubled our number of active contributors

  • to open source.

  • So you can see that IBM is really big in open source.

  • And more and more, the bulk of our contributions in the open

  • are going towards the foundations of AI.

  • And when I say AI, I mean AI in production.

  • I mean AI at scale.

  • AI at scale is not an algorithm.

  • It's not a tool.

  • It's a process.

  • It's a process that starts with data,

  • and then that data turns into features.

  • And those features train models, and those models

  • get deployed in applications, and those applications

  • produce more data.

  • And the whole thing starts all over again.

  • And at the core of this process is an ecosystem

  • of open source software.

  • And at the core of this ecosystem

  • is TensorFlow, which is why I'm here,

  • on behalf of IBM open source, to welcome you

  • to TensorFlow World.

  • Now throughout this conference, you're

  • going to see talks that speak to all of the different stages

  • of this AI lifecycle.

  • But I think you're going to see a special emphasis

  • on this part--

  • moving models into production.

  • And one of the most important aspects of moving models

  • into production is that when your model gets deployed

  • in a real-world application, it's

  • going to start having effects on the real world.

  • And it becomes important to ensure

  • that those effects are positive and that they're

  • fair to your clients, to your users.

  • Now, at IBM, here's a hypothetical example

  • that our researchers put together about a little

  • over a year ago.

  • They took some real medical records data,

  • and they produced a model that predicts which patients

  • are more likely to get sick and therefore should

  • get additional screening.

  • And they showed that if you naively trained this model,

  • you end up with a model that has significant racial bias,

  • but that by deploying state-of-the-art techniques

  • to adjust the data set and the process of making the model,

  • they could substantially reduce this bias to produce a model

  • that is much more fair.

  • You can see a Jupyter Notebook with the entire scenario

  • from end to end, including code and equations and results,

  • at the URL down here.

  • Again, I need to emphasize this was a hypothetical example.

  • We built a flawed model deliberately,

  • so we could show how to make it better.

  • But no patients were harmed in this exercise.

  • However, last Friday, I sat down with my morning coffee,

  • and I opened up the "Wall Street Journal."

  • And I saw this article at the bottom of page three,

  • describing a scenario eerily similar to our hypothetical.

  • When your hypothetical starts showing up

  • as newspapers headlines, that's kind of scary.

  • And I think it is incumbent upon us as an industry to move

  • forward the process, the technology of trust in AI,

  • trust and transparency in AI, which is why IBM and IBM

  • Research have released our toolkits of state-of-the-art

  • algorithms in this space as open source under AI Fairness 360,

  • AI Explainability 360, and Adversarial Robustness 360.

  • It is also why IBM is working with other members of the Linux

  • Foundation AI, a trusted AI committee,

  • to move forward open standards in this area

  • so that we can all move more quickly to trusted AI.

  • Now if you'd like to hear more on this topic,

  • my colleague, Animesh Singh, will

  • be giving a talk this afternoon at 1:40

  • on trusted AI for the full 40 minute session.

  • Also I'd like to give a quick shout out

  • to my other co-workers from CODAIT

  • who have come down here to show you cool open source

  • demos at the IBM booth.

  • That's booth 201.

  • Also check out our websites, developer.ibm.com

  • and codait.org.

  • On behalf of IBM, I'd like to welcome you

  • all to TensorFlow World.

  • Enjoy the conference.

  • Thank you.

  • [APPLAUSE]

  • THEODORE SUMME: Hi, I'm Ted Summe from Twitter.

  • Before I get started with my conversation today,

  • I want to do a quick plug for Twitter.

  • What's great about events like this is you

  • get to hear people like Jeff Dean talk.

  • And you also get to hear from colleagues and people

  • in the industry that are facing similar challenges as you

  • and have conversations around developments in data science

  • and machine learning.

  • But what's great is that's actually available

  • every day on Twitter.

  • Twitter's phenomenal for conversation on data science

  • and machine learning.

  • People like Jeff Dean and other thought leaders

  • are constantly sharing their thoughts

  • and their developments.

  • And you can follow that conversation and engage in it.

  • And not only that, but you can bring that conversation back

  • to your workplace and come off looking like a hero--

  • just something to consider.

  • So without that shameless plug, my name's Ted Summe.

  • I lead product for Cortex.

  • Cortex is Twitter's central machine learning organization.

  • If you have any questions for me or the team,

  • feel free to connect with me on Twitter,

  • and we can follow up later.

  • So before we get into how we're accelerating ML at Twitter,

  • let's talk a little bit about how

  • we're even using ML at Twitter.

  • Twitter is largely organized against three customer needs,

  • the first of which is our health initiative.

  • That might be a little bit confusing to you.

  • You might think of it as user safety.

  • But we think about it as improving the health

  • of conversations on Twitter.

  • And machine learning is already at use here.

  • We use it to detect spam.

  • We can algorithmically and at scale

  • detect spam and protect our users from it.

  • Similarly, in the abuse space, we

  • can proactively flag content as potentially abuse,

  • toss it up for human reviews, and act on it

  • before our users even get impacted by it.

  • A third space where we're using machine learning here

  • is something called NSFW, Not Safe For Work.

  • I think you're all familiar with the acronym.

  • So how can we, at scale, identify this content

  • and handle it accordingly?

  • Another use of machine learning in this space.

  • There's more that we want to do here,

  • and there's more that we're already doing.

  • Similarly, the consumer organization-- this

  • is largely what you think of, the big blue app of Twitter.

  • And here, the customer job that we're serving

  • is helping connect our customers with the conversations

  • on Twitter that they're interested in.

  • And one of the primary veins in which we do this

  • is our timeline.

  • Our timeline today is ranked.

  • So if you're not familiar, users follow accounts.

  • Content and tweets associated with those accounts

  • get funneled into a central feed.

  • And we rank that based on your past engagement and interest

  • to make sure we bring forth the most relevant conversations

  • for you.

  • Now, there's lots of conversations on Twitter,

  • and you're not following everyone.

  • And so there's also a job that we

  • have to serve about bringing forth all the conversations

  • that you're not proactively following,

  • but are still relevant to you.

  • This has surfaced in our Recommendations product, which

  • uses machine learning to scan the corpus of content

  • on Twitter, and identify what conversations would

  • be most interesting to you, and push it

  • to you in a notification.

  • The inverse of that is when you know

  • what the topics you want to explore are,

  • but you're looking for the conversations around that.

  • That's where we use Twitter Search.

  • This is another surface area in the big blue app

  • that we're using machine learning.

  • The third job to be done for our customers

  • is helping connect brands with their customers.

  • You might think of this as the ads product.

  • And this is actually the OG of machine

  • learning at Twitter, the first team that implemented it.

  • And here, we use it for what you might expect, ads ranking.

  • That's kind of like the timeline ranking, but instead of tweets,

  • it's ads and identifying the most relevant ads

  • for our users.

  • And as signals to go into that, we also

  • do user targeting to understand your past engagement ads,

  • understand which ads are in your interest space.

  • And the third-- oh.

  • Yeah, we're still good.

  • And the third is brand safety.

  • You might not think about this when you think about machine

  • learning and advertising.

  • But if you're a company like United

  • and you want to advertise on Twitter,

  • you want to make sure that your ad never

  • shows up next to a tweet about a plane crash.

  • So how do we, at scale, protect our brands

  • from those off-brand conversations?

  • We use machine learning for this as well.

  • So as you can tell, machine learning

  • is a big part of all of these organizations today.

  • And where we have shared interests and shared

  • investment, we want to make sure we have a shared

  • organization that serves that.

  • And that's the need for Cortex.

  • Cortex is Twitter's central machine learning team,

  • and our purpose is really quite simple--

  • to enable Twitter with ethical and advanced AI.

  • And to serve that purpose, we've organized in three ways.

  • The first is our applied research group.

  • This group applies the most advanced ML techniques

  • from industry and research to our most important surface

  • areas, whether they be new initiatives or existing places.

  • This team you can kind of think of as like an internal task

  • force or consultancy that we can redeploy against the company's

  • top initiatives.

  • The second is signals.

  • When using machine learning, having

  • shared data assets that are broadly useful

  • can provide us more leverage.

  • Examples of this would be our language understanding team

  • that looks at tweets and identifies

  • named entities inside them.

  • Those can then be offered up as features for other teams

  • to consume in their own applications of machine

  • learning.

  • Similarly, our media understanding team

  • looks at images and can create a fingerprint of any image.

  • And therefore, we can identify every use of that image

  • across the platform.

  • These are examples of shared signals that we're

  • producing that can be used for machine learning

  • at scale inside the company.

  • And the third organization is our platform team.

  • And this is really the origins of Cortex.

  • Here, we provide tools and infrastructure

  • to accelerate ML development at Twitter,

  • increase the velocity of our ML practitioners.

  • And this is really the focus of the conversation today.

  • When we set out to build this ML platform,

  • we decided we wanted a shared ML platform across all of Twitter.

  • And why is that important that it be

  • shared across all of Twitter?

  • Well, we want transferability.

  • We want the great work being done in the ads team

  • to be, where possible, transferable

  • to benefit the health initiative where that's relevant.

  • And similarly, if we have great talent in the consumer team

  • that's interested in moving to the ads team,

  • if they're on the same platform, they

  • can transfer without friction and be able to ramp up quickly.

  • So we set out with this goal of having a shared ML

  • platform across all of Twitter.

  • And when we did that, we looked at a couple product

  • requirements.

  • First, it needs to be scalable.

  • It needs to be able to operate at Twitter scale.

  • The second, it needs to be adaptable.

  • This space is developing quickly so we

  • need a platform that can evolve with data science and machine

  • learning developments.

  • Third is the talent pool.

  • We want to make sure that we have a development environment

  • at Twitter that appeals to the ML researchers and engineers

  • that we're hiring and developing.

  • Fourth is the ecosystem.

  • We want to be able to lean on the partners that

  • are developing industry leading tools

  • so that we can focus on technologies

  • that are Twitter specific.

  • Fourth is documentation.

  • You ought to understand that.

  • We want to be able to quickly unblock

  • our practitioners as they hit issues, which

  • is inevitable in any platform.

  • And finally, usability.

  • We want to remove friction and frustration

  • from the lives of our team, so that they

  • can focus on delivering value for our end customers.

  • So considering these product requirements,

  • let's see how TensorFlow is done against them.

  • First is scalability.

  • We validated this by putting TensorFlow

  • by way of our implementation we called Deep Bird

  • against timeline ranking.

  • So every tweet that's ranked in the timeline today

  • runs through TensorFlow.

  • So we can consider that test validated.

  • Second is adaptability.

  • The novel architectures that TensorFlow can support,

  • as well as the custom lost functions, allows

  • us to react to the latest research

  • and employ that inside the company.

  • An example that we published on this

  • publicly is our use of a SplitNet architecture and ads

  • ranking.

  • So TensorFlow has been very adaptable for us.

  • Third is the talent pool, and we think about the talent

  • pool in kind of two types.

  • There's the ML engineer and the ML researcher.

  • And as a proxy of these audiences,

  • we looked at the GitHub data on these.

  • And clearly, TensorFlow is widely adopted

  • amongst ML engineers.

  • And similarly, the archive community

  • shows strong evidence of wide adoption

  • in the academic community.

  • On top of this proxy data, we also

  • have anecdotal evidence of the speed

  • of ramp-up for ML researchers and ML

  • engineers inside the company.

  • The fourth is the ecosystem.

  • Whether it's TensorBoard, TF Data Validation, TF Model

  • Analysis, TF Metastore, TF Hub, TFX Pipelines,

  • there's a slew of these products out there,

  • and they're phenomenal.

  • They allow us to focus on developing tools

  • and infrastructure that is specific to Twitter's needs

  • and lead on the great work of others.

  • So we're really grateful for this,

  • and TensorFlow does great here.

  • Fifth being documentation.

  • Now, this is what you would go to when you go to TensorFlow,

  • and you see that phenomenal documentation, as well as

  • great education resources.

  • But what you might not appreciate

  • and what we've come to really appreciate

  • is the value of the user generated content.

  • What Stack Overflow and other platforms

  • can provide in terms of user generated

  • content is almost as valuable as anything

  • TensorFlow itself can create.

  • And so TensorFlow, given its widespread adoption,

  • its great TensorFlow website, has

  • provided phenomenal documentation

  • for ML practitioners.

  • Finally, usability.

  • And this is why we're really excited about TensorFlow 2.0.

  • The orientation around the carrier's API

  • makes it more user friendly.

  • It also still continues to allow for flexibility

  • for more advanced users.

  • The eager execution enables more rapid and intuitive debugging,

  • and it closes the gap between ML engineers and modelers.

  • So clearly from this checklist, we're

  • pretty happy with our engagement with TensorFlow.

  • And we're excited about continuing

  • to develop the platform with them

  • and push the limits on what it can

  • do, with gratitude to the community

  • for their participation and involvement in the product

  • and appreciate their conversation on Twitter, as we

  • advance it.

  • So if you have any questions for me, as I said before,

  • you can connect with me, but I'm not alone here today.

  • A bunch of my colleagues are here as well.

  • So if you see them roaming the halls,

  • feel free to engage with them.

  • Or as I shared before, you can continue the conversation

  • on Twitter.

  • Here are their handles.

  • Thank you for your time.

  • Cheers.

  • [APPLAUSE]

  • CRAIG WILEY: I just want to begin

  • by saying I've been dabbling in Cloud AI and Cloud Machine

  • Learning for a while.

  • And during that time, it never occurred to me

  • that we'd be able to come out with something like we

  • did today because this is only possible because Google Cloud

  • and TensorFlow can collaborate unbelievably

  • closely together within Google.

  • So to begin, let's talk a little bit about TensorFlow--

  • 46 million downloads.

  • TensorFlow has been massive growth the last few years.

  • It's expanded from the forefront of research, which we've

  • seen earlier this morning, to businesses taking it

  • on as a dependency for their business to operate on a day

  • in, day out basis.

  • It's a super exciting piece.

  • As someone who spends most all of their time thinking

  • about how we can bring AI and machine learning

  • into businesses, seeing TensorFlow's commitment

  • and focus on deploying actual ML in production

  • is super exciting to me.

  • With this growth, though, comes growing pains.

  • And part of that is things like support, right?

  • When my model doesn't do what I expected it to

  • or my training job fails, what options do I have?

  • And how well does your boss respond when you say,

  • hey, yes, I don't know why my model's not training,

  • but not to worry, I've put a question on Slack.

  • And hopefully, someone will get back to me.

  • We understand that businesses who

  • are taking a bet on TensorFlow as a critical piece

  • of their hardware architecture or their stack

  • need more than this.

  • Second, it can be a challenge to unlock the scale

  • and performance of cloud.

  • For those of you who, like me, have gone through this journey

  • over the last couple of years, for me,

  • it started on my laptop.

  • Right?

  • And then eventually, I outgrew my laptop,

  • and so I had a gaming rig under my desk, right?

  • With the GPU and eventually, there

  • were eight gaming rigs under my desk.

  • And when you opened the door to my office,

  • the whole floor knew because it sounded like [INAUDIBLE]..

  • Right?

  • And but now with today's cloud, that

  • doesn't have to be the case.

  • You can go from that single instance

  • all the way up to a massive scale seamlessly.

  • So with that, today, we bring you TensorFlow Enterprise.

  • TensorFlow Enterprise is designed to do three things--

  • one, give you Enterprise grade support; two, cloud scale

  • performance; and three, managed services

  • when and where you want them, at the abstraction level

  • you want them.

  • Enterprise grade support, what does that mean?

  • Fundamentally what that means is that as these businesses take

  • a bet on TensorFlow, many of these businesses

  • have IT policies or requirements that the software have

  • a certain longevity before they're willing to commit

  • to it in production.

  • And so today, for certain versions of TensorFlow,

  • when used on Google Cloud, we will extend that one year

  • of support a full three years.

  • That means that if you're building models on 1.15 today,

  • you can know that for the next three years,

  • you'll get bug fixes and security patches when

  • and where you need them.

  • Simple and scalable.

  • Scaling from an idea on a single node

  • to production at massive scale can be daunting, right?

  • Saying to my boss, hey, I took a sample of the data

  • was something that previously seemed totally reasonable,

  • but now we're asked to train on the entire corpus of data.

  • And that can take days, weeks.

  • We can help with all of that by deploying TensorFlow

  • on Google Cloud, a network that's been running TensorFlow

  • successfully for years and has been highly

  • optimized for this purpose.

  • So scalable across our world class architecture,

  • the products are compatibility tested

  • with the cloud, their performance optimized

  • for the cloud and for Google's world class infrastructure.

  • What does this mean?

  • So if any of you have ever had the opportunity

  • to use BigQuery, BigQuery is Google Cloud's kind

  • of massively parallel cloud hosted data warehouse.

  • And by the way, if you haven't tried using BigQuery,

  • I highly recommend going out and trying it.

  • It returns results faster than can be imagined.

  • That speed in BigQuery, we wanted

  • to make sure we were taking full advantage of that.

  • And so recent changes and recent pieces

  • included in TensorFlow Enterprise

  • have increased the speed of the connection between the data

  • warehouse and TensorFlow by three times.

  • Right?

  • Now, all of sudden, those jobs that were taking days

  • take hours.

  • Unity gaming, wonderful customer and partner with us.

  • You can see the quote here.

  • Unity leverages these aspects of TensorFlow Enterprise

  • in their business.

  • Their monetization products reach more than three billion

  • devices--

  • three billion devices worldwide.

  • Game developers rely on a mix of scale and products

  • to drive installs and revenue and player engagement.

  • And Unity needs to be able to quickly test, build, scale,

  • deploy models all at massive scale.

  • This allows them to serve up the best

  • results for their developers and their advertisers.

  • Managed services.

  • As I said, TensorFlow Enterprise will

  • be available on Google Cloud and will

  • be available as part of Google Cloud's AI platform.

  • It will also be available in VMs if you'd prefer that,

  • or in containers if you want to run them on Google Cloud

  • Kubernetes Engine, or using Kubeflow on Kubernetes Engine.

  • In summary, TensorFlow Enterprise

  • offers Enterprise grade support--

  • that continuation, that full three years of support

  • that IT departments are accustomed to--

  • cloud scale performance so that you can run at massive scale,

  • and works seamlessly with our managed services.

  • And all of this is free and fully included

  • for all Google Cloud users.

  • Google Cloud becomes the best place to run TensorFlow.

  • But there's one last piece, which

  • is for companies for whom AI is their business--

  • not companies for whom AI might help

  • with this part of their business or that or might help

  • optimize this campaign or this backend system,

  • but for companies where AI is their business, right?

  • Where they're training hundreds of thousands

  • of hours of training a year, petabytes of data, right?

  • Using cutting edge models to meet their unique requirements,

  • we are introducing TensorFlow Enterprise

  • with white-glove support.

  • This is really for cutting edge AI, right?

  • Engineering to engineering assistance when needed.

  • Close collaboration across Google

  • allows us to fix bugs faster if needed.

  • One of the great opportunities of working in cloud,

  • if you ask my kids, they'll tell you

  • that the reason I work in cloud AI

  • and in kind of machine learning is in an effort to keep them

  • ever from learning to drive.

  • They're eight and 10 years old, so I need people

  • to kind of hurry along this route, if you will.

  • But one of the customers and partners we have

  • is Cruise Automotive.

  • And you can see here, they're a shining example

  • of the work we're doing.

  • On their quest towards self-driving cars,

  • they've also experienced hiccups and challenges

  • and scaling problems.

  • And we've been a critical partner for them

  • in helping ensure that they can achieve the results they need

  • to, to solve this kind of generational defining

  • problem of autonomous vehicles.

  • You can see not only did we improve

  • the accuracy of their models, but also reduce training times

  • from four days down to one day.

  • This allows them to iterate at speeds previously unthinkable.

  • So none of this, as I said, would have been possible

  • without the close collaboration between Google Cloud

  • and TensorFlow.

  • I look back on Megan's recent announcement

  • of TensorBoard.dev.

  • We will be looking at bringing that type of functionality

  • into an enterprise environment as well in the coming months.

  • But we're really, really excited to get TensorFlow Enterprise

  • into your hands today.

  • To learn more and get started, you

  • can go to the link, as well as sessions later today.

  • And if you are on the cutting edge of AI,

  • we are accepting applications for the white-glove service

  • as well.

  • We're excited to bring this offering to teams.

  • We're excited to bring this offering to businesses

  • that want to move into a place where machine learning is

  • increasingly a part of how they create value.

  • Thank you very much for your time today

  • KEMAL EL MOUJAHID: Hi, my name is Kemal.

  • I'm the product director for TensorFlow.

  • So earlier, you heard from Jeff and Megan

  • about the prod direction.

  • Now, what I'd like to talk about is the most important part

  • of what we're building, and that's the community.

  • That's you.

  • Sorry.

  • Where's the slide?

  • Thank you.

  • So as you've seen in the video, we've

  • got a great roadshow, 11 events spanning five continents

  • to connect the community with the TensorFlow team.

  • I, personally, was very lucky this summer,

  • because I got to travel to Morocco and Ghana and Shanghai,

  • amongst other places, just to meet the community,

  • and to listen to your feedback.

  • And we heard a lot of great things.

  • So as we're thinking about, how can we best help the community?

  • It really came down to three things.

  • First, we would like to help you to connect with the larger

  • community, and to share the latest and greatest of what

  • you've been building.

  • Then, we also would like you--

  • we want to help you learn, learn about ML,

  • learn about TensorFlow.

  • And then, we want to help you contribute and give back

  • to the community.

  • So let's start with Connect.

  • So why connect?

  • Well, first the community-- the TensorFlow community

  • has really grown a lot.

  • It's huge-- 46 million downloads, 2,100 committers,

  • and--

  • again, I know that we've been saying that all along,

  • but I really want to say a huge thank you

  • on behalf of the TensorFlow team for making the community what

  • it is today.

  • Another aspect of the community that we're very proud of

  • is that it's truly global.

  • This is a revised map of our GitHub stars.

  • And, as you can see, we're covering all time zones

  • and we keep growing.

  • So the community is huge.

  • It's truly global.

  • And we really want to think about,

  • how can we bring the community closer together?

  • And this is really what initiated

  • the idea of TensorFlow World.

  • We wanted to create an event for you.

  • We wanted an event where you could come up and connect

  • with the rest of the community, and share

  • what you've been working on.

  • And this has actually started organically.

  • Seven months ago, the TensorFlow User Groups started,

  • and I think now we have close to 50.

  • The largest one is in Korea.

  • It has 46,000 members.

  • We have 50 in China.

  • So if you're in the audience or in the livestream,

  • and you're looking to this map, and you're thinking, wait,

  • I don't see a dot where I live--

  • and you have a TensorFlow member that you're connecting with,

  • and you want to start a TensorFlow User Group-- well,

  • we'd like to help you.

  • So please go to tensorflow.org/community,

  • and we'll help you get it started.

  • So that next year, when we look at this map,

  • we have dots all over the place.

  • So what about businesses?

  • We've talked about developers.

  • What about businesses?

  • One thing we heard from businesses

  • is they have this business problem.

  • They think ML can help them, but they're not sure how.

  • And that's a huge missed opportunity

  • when we look at the staggering $13 trillion

  • that AI will bring to the global economy over the next decade.

  • So you have those businesses on one side,

  • and then you have partners on the other side, who

  • know about ML, they know how to use TensorFlow,

  • so how do we connect those two?

  • Well, this was the inspiration for launching our Trusted

  • Partner Pilot Program, which helps you, as a business,

  • connect to a partner who will help you solve your ML problem.

  • So if you go on tensorflow.org, you'll

  • find more about our Trusted Partner program.

  • Just a couple of examples of cool things

  • that they've been working on.

  • One partner helped a car insurance company

  • shorten the insurance claim processing time using

  • image processing techniques.

  • Another partner helped the global med tech company

  • by automating the shipping labeling process using

  • object recognition techniques.

  • And you'll hear more from these partners later today.

  • I encourage you to go check out their talks.

  • Another aspect is that if you're a partner,

  • and you're interested in getting into this program,

  • we also would like to hear from you.

  • So let's talk about Learn.

  • We've invested a lot in producing quality material

  • to help you learn about ML and about TensorFlow.

  • One thing that we did over the summer, which was very exciting

  • is for the first time, we're a part of the Google Summer

  • of Code.

  • We had a lot of interest.

  • We were able to select 20 very talented students,

  • and they got to work the whole summer

  • with amazing mentors on the TensorFlow engineering team.

  • And they worked on very inspiring projects

  • going from 2.0 to Swift to JS to TF-Agents.

  • So we were so excited with the success of this program

  • that we decided to participate, for the first time,

  • in a Google Code-in program.

  • So this is the same program, but for pre-university students

  • from 13 to 17.

  • It's a global online contest.

  • And it introduces teenagers to the world of contributing

  • to open source development.

  • So as I mentioned, we've invested a lot this year

  • on ML education material, but one thing we heard

  • is that there's a lot of different things.

  • And what you want is to be guided

  • through pathways of learning.

  • So we've worked hard on that, and we've

  • decided to announce the new Learn ML page tensorflow.org.

  • And what this is a learning path curated

  • for you by the TensorFlow team, and organized by level.

  • So you have from beginners to advanced.

  • You can explore books, courses, and videos

  • to help you improve your knowledge of machine learning,

  • and use that knowledge and use TensorFlow, to solve

  • your real-world problem.

  • And for more exciting news that will

  • be available on the website, I'd like

  • to play a brief video by a friend Andrew Ng.

  • [VIDEO PLAYBACK]

  • - Hi, everyone.

  • I'm in New York right now, and wish

  • I could be there to enjoy the conference.

  • But I want to share with you some exciting updates.

  • Deeplearning.ai started a partnership

  • with the TensorFlow team with a goal

  • of making world-class education available for developers

  • on the Coursera platform.

  • Since releasing the Deep Learning Specialization,

  • I've seen so many of you, hundreds of thousands,

  • learn the fundamental skills of deep learning.

  • I'm delighted we've been able to complement that

  • with the TensorFlow in Practice Specialization

  • to help developers learn how to build

  • ML applications for computer vision, NLP,

  • sequence models, and more.

  • Today, I want to share with you an exciting new project

  • that the deeplearning.ai and TensorFlow teams have

  • been working on together.

  • Being able to use your models in a real-world scenario

  • is when machine learning gets particularly exciting.

  • So we're producing a new four-course specialization

  • called TensorFlow Data and Deployment that

  • will let you take your ML skills to the real world,

  • deploying models to the web, mobile devices, and more.

  • It will be available on Coursera in early December.

  • I'm excited to see what you do with these new resources.

  • Keep learning.

  • [END PLAYBACK]

  • KEMAL EL MOUJAHID: All right.

  • This is really cool.

  • Since we started working on these programs,

  • it's been pretty amazing to see hundreds of thousands of people

  • take those courses.

  • And the goal of these educational resources

  • is to let everyone participate in the ML revolution,

  • regardless of what your experience with machine

  • learning is.

  • And now, Contribute.

  • So a great way to get involved is to connect with your GDE.

  • We now have 126 machine learning GDEs globally.

  • We love our GDEs.

  • They're amazing.

  • They do amazing things for the community.

  • This year alone, they gave over 400 tech talks, 250 workshops.

  • They wrote 221 articles reaching tens of thousands

  • of developers.

  • And one thing that was new this year

  • is that they helped with doc sprints.

  • So docs are really important.

  • They're critical, right?

  • You really need good quality docs

  • to work on machine learning, and often the documentation

  • is not available in people's native languages.

  • And so this is why when we partnered with our GDEs,

  • we launched the doc sprints.

  • Over 9,000 API docs were updated by members of the TensorFlow

  • community in over 15 countries.

  • We heard amazing stories of power outage,

  • and power running out, and people coming back later

  • to finish a doc sprint, and actually writing

  • docs on their phones.

  • So if you've been helping with docs,

  • thank you, if you're in the room,

  • if you're over livestream, thank you so much.

  • If you're interested in helping translate documentation

  • in your native language, please reach out,

  • and we'll help you organize a doc sprint.

  • Another thing that the GDEs help with

  • is experimenting with the latest features.

  • So I want to call out Sam Witteveen, an ML

  • GDE from Singapore, who's already experimenting

  • with 2.x TPUs, and you can hear him talk later today to hear

  • about his experience.

  • So if you want to get involved, please reach out to your GDE

  • and start working on TensorFlow.

  • Another really great way to help is to join a SIG.

  • A SIG is a Special Interest Group,

  • and it helps you work on the things

  • that you're the most excited about on TensorFlow.

  • We have, now, 11 SIGs available.

  • Addons, IO, and Networking, in particular,

  • really supported the transition to 2.0

  • by embracing the parts of contrib

  • and putting them into 2.0.

  • And SIG Build ensures that TF runs well everywhere

  • on any OS, any architecture, and plays well

  • with the Python library.

  • And we have many other really exciting SIGs, so I really

  • encourage you to join one.

  • Another really great way to contribute

  • is through competition.

  • And for those of you who were there

  • at the Dev Summit back in March, we launched our 2.0 challenge

  • on DevPost.

  • And the grand prize was an invitation

  • to this event, TensorFlow World.

  • And so we would like to honor our 2.0 Challenge winners,

  • and I think we are lucky to have two of them in the room--

  • Victor and Kyle, if you're here.

  • [APPLAUSE]

  • So Victor worked on Handtrack.js, a library

  • for prototyping hand gesture in the browser.

  • And then Kyle worked on a Python 3 package to simulate N-body,

  • to generate N-body simulations.

  • So one thing we heard, too, during our travels

  • is, oh, that hackathon was great, but I totally missed it.

  • Can we have another one?

  • Well, yes.

  • Let's do another one.

  • So if you go on tfworld.devpost.com,

  • we're launching a new challenge.

  • You can apply your 2.0 skills and share

  • the latest and greatest, and win cool prizes.

  • So we're really excited to see what you're going to build.

  • Another great community that we're very excited to partner

  • with is Kaggle.

  • So we've launched a contest on Kaggle

  • to challenge you with question answering model based

  • on Wikipedia articles.

  • You can put your natural language processing skills

  • to the test and earn $50,000 in prizes.

  • It's open for entry until January 22, so best of luck.

  • So we have a few action items for you,

  • and they're listed on this slide.

  • But remember, we created TensorFlow World for you,

  • to help you connect and share what you've been working on.

  • So our main action item for you in the next two days

  • is really to get to know the community better.

  • And with that, I'd like to thank you,

  • and I hope you enjoy the rest of TF World.

  • Thank you.

  • [APPLAUSE]

JEFF DEAN: I'm really excited to be here.

字幕與單字

單字即點即查 點擊單字可以查詢單字解釋

B1 中級

TensorFlow World 2019主題演講 (TensorFlow World 2019 Keynote)

  • 3 0
    林宜悉 發佈於 2021 年 01 月 14 日
影片單字