字幕列表 影片播放 列印英文字幕 JEFF DEAN: I'm really excited to be here. I think it was almost four years ago to the day that we were about 20 people sitting in a small conference room in one of the Google buildings. We've woken up early because we wanted to kind of time this for an early East Coast launch where we were turning on the TensorFlow.org website and releasing the first version of TensorFlow as an open source project. And I'm really, really excited to see what it's become. It's just remarkable to see the growth and all the different kinds of ways in which people have used this system for all kinds of interesting things around the world. So one thing that's interesting is the growth in the use of TensorFlow also kind of mirrors the growth in interest in machine learning and machine learning research generally around the world. So this is a graph showing the number of machine learning archive papers that have been posted over the last 10 years or so. And you can see it's growing quite, quite rapidly, much more quickly than you might expect. And that lower red line is kind of the nice doubling every couple of years growth rate, exponential growth rate we got used to in computing power, due to Moore's law for so many years. That's now kind of slowed down. But you can see that the machine learning research community is generating research ideas at faster than that rate, which is pretty remarkable. We've replaced computational growth with growth of ideas, and we'll see those both together will be important. And really, the excitement about machine learning is because we can now do things we couldn't do before, right? As little as five or six years ago, computers really couldn't see that well. And starting in about 2012, 2013, we started to have people use deep neural networks to try to tackle computer vision problems, image classification, object detection, things like that. And so now, using deep learning and deep neural networks, you can feed in the raw pixels of an image and fairly reliably get a prediction of what kind of object is in that image. Feed in the pixels there. Red, green, and blue values in a bunch of different coordinates, and you get out the prediction leopard. This works for speech as well. You can feed an audio wave forms, and by training on lots of audio wave forms and transcripts of what's being said in those wave forms, we can actually take a completely new recording and tell you what is being said amid a transcript. Bonjour, comment allez-vous? You can even combine these ideas and have models that take in pixels, and instead of just predicting classifications of what are in the object, it can actually write a short sentence, a short caption, that a human might write about the image-- a cheetah lying on top of a car. That's one of my vacation photos, which was kind of cool. And so just to show the progress in computer vision, in 2011, Stanford hosts an ImageNet contest every year to see how well computer vision systems can predict one of 1,000 categories in a full color image. And you get about a million images to train on, and then you get a bunch of test images your model has never seen before. And you need to make a prediction. In 2011, the winning entrant got 26% error, right? So you can kind of make out what that is. But it's pretty hard to tell. We know from human experiment that human error of a well-trained human, someone who's practiced at this particular task and really understands 1,000 categories, gets about 5% error. So this is not a trivial task. And in 2016, the winning entrant got 3% error. So just look at that tremendous progress in the ability of computers to resolve and understand computer imagery and have computer vision that actually works. This is remarkably important in the world, because now we have systems that can perceive the world around us and we can do all kinds of really interesting things about. We've seen similar progress in speech recognition and language translation and things like that. So for the rest of the talk, I'd like to kind of structure it around this nice list of 14 challenges that the US National Academy of Engineering put out and felt like these were important things for the science and engineering communities to work on for the next 100 years. They put this out in 2008 and came up with this list of 14 things after some deliberation. And I think you'll agree that these are sort of pretty good large challenging problems, that if we actually make progress on them, that we'll actually have a lot of progress in the world. We'll be healthier. We'll be able to learn things better. We'll be able to develop better medicines. We'll have all kinds of interesting energy solutions. So I'm going to talk about a few of these. And the first one I'll talk about is restoring and improving urban infrastructure. So we're on the cusp of the sort of widespread commercialization of a really interesting new technology that's going to really change how we think about transportation. And that is autonomous vehicles. And this is a problem that has been worked on for quite a while, but it's now starting to look like it's actually completely possible and commercially viable to produce these things. And a lot of the reason is that we now have computer vision and machine learning techniques that can take in sort of raw forms of data that the sensors on these cars collect. So they have the spinning LIDARs on the top that give them 3D point cloud data. They have cameras in lots of different directions. They have radar in the front bumper and the rear bumper. And they can really take all this raw information in, and with a deep neural network, fuse it all together to build a high level understanding of what is going on around the car. Or is it another car to my side, there's a pedestrian up here to the left, there's a light post over there. I don't really need to worry about that moving. And really help to understand the environment in which they're operating and then what actions can they take in the world that are both legal, safe, obey all the traffic laws, and get them from A to B. And this is not some distant far-off dream. Alphabet's Waymo subsidiary has actually been running tests in Phoenix, Arizona. Normally when they run tests, they have a safety driver in the front seat, ready to take over if the car does something kind of unexpected. But for the last year or so, they've been running tests in Phoenix with real passengers in the backseat and no safety drivers in the front seat, running around suburban Phoenix. So suburban Phoenix is a slightly easier training ground than, say, downtown Manhattan or San Francisco. But it's still something that is like not really far off. It's something that's actually happening. And this is really possible because of things like machine learning and the use of TensorFlow in these systems. Another one that I'm really, really excited about is advance health informatics. This is a really broad area, and I think there's lots and lots of ways that machine learning and the use of health data can be used to make better health care decisions for people. So I'll talk about one of them. And really, I think the potential here is that we can use machine learning to bring the wisdom of experts through a machine learning model anywhere in the world. And that's really a huge, huge opportunity. So let's look at this through one problem we've been working on for a while, which is diabetic retinopathy. So diabetic retinopathy is the fastest growing cause of preventable blindness in the world. And screening every year, if you're at risk for this, and if you have diabetes or early sort of symptoms that make it likely you might develop diabetes, you should really get screened every year. So there's 400 million people around the world that should be screened every year. But the screening is really specialized. Doctors can't do it. You really need ophthalmologist level of training in order to do this effectively. And the impact of the shortage is significant. So in India, for example, there's a shortage of 127,000 eye doctors to do this sort of screening. And as a result, 45% of patients who are diagnosed to this disease actually have suffered either full or partial vision loss before they're actually diagnosed and then treated. And this is completely tragic because this disease, if you catch it in time, is completely treatable. There's a very simple 99% effective treatment that we just need to make sure that the right people get treated at the right time. So what can you do? So, it turns out diabetic retinopathy screening is also a computer vision problem, and the progress we've made on general computer vision problems where you want to take a picture and tell if that's a leopard or an aircraft carrier or a car actually also works for diabetic retinopathy. So you can take a retinal image, which is what the screening camera, sort of the raw data that comes off the screening camera, and try to feed that into a model that predicts 1, 2, 3, 4, or 5. That's how these things are graded, 1 being no diabetic retinopathy, 5 being proliferative, and the other numbers being in between. So it turns out you can get a collection of data of retinal images and have ophthalmologists label them. Turns out if you ask two ophthalmologists to label the same image, they agree with each other 60% of the time on the number 1, 2, 3, 4, or 5. But perhaps slightly scarier if you ask the same ophthalmologist to grade the same image a few hours apart, they agree with themselves 65% of the time. But you can fix this by actually getting each image labeled by a lot of ophthalmologists, so you'll get it labeled by seven ophthalmologists. If five of them say it's a 2, and two of them say it's a 3, it's probably more like a 2 than a 3. Eventually, you have a nice, high quality data set you can train on. Like many machine learning problems, high quality data is the right raw ingredient. But then you can apply, basically, an off-the-shelf computer vision model trained on this data set. And now you can get a model that is on par or perhaps slightly better than the average board certified ophthalmologist in the US, which is pretty amazing. It turns out you can actually do better than that. And if you get the data labeled by retinal specialists, people who have more training in retinal disease and change the protocol by which you label things, you get three retinal specialists to look at an image, discuss it amongst themselves, and come up with what's called a sort of coordinated assessment and one number. Then you can train a model and now be on par with retinal specialists, which is kind of the gold standard of care in this area. And that's something you can now take and distribute widely around the world. So one issue particularly with health care kinds of problems is you want explainable models. You want to be able to explain to a clinician why is this person, why do we think this person has moderate diabetic retinopathy. So you can take a retinal image like this, and one of the things that really helps is if you can show in the model's assessment why this is a 2 and not a 3. And by highlighting parts of the input data, you can actually make this more understandable for clinicians and enable them to really sort of get behind the assessment that the model is making. And we've seen this in other areas as well. There's been a lot of work on explainability, so I think the notion that deep neural networks are sort of complete black boxes is a bit overdone. There's actually a bunch of good techniques that are being developed and more all the time that will improve this. So a bunch of advances depend on being able to understand text. And we've had a lot of really good improvements in the last few years on language understanding. So this is a bit of a story of research and how research builds on other research. So in 2017, a collection of Google researchers and interns came up with a new kind of model for text called the Transformer model. So unlike recurrent models where you have kind of a sequential process where you absorb one word or one token at a time and update some internal state and then go on to the next token, the Transformer model enables you to process a whole bunch of text, all at once in parallel, making it much more computationally efficient, and then to use attention on previous texts to really focus on if I'm trying to predict what the next word is, what are other parts of the context to the left that are relevant to predicting that? So that paper was quite successful and showed really good results on language translation tasks with a lot less compute. So the blue score there and the first two columns for English to German and English to French, higher is better. And then the compute cost of these models shows that this is getting sort of state of the art results at that time, with 10 to 100x less compute than other approaches. Then in 2018, another team of Google researchers built on the idea of Transformers. So everything you see there in a blue oval is a Transformer module, and they came up with this approach called Bidirectional Encoding Representations from Transformers, or BERT. It's a little bit shorter and more catchy. So BERT has this really nice property that, in addition to using context to the left, it uses context all around the language, sort of the surrounding text, in order to make predictions about text. And the way it works is you start with a self-supervised objective. So the one really nice thing about this is there's lots and lots of text in the world. So if you can figure out a way to use that text to train a model to be able to understand text better, that would be great. So we're going to take this text, and in the BERT training objective, to make it self-supervised, we're going to drop about 15% of the words. And this is actually pretty hard, but the model is then going to try to fill in the blanks, essentially. Try to predict what are the missing words that were dropped. And because we actually have the original words, we now know if the model is correct in its guesses about what goes in the box. And by processing trillions of words of text like this, you actually get a very good understanding of contextual cues in language and how to actually fill in the blanks in a really intelligent way. And so that's essentially the training objective for BERT. You take text, you drop 15% of it, and then you try to predict those missing words. And one key thing that works really well is that step one. You can pre-train a model on lots and lots of text, using this fill-in-the-blank self-supervised objective function. And then step two, you can then take a language task you really care about. Like maybe you want to predict, is this a five-star review or a one-star review for some hotel, but you don't have very much labeled text for that actual task. You might have 10,000 reviews and know the star count of each review. But you can then finetune the model, starting with the model trained in step one on trillions of words of text and now use your paltry 10,000 examples for the text task you really care about. And that works extremely well. So in particular, BERT gave state-of-the-art results across a broad range of different text understanding benchmarks in this GLUE benchmark suite, which was pretty cool. And people have been using BERT now in this way to improve all kinds of different things all across the language understanding and NLP space. So one of the grand challenges was engineer the tools of scientific discovery. And I think it's pretty clear machine learning is actually going to be an important component of making advances in a lot of these other grand challenge areas, things like autonomous vehicles or other kinds of things. And it's been really satisfying to see what we'd hoped would happen when we released TensorFlow as an open source project has actually kind of come to pass, as we were hoping, in that lots of people would sort of pick up TensorFlow, use it for all kinds of things. People would improve the core system. They would use it for tasks we would never imagine. And that's been quite satisfying. So people have done all kinds of things. Some of these are uses inside of Google. Some are outside in academic institutions. Some are scientists working on conserving whales or understanding ancient scripts, many kinds of things, which is pretty neat. The breadth of uses is really amazing. These are the 20 winners of the Google.org AI Impact Challenge, where people could submit proposals for how they might use machine learning and AI to really tackle a local challenge they saw in their communities. And they have all kinds of things, ranging from trying to predict better ambulance dispatching to identifying sort of illegal logging using speech recognition or audio processing. Pretty neat. And many of them are using TensorFlow. So one of the things we're pretty excited about is AutoML, which is this idea of automating some of the process by which machine learning experts sit down and sort of make decisions to solve machine learning problems. So currently, you have a machine learning expert sit down, they take data, they have computation. They run a bunch of experiments. They kind of stir it all together. And eventually, you get a solution to a problem you actually care about. One of the things we'd like to be able to do, though, is see if we could eliminate a lot of the need for the human machine learning expert to run these experiments and instead, automate the experimental process by which a machine learning expert comes by a high quality solution for a problem you care about. So lots and lots of organizations around the world have machine learning problems, but many, many of them don't even realize they have a machine learning problem, let alone have people in their organization that can tackle the problem. So one of the earliest pieces of work our researchers did in the space was something called neural architecture search. So when you sit down and design a neural network to tackle a particular task, you make a lot of decisions about shapes of this or that, and should it be used 3 by 3 filters at layer 17 or 5 by 5, all kinds of things like this. It turns out you can automate this process by having a model generating model and train the model generating model based on feedback about how well the models that it generates work on the problem you care about. So the way this will work, we're going to generate a bunch of models. Those are just descriptions of different neural network architectures. We're going to train each of those for a few hours, and then we're going to see how well they work. And then use the accuracy of those models as a reinforcement learning signal for the model generating model, to steer it away from models that didn't work very well and towards models that worked better. And we're going to repeat many, many times. And over time, we're going to get better and better by steering the search to the parts of the space of models that worked well. And so it comes up with models that look a little strange, admittedly. A human probably would not sit down and wire up a sort of machine learning, computer vision model exactly that way. But they're pretty effective. So if you look at this graph, this shows kind of the best human machine learning experts, computer vision experts, machine learning researchers in the world, producing a whole bunch of different kinds of models in the last four or five years, things like ResNet 50, DenseNet-201, Inception-ResNet, all kinds of things. That black dotted line is kind of the frontier of human machine learning expert model quality on the y-axis and computational cost on the x-axis. So what you see is as you go out the x-axis, you tend to get more accuracy because you're applying more computational cost. But what you see is the blue dotted line is AutoML-based solutions, systems where we've done this automated experimentation instead of pre-designing any particular architecture. And you see that it's better both at the high end, where you care about the most accurate model you can get, regardless of computational cost, but it's also accurate at the low end, where you care about a really lightweight model that might run on a phone or something like that. And in 2019, we've actually been able to improve that significantly. This is a set of models called Efficient Net and it has a very kind of a slider about you can trade off computational cost and accuracy. But they're all way better than human sort of guided experimentation on the black dotted line there. And this is true for image recognition, for [INAUDIBLE].. It's true for object detection. So the red line there is AutoML. The other things are not. It's true for language translation. So the black line there is various kinds of Transformers. The red line is we gave the basic components of Transformers to an AutoML system and allowed it to fiddle with it and come up with something better. It's true for computer vision models used in autonomous vehicles. So this was a collaboration between Waymo and Google Research. We were able to come up with models that were significantly lower latency for the same quality, or they could trade it off and get significantly lower error rate at the same latency. It actually works for tabular data. So if you have lots of customer records, and you want to predict which customers are going to be spending $1,000 with your business next month, you can use AutoML to come up with a high quality model for that kind of problem. OK. So what do we want? I think we want the following properties in a machine learning model. So one is we tend to train separate models for each different problem we care about. And I think this is a bit misguided. Like, really, we want one model that does a lot of things so that it can build on the knowledge in how it does thousands or millions of different things, so that when the million and first thing comes along, it can actually use its expertise from all the other things it knows how to do to know how to get into a good state for the new problem with relatively little data and relatively little computational cost. So these are some nice properties. I have kind of a cartoon diagram of something I think might make sense. So imagine we have a model like this where it's very sparsely activated, so different pieces of the model have different kinds of expertise. And they're called upon when it makes sense, but they're mostly idle, so it's relatively computationally [INAUDIBLE] power efficient. But it can do many things. And now, each component here is some piece of machine learning model with different kinds of state, parameters in the model, and different operations. And a new task comes along. Now you can imagine something like neural architecture search becoming-- squint at it just right and now turn it into neural pathway search. We're going to look for components that are really good for this new task we care about, and maybe we'll search and find that this path through the model actually gets us into a pretty good state for this new task. Because maybe it goes through components that are trained on related tasks already. And now maybe we want that model to be more accurate for the purple task, so we can add a bit more computational capacity, add a new component, start to use that component for this new task, continue training it, and now, that new component can also be used for solving other related tasks. And each component itself might be running some sort of interesting architectural search inside it. So I think something like that is the direction we should be exploring as a community. It's not what we're doing today, but I think it could be a pretty interesting direction. OK, and finally, I'd like to touch on thoughtful use of AI in society. As we've seen more and more uses of machine learning in our products and around the world, it's really, really important to be thinking carefully about how we want to apply these technologies. Like any technology, these systems can be used for amazing things or things we might find a little detrimental in various ways. And so we've come up with a set of principles by which we think about applying sort of machine learning and AI to our products. And we've made these public about a year and a half ago as a way of sort of sharing our thought process with the rest of the world. And I particularly like these. I'll point out many of these are sort of areas of research that are not fully understood yet, but we aim to apply the best in the state of the art methods, for example, for reducing bias in machine learning models, but also continue to do research and advance the state of the art in these areas. And so this is just kind of a taste of different kinds of work we're doing in this area-- how do we do machine learning with more privacy, using things like federated learning? How do we make models more interpretable so that a clinician can understand the predictions it's making on diabetic retinopathy sort of examples? How do we make machine learning more fair? OK, and with that, I hope I've convinced you that deep neural nets and machine learning-- you're already here, so maybe you're already convinced of this-- but are helping make sort of significant advances in a lot of hard computer science problems, computer vision, speech recognition, language understanding. General use of machine learning is going to push the world forward. So thank you very much, and I appreciate you all being here. [APPLAUSE] MEGAN KACHOLIA: Hey, everyone. Good morning. Just want to say, first of all, welcome. Today, I want to talk a little bit about TensorFlow 2.0 and some of the new updates that we have that are going to make your experience with TensorFlow even better. But before I dive into a lot of those details, I want to start off by thanking you, everyone here, everyone on the livestream, everyone who's been contributing to TensorFlow, all of you who make up the community. TensorFlow was open source to help accelerate the AI field for everyone. You've used it in your experiments. You've deployed in your businesses. You've made some amazing different applications that we're so excited to showcase and talk about, some that we get to see a bit here today, which is one of my favorite parts about conferences like this. And you've done so much more. And all of this has helped make TensorFlow what it is today. It's the most popular ML ecosystem in the world. And honestly, that would not happen without the community being excited and embracing and using this and giving back. So on behalf of the entire TensorFlow team, I really just first want to say thank you because it's so amazing to see how TensorFlow is used. That's one of the greatest things I get to see about my job, is the applications and the way folks are using TensorFlow. I want to take a step back and talk a little bit about some of the different user groups and how we see them making use of TensorFlow. TensorFlow was being used across a wide range of experiments and applications. So here, calling out researchers, data scientists and developers, and there's other groups kind of in-between as well. Researchers use it because it's flexible. It's flexible enough to experiment with and push the state-of-the-art deep learning. You heard this even just a few minutes ago, with folks from Twitter talking about how they're able to use TensorFlow and expand on top of it in order to do some of the amazing things that they want to make use of on their own platform. And at Google, we see examples of this when researchers are creating advanced models like Excel NAT and some of the other things that Jeff referenced in his talk earlier. Taking a step forward, looking at data scientists, data scientists and enterprise engineers have said they rely on TensorFlow for performance and scale in training and production environments. That's one of the big things about TensorFlow that we've always emphasized and looked at from the beginning. How can we make sure this can scale to large production use cases? For example, Quantify and BlackRock use TensorFlow to test and deploy BERT in real world NLP instances, such as text tokenization, as well as classification. Hopping one step forward, looking a bit at application developers, application developers use TensorFlow because it's easy to learn ML on the platforms that they care about. Arduino wants to make ML simple on microcontrollers, so they rely on TensorFlow pre-trained models and TensorFlow Lite Micro for deployment. Each of these groups is a critical part of the TensorFlow ecosystem. And this is why we really wanted to make sure that TensorFlow 2.0 works for everyone. We announced the alpha at our Dev Summit earlier this year. And over the past few months, the team has been working very hard to incorporate early feedback. Again, thank you to the community for giving us that early feedback, so we can make sure we're developing something that works well for you. And we've been working to resolve bugs and issues and things like that. And just last month in September, we were excited to announce the final general release for TensorFlow 2.0. You might be familiar with TensorFlow's architecture, which has always supported the ML lifecycle from training through deployment. Again, one of the things we've emphasized since the beginning when TensorFlow was initially open sourced a few years ago. But I want to emphasize how TensorFlow 2.0 makes this workflow even easier and more intuitive. First, we invested in Keras, an easy-to-use package in TensorFlow, making it the default high level API. Many developers love Keras because it's easy to use and understand. Again, you heard this already mentioned a little bit earlier, and hopefully, we'll hear more about it throughout the next few days. By tightly integrating Keras into 2.0, we can make Keras work even better with primitives like TF data. We can do performance optimizations behind the scenes and run distributed training. Again, we really wanted 2.0 to focus on usability. How can we make it easier for developers? How can we make it easier for users to get what they need out of TensorFlow? For instance, Lose It, a customized weight loss app, said they use tf.keras for designing their network. By leveraging [INAUDIBLE] strategy distribution in 2.0, they were able to utilize the full power of their GPUs. It's feedback like this that we love to hear, and again, it's very important for us to know how the community is making use of things, how the community is using 2.0, the things they want to see, so that we can make sure we're developing the right framework and also make sure you can contribute back. When you need a bit more control to create advanced algorithms, 2.0 comes fully loaded with eager execution, making it familiar for Python developers. This is especially useful when you're stepping through, doing debugging, making sure you can really understand step by step what's happening. This also means there's less coding required when training your model, all without having to use session.run. Again, usability is a focus. To demonstrate the power of training models with 2.0, I'll show you how you can train a state-of-the-art NLP model in 10 lines of code, using the Transformers NLP library by Hugging Face-- again, a community contribution. This popular package hosts some of the most advanced NLP models available today, like BERT, GPT, Transformer-XL, XLNet, and now supports TensorFlow 2.0. So let's take a look. Here, kind of just looking through the code, you can see how you can use 2.0 to train Hugging Face's DistilBERT model for text classification. You can see just simply load the tokenizer, model, and the data set. Then prepare the data set and use tf.keras compile and fit APIs. And with a few lines of code, I can now train my model. And with just a few more lines, we can use the train model for tasks such as text classification using eager execution. Again, it's examples like this where we can see how the community takes something and is able to do something very exciting and amazing by making use of the platform and the ecosystem that TensorFlow is providing. But building and training a model is only one part of TensorFlow 2.0. You need the performance to match. That's why we worked hard to continue to improve performance with TensorFlow 2.0. It delivers up to 3x faster training performance using mixed precision on NVIDIA Volta and Turing GPUs in a few lines of code with models like ResNet-50 and BERT. As we continue to double down on 2.0 in the future, performance will remain a focus with more models and with hardware accelerators. For example, in 2.1, so the next upcoming TensorFlow release, you can expect TPU and TPU pod support, along with mixed precision for GPUs. So performance is something that we're keeping a focus on as well, while also making sure usability really stands to the forefront. But there's a lot more to the ecosystem. So beyond model building and performance, there are many other pieces that help round out the TensorFlow ecosystem. Add-ons and extensions are a very important piece here, which is why we wanted to make sure that they're also compatible with TensorFlow 2.0. So you can use popular libraries, like some other ones called out here, whether it's TensorFlow Probability, TF Agents, or TF Text. We've also introduced a host of new libraries to help researchers and ML practitioners in more useful ways. So for example, neural structure learning helps to train neural networks with structured signals. And the new Fairness Indicators add-on enables regular computation and visualization of fairness metrics. And these are just the types of things that you can see kind of as part of the TensorFlow ecosystem, these add-ons that, again, can help you make sure you're able to do the things you need to do not with your models, but kind of beyond just that. Another valuable aspect of the TensorFlow ecosystem is being able to analyze your ML experiments in detail. So this is showing TensorBoard. TensorBoard is TensorFlow's visualization toolkit, which is what helps you accomplish this. It's a popular tool among researchers and ML practitioners for tracking metrics, visualizing model graphs and parameters, and much more. It's very interesting that we've seen users enjoy TensorBoard so much, they'll even take screenshots of their experiments and then use those screenshots to be able to share with others what they're doing with TensorFlow. This type of sharing and collaboration in the ML community is something we really want to encourage with TensorFlow. Again, there's so much that can happen by enabling the community to do good things. That's why I'm excited to share the preview of TensorBoard.dev, a new, free, managed TensorBoard experience that lets you upload and share your ML experiment results with anyone. You'll now be able to host and track your ML experiments and share them publicly. No setup required. Simply upload your logs, and then share the URL, so that others can see the experiments and see the things that you are doing with TensorFlow. As a preview, we're starting off with the [INAUDIBLE] dashboard, but over time, we'll be adding a lot more functionality to make the sharing experience even better. But if you're not looking to build models from scratch and want to reduce some computational cost, TensorFlow was always made pre-trained models available through TensorFlow Hub. And today, we're excited to share an improved experience of TensorFlow Hub that's much more intuitive, where you can find a comprehensive repository of pre-trained models in the TensorFlow ecosystem. This means you can find models like BERT and others related to image, text, video, and more that are ready to use with TensorFlow Lite and TensorFlow.js. Again, we wanted to make sure the experience here was vastly improved to make it easier for you to find what you need in order to more quickly get to the task at hand. And since TensorFlow is driven by all of you, TensorFlow Hub is hosting more pre-trained models from the community. You'll be able to find curated models by DeepMind, Google, Microsoft's AI for Earth, and NVIDIA ready to use today with many more to come. We want to make sure that TensorFlow Hub is a great place to find some of these excellent pre-trained models. And again, there's so much the community is doing. We want to be able to showcase those models as well. TensorFlow 2.0 also highlights TensorFlow's core strengths and areas of focus, which is being able to go from model building, experimentation, through to production, no matter what platform you work on. You can deploy end-to-end ML pipelines with TensorFlow Extended or TFX. You can use your models on mobile and embedded devices with TensorFlow Lite for on device inference, and you can train and run models in the browser or Node.js with TensorFlow.js. You'll learn more about what's new in TensorFlow in production during the keynote sessions tomorrow. You can learn more about these updates by going to tensorflow.org where you'll also find the latest documentation, examples, and tutorials for 2.0. Again, we want to make sure it's easy for the community to see what's happening, what's new, and enable you to just do what you need to do with TensorFlow. We've been thrilled to see the positive response to 2.0, and we hope you continue to share your feedback. Thank you, and I hope you enjoy the rest of TF World. [APPLAUSE] FREDERICK REISS: Hello, everyone. I'm Fred Reiss. I work for IBM. I've been working for IBM since 2006. And I've been contributing to TensorFlow Core since 2017. But my primary job at IBM is to serve as tech lead for CODAIT. That's the Center for Open Source Data and AI Technologies. We are an open source lab located in downtown San Francisco, and we work on open source technologies that are foundational to AI. And we have on staff 44 full-time developers who work only on open source software. And that's a lot of developers, a lot of open source developers. Or is it? Well, if you look across IBM at all of the IBM-ers who are active contributors to open source, in that they have committed code to GitHub in the last 30 days, you'll find that there are almost 1,200 IBM-ers in that category. So our 44 developers are actually a very small slice of a very large pie. Oh, and those numbers, they don't include Red Hat. When we closed that acquisition earlier this year, we more than doubled our number of active contributors to open source. So you can see that IBM is really big in open source. And more and more, the bulk of our contributions in the open are going towards the foundations of AI. And when I say AI, I mean AI in production. I mean AI at scale. AI at scale is not an algorithm. It's not a tool. It's a process. It's a process that starts with data, and then that data turns into features. And those features train models, and those models get deployed in applications, and those applications produce more data. And the whole thing starts all over again. And at the core of this process is an ecosystem of open source software. And at the core of this ecosystem is TensorFlow, which is why I'm here, on behalf of IBM open source, to welcome you to TensorFlow World. Now throughout this conference, you're going to see talks that speak to all of the different stages of this AI lifecycle. But I think you're going to see a special emphasis on this part-- moving models into production. And one of the most important aspects of moving models into production is that when your model gets deployed in a real-world application, it's going to start having effects on the real world. And it becomes important to ensure that those effects are positive and that they're fair to your clients, to your users. Now, at IBM, here's a hypothetical example that our researchers put together about a little over a year ago. They took some real medical records data, and they produced a model that predicts which patients are more likely to get sick and therefore should get additional screening. And they showed that if you naively trained this model, you end up with a model that has significant racial bias, but that by deploying state-of-the-art techniques to adjust the data set and the process of making the model, they could substantially reduce this bias to produce a model that is much more fair. You can see a Jupyter Notebook with the entire scenario from end to end, including code and equations and results, at the URL down here. Again, I need to emphasize this was a hypothetical example. We built a flawed model deliberately, so we could show how to make it better. But no patients were harmed in this exercise. However, last Friday, I sat down with my morning coffee, and I opened up the "Wall Street Journal." And I saw this article at the bottom of page three, describing a scenario eerily similar to our hypothetical. When your hypothetical starts showing up as newspapers headlines, that's kind of scary. And I think it is incumbent upon us as an industry to move forward the process, the technology of trust in AI, trust and transparency in AI, which is why IBM and IBM Research have released our toolkits of state-of-the-art algorithms in this space as open source under AI Fairness 360, AI Explainability 360, and Adversarial Robustness 360. It is also why IBM is working with other members of the Linux Foundation AI, a trusted AI committee, to move forward open standards in this area so that we can all move more quickly to trusted AI. Now if you'd like to hear more on this topic, my colleague, Animesh Singh, will be giving a talk this afternoon at 1:40 on trusted AI for the full 40 minute session. Also I'd like to give a quick shout out to my other co-workers from CODAIT who have come down here to show you cool open source demos at the IBM booth. That's booth 201. Also check out our websites, developer.ibm.com and codait.org. On behalf of IBM, I'd like to welcome you all to TensorFlow World. Enjoy the conference. Thank you. [APPLAUSE] THEODORE SUMME: Hi, I'm Ted Summe from Twitter. Before I get started with my conversation today, I want to do a quick plug for Twitter. What's great about events like this is you get to hear people like Jeff Dean talk. And you also get to hear from colleagues and people in the industry that are facing similar challenges as you and have conversations around developments in data science and machine learning. But what's great is that's actually available every day on Twitter. Twitter's phenomenal for conversation on data science and machine learning. People like Jeff Dean and other thought leaders are constantly sharing their thoughts and their developments. And you can follow that conversation and engage in it. And not only that, but you can bring that conversation back to your workplace and come off looking like a hero-- just something to consider. So without that shameless plug, my name's Ted Summe. I lead product for Cortex. Cortex is Twitter's central machine learning organization. If you have any questions for me or the team, feel free to connect with me on Twitter, and we can follow up later. So before we get into how we're accelerating ML at Twitter, let's talk a little bit about how we're even using ML at Twitter. Twitter is largely organized against three customer needs, the first of which is our health initiative. That might be a little bit confusing to you. You might think of it as user safety. But we think about it as improving the health of conversations on Twitter. And machine learning is already at use here. We use it to detect spam. We can algorithmically and at scale detect spam and protect our users from it. Similarly, in the abuse space, we can proactively flag content as potentially abuse, toss it up for human reviews, and act on it before our users even get impacted by it. A third space where we're using machine learning here is something called NSFW, Not Safe For Work. I think you're all familiar with the acronym. So how can we, at scale, identify this content and handle it accordingly? Another use of machine learning in this space. There's more that we want to do here, and there's more that we're already doing. Similarly, the consumer organization-- this is largely what you think of, the big blue app of Twitter. And here, the customer job that we're serving is helping connect our customers with the conversations on Twitter that they're interested in. And one of the primary veins in which we do this is our timeline. Our timeline today is ranked. So if you're not familiar, users follow accounts. Content and tweets associated with those accounts get funneled into a central feed. And we rank that based on your past engagement and interest to make sure we bring forth the most relevant conversations for you. Now, there's lots of conversations on Twitter, and you're not following everyone. And so there's also a job that we have to serve about bringing forth all the conversations that you're not proactively following, but are still relevant to you. This has surfaced in our Recommendations product, which uses machine learning to scan the corpus of content on Twitter, and identify what conversations would be most interesting to you, and push it to you in a notification. The inverse of that is when you know what the topics you want to explore are, but you're looking for the conversations around that. That's where we use Twitter Search. This is another surface area in the big blue app that we're using machine learning. The third job to be done for our customers is helping connect brands with their customers. You might think of this as the ads product. And this is actually the OG of machine learning at Twitter, the first team that implemented it. And here, we use it for what you might expect, ads ranking. That's kind of like the timeline ranking, but instead of tweets, it's ads and identifying the most relevant ads for our users. And as signals to go into that, we also do user targeting to understand your past engagement ads, understand which ads are in your interest space. And the third-- oh. Yeah, we're still good. And the third is brand safety. You might not think about this when you think about machine learning and advertising. But if you're a company like United and you want to advertise on Twitter, you want to make sure that your ad never shows up next to a tweet about a plane crash. So how do we, at scale, protect our brands from those off-brand conversations? We use machine learning for this as well. So as you can tell, machine learning is a big part of all of these organizations today. And where we have shared interests and shared investment, we want to make sure we have a shared organization that serves that. And that's the need for Cortex. Cortex is Twitter's central machine learning team, and our purpose is really quite simple-- to enable Twitter with ethical and advanced AI. And to serve that purpose, we've organized in three ways. The first is our applied research group. This group applies the most advanced ML techniques from industry and research to our most important surface areas, whether they be new initiatives or existing places. This team you can kind of think of as like an internal task force or consultancy that we can redeploy against the company's top initiatives. The second is signals. When using machine learning, having shared data assets that are broadly useful can provide us more leverage. Examples of this would be our language understanding team that looks at tweets and identifies named entities inside them. Those can then be offered up as features for other teams to consume in their own applications of machine learning. Similarly, our media understanding team looks at images and can create a fingerprint of any image. And therefore, we can identify every use of that image across the platform. These are examples of shared signals that we're producing that can be used for machine learning at scale inside the company. And the third organization is our platform team. And this is really the origins of Cortex. Here, we provide tools and infrastructure to accelerate ML development at Twitter, increase the velocity of our ML practitioners. And this is really the focus of the conversation today. When we set out to build this ML platform, we decided we wanted a shared ML platform across all of Twitter. And why is that important that it be shared across all of Twitter? Well, we want transferability. We want the great work being done in the ads team to be, where possible, transferable to benefit the health initiative where that's relevant. And similarly, if we have great talent in the consumer team that's interested in moving to the ads team, if they're on the same platform, they can transfer without friction and be able to ramp up quickly. So we set out with this goal of having a shared ML platform across all of Twitter. And when we did that, we looked at a couple product requirements. First, it needs to be scalable. It needs to be able to operate at Twitter scale. The second, it needs to be adaptable. This space is developing quickly so we need a platform that can evolve with data science and machine learning developments. Third is the talent pool. We want to make sure that we have a development environment at Twitter that appeals to the ML researchers and engineers that we're hiring and developing. Fourth is the ecosystem. We want to be able to lean on the partners that are developing industry leading tools so that we can focus on technologies that are Twitter specific. Fourth is documentation. You ought to understand that. We want to be able to quickly unblock our practitioners as they hit issues, which is inevitable in any platform. And finally, usability. We want to remove friction and frustration from the lives of our team, so that they can focus on delivering value for our end customers. So considering these product requirements, let's see how TensorFlow is done against them. First is scalability. We validated this by putting TensorFlow by way of our implementation we called Deep Bird against timeline ranking. So every tweet that's ranked in the timeline today runs through TensorFlow. So we can consider that test validated. Second is adaptability. The novel architectures that TensorFlow can support, as well as the custom lost functions, allows us to react to the latest research and employ that inside the company. An example that we published on this publicly is our use of a SplitNet architecture and ads ranking. So TensorFlow has been very adaptable for us. Third is the talent pool, and we think about the talent pool in kind of two types. There's the ML engineer and the ML researcher. And as a proxy of these audiences, we looked at the GitHub data on these. And clearly, TensorFlow is widely adopted amongst ML engineers. And similarly, the archive community shows strong evidence of wide adoption in the academic community. On top of this proxy data, we also have anecdotal evidence of the speed of ramp-up for ML researchers and ML engineers inside the company. The fourth is the ecosystem. Whether it's TensorBoard, TF Data Validation, TF Model Analysis, TF Metastore, TF Hub, TFX Pipelines, there's a slew of these products out there, and they're phenomenal. They allow us to focus on developing tools and infrastructure that is specific to Twitter's needs and lead on the great work of others. So we're really grateful for this, and TensorFlow does great here. Fifth being documentation. Now, this is what you would go to when you go to TensorFlow, and you see that phenomenal documentation, as well as great education resources. But what you might not appreciate and what we've come to really appreciate is the value of the user generated content. What Stack Overflow and other platforms can provide in terms of user generated content is almost as valuable as anything TensorFlow itself can create. And so TensorFlow, given its widespread adoption, its great TensorFlow website, has provided phenomenal documentation for ML practitioners. Finally, usability. And this is why we're really excited about TensorFlow 2.0. The orientation around the carrier's API makes it more user friendly. It also still continues to allow for flexibility for more advanced users. The eager execution enables more rapid and intuitive debugging, and it closes the gap between ML engineers and modelers. So clearly from this checklist, we're pretty happy with our engagement with TensorFlow. And we're excited about continuing to develop the platform with them and push the limits on what it can do, with gratitude to the community for their participation and involvement in the product and appreciate their conversation on Twitter, as we advance it. So if you have any questions for me, as I said before, you can connect with me, but I'm not alone here today. A bunch of my colleagues are here as well. So if you see them roaming the halls, feel free to engage with them. Or as I shared before, you can continue the conversation on Twitter. Here are their handles. Thank you for your time. Cheers. [APPLAUSE] CRAIG WILEY: I just want to begin by saying I've been dabbling in Cloud AI and Cloud Machine Learning for a while. And during that time, it never occurred to me that we'd be able to come out with something like we did today because this is only possible because Google Cloud and TensorFlow can collaborate unbelievably closely together within Google. So to begin, let's talk a little bit about TensorFlow-- 46 million downloads. TensorFlow has been massive growth the last few years. It's expanded from the forefront of research, which we've seen earlier this morning, to businesses taking it on as a dependency for their business to operate on a day in, day out basis. It's a super exciting piece. As someone who spends most all of their time thinking about how we can bring AI and machine learning into businesses, seeing TensorFlow's commitment and focus on deploying actual ML in production is super exciting to me. With this growth, though, comes growing pains. And part of that is things like support, right? When my model doesn't do what I expected it to or my training job fails, what options do I have? And how well does your boss respond when you say, hey, yes, I don't know why my model's not training, but not to worry, I've put a question on Slack. And hopefully, someone will get back to me. We understand that businesses who are taking a bet on TensorFlow as a critical piece of their hardware architecture or their stack need more than this. Second, it can be a challenge to unlock the scale and performance of cloud. For those of you who, like me, have gone through this journey over the last couple of years, for me, it started on my laptop. Right? And then eventually, I outgrew my laptop, and so I had a gaming rig under my desk, right? With the GPU and eventually, there were eight gaming rigs under my desk. And when you opened the door to my office, the whole floor knew because it sounded like [INAUDIBLE].. Right? And but now with today's cloud, that doesn't have to be the case. You can go from that single instance all the way up to a massive scale seamlessly. So with that, today, we bring you TensorFlow Enterprise. TensorFlow Enterprise is designed to do three things-- one, give you Enterprise grade support; two, cloud scale performance; and three, managed services when and where you want them, at the abstraction level you want them. Enterprise grade support, what does that mean? Fundamentally what that means is that as these businesses take a bet on TensorFlow, many of these businesses have IT policies or requirements that the software have a certain longevity before they're willing to commit to it in production. And so today, for certain versions of TensorFlow, when used on Google Cloud, we will extend that one year of support a full three years. That means that if you're building models on 1.15 today, you can know that for the next three years, you'll get bug fixes and security patches when and where you need them. Simple and scalable. Scaling from an idea on a single node to production at massive scale can be daunting, right? Saying to my boss, hey, I took a sample of the data was something that previously seemed totally reasonable, but now we're asked to train on the entire corpus of data. And that can take days, weeks. We can help with all of that by deploying TensorFlow on Google Cloud, a network that's been running TensorFlow successfully for years and has been highly optimized for this purpose. So scalable across our world class architecture, the products are compatibility tested with the cloud, their performance optimized for the cloud and for Google's world class infrastructure. What does this mean? So if any of you have ever had the opportunity to use BigQuery, BigQuery is Google Cloud's kind of massively parallel cloud hosted data warehouse. And by the way, if you haven't tried using BigQuery, I highly recommend going out and trying it. It returns results faster than can be imagined. That speed in BigQuery, we wanted to make sure we were taking full advantage of that. And so recent changes and recent pieces included in TensorFlow Enterprise have increased the speed of the connection between the data warehouse and TensorFlow by three times. Right? Now, all of sudden, those jobs that were taking days take hours. Unity gaming, wonderful customer and partner with us. You can see the quote here. Unity leverages these aspects of TensorFlow Enterprise in their business. Their monetization products reach more than three billion devices-- three billion devices worldwide. Game developers rely on a mix of scale and products to drive installs and revenue and player engagement. And Unity needs to be able to quickly test, build, scale, deploy models all at massive scale. This allows them to serve up the best results for their developers and their advertisers. Managed services. As I said, TensorFlow Enterprise will be available on Google Cloud and will be available as part of Google Cloud's AI platform. It will also be available in VMs if you'd prefer that, or in containers if you want to run them on Google Cloud Kubernetes Engine, or using Kubeflow on Kubernetes Engine. In summary, TensorFlow Enterprise offers Enterprise grade support-- that continuation, that full three years of support that IT departments are accustomed to-- cloud scale performance so that you can run at massive scale, and works seamlessly with our managed services. And all of this is free and fully included for all Google Cloud users. Google Cloud becomes the best place to run TensorFlow. But there's one last piece, which is for companies for whom AI is their business-- not companies for whom AI might help with this part of their business or that or might help optimize this campaign or this backend system, but for companies where AI is their business, right? Where they're training hundreds of thousands of hours of training a year, petabytes of data, right? Using cutting edge models to meet their unique requirements, we are introducing TensorFlow Enterprise with white-glove support. This is really for cutting edge AI, right? Engineering to engineering assistance when needed. Close collaboration across Google allows us to fix bugs faster if needed. One of the great opportunities of working in cloud, if you ask my kids, they'll tell you that the reason I work in cloud AI and in kind of machine learning is in an effort to keep them ever from learning to drive. They're eight and 10 years old, so I need people to kind of hurry along this route, if you will. But one of the customers and partners we have is Cruise Automotive. And you can see here, they're a shining example of the work we're doing. On their quest towards self-driving cars, they've also experienced hiccups and challenges and scaling problems. And we've been a critical partner for them in helping ensure that they can achieve the results they need to, to solve this kind of generational defining problem of autonomous vehicles. You can see not only did we improve the accuracy of their models, but also reduce training times from four days down to one day. This allows them to iterate at speeds previously unthinkable. So none of this, as I said, would have been possible without the close collaboration between Google Cloud and TensorFlow. I look back on Megan's recent announcement of TensorBoard.dev. We will be looking at bringing that type of functionality into an enterprise environment as well in the coming months. But we're really, really excited to get TensorFlow Enterprise into your hands today. To learn more and get started, you can go to the link, as well as sessions later today. And if you are on the cutting edge of AI, we are accepting applications for the white-glove service as well. We're excited to bring this offering to teams. We're excited to bring this offering to businesses that want to move into a place where machine learning is increasingly a part of how they create value. Thank you very much for your time today KEMAL EL MOUJAHID: Hi, my name is Kemal. I'm the product director for TensorFlow. So earlier, you heard from Jeff and Megan about the prod direction. Now, what I'd like to talk about is the most important part of what we're building, and that's the community. That's you. Sorry. Where's the slide? Thank you. So as you've seen in the video, we've got a great roadshow, 11 events spanning five continents to connect the community with the TensorFlow team. I, personally, was very lucky this summer, because I got to travel to Morocco and Ghana and Shanghai, amongst other places, just to meet the community, and to listen to your feedback. And we heard a lot of great things. So as we're thinking about, how can we best help the community? It really came down to three things. First, we would like to help you to connect with the larger community, and to share the latest and greatest of what you've been building. Then, we also would like you-- we want to help you learn, learn about ML, learn about TensorFlow. And then, we want to help you contribute and give back to the community. So let's start with Connect. So why connect? Well, first the community-- the TensorFlow community has really grown a lot. It's huge-- 46 million downloads, 2,100 committers, and-- again, I know that we've been saying that all along, but I really want to say a huge thank you on behalf of the TensorFlow team for making the community what it is today. Another aspect of the community that we're very proud of is that it's truly global. This is a revised map of our GitHub stars. And, as you can see, we're covering all time zones and we keep growing. So the community is huge. It's truly global. And we really want to think about, how can we bring the community closer together? And this is really what initiated the idea of TensorFlow World. We wanted to create an event for you. We wanted an event where you could come up and connect with the rest of the community, and share what you've been working on. And this has actually started organically. Seven months ago, the TensorFlow User Groups started, and I think now we have close to 50. The largest one is in Korea. It has 46,000 members. We have 50 in China. So if you're in the audience or in the livestream, and you're looking to this map, and you're thinking, wait, I don't see a dot where I live-- and you have a TensorFlow member that you're connecting with, and you want to start a TensorFlow User Group-- well, we'd like to help you. So please go to tensorflow.org/community, and we'll help you get it started. So that next year, when we look at this map, we have dots all over the place. So what about businesses? We've talked about developers. What about businesses? One thing we heard from businesses is they have this business problem. They think ML can help them, but they're not sure how. And that's a huge missed opportunity when we look at the staggering $13 trillion that AI will bring to the global economy over the next decade. So you have those businesses on one side, and then you have partners on the other side, who know about ML, they know how to use TensorFlow, so how do we connect those two? Well, this was the inspiration for launching our Trusted Partner Pilot Program, which helps you, as a business, connect to a partner who will help you solve your ML problem. So if you go on tensorflow.org, you'll find more about our Trusted Partner program. Just a couple of examples of cool things that they've been working on. One partner helped a car insurance company shorten the insurance claim processing time using image processing techniques. Another partner helped the global med tech company by automating the shipping labeling process using object recognition techniques. And you'll hear more from these partners later today. I encourage you to go check out their talks. Another aspect is that if you're a partner, and you're interested in getting into this program, we also would like to hear from you. So let's talk about Learn. We've invested a lot in producing quality material to help you learn about ML and about TensorFlow. One thing that we did over the summer, which was very exciting is for the first time, we're a part of the Google Summer of Code. We had a lot of interest. We were able to select 20 very talented students, and they got to work the whole summer with amazing mentors on the TensorFlow engineering team. And they worked on very inspiring projects going from 2.0 to Swift to JS to TF-Agents. So we were so excited with the success of this program that we decided to participate, for the first time, in a Google Code-in program. So this is the same program, but for pre-university students from 13 to 17. It's a global online contest. And it introduces teenagers to the world of contributing to open source development. So as I mentioned, we've invested a lot this year on ML education material, but one thing we heard is that there's a lot of different things. And what you want is to be guided through pathways of learning. So we've worked hard on that, and we've decided to announce the new Learn ML page tensorflow.org. And what this is a learning path curated for you by the TensorFlow team, and organized by level. So you have from beginners to advanced. You can explore books, courses, and videos to help you improve your knowledge of machine learning, and use that knowledge and use TensorFlow, to solve your real-world problem. And for more exciting news that will be available on the website, I'd like to play a brief video by a friend Andrew Ng. [VIDEO PLAYBACK] - Hi, everyone. I'm in New York right now, and wish I could be there to enjoy the conference. But I want to share with you some exciting updates. Deeplearning.ai started a partnership with the TensorFlow team with a goal of making world-class education available for developers on the Coursera platform. Since releasing the Deep Learning Specialization, I've seen so many of you, hundreds of thousands, learn the fundamental skills of deep learning. I'm delighted we've been able to complement that with the TensorFlow in Practice Specialization to help developers learn how to build ML applications for computer vision, NLP, sequence models, and more. Today, I want to share with you an exciting new project that the deeplearning.ai and TensorFlow teams have been working on together. Being able to use your models in a real-world scenario is when machine learning gets particularly exciting. So we're producing a new four-course specialization called TensorFlow Data and Deployment that will let you take your ML skills to the real world, deploying models to the web, mobile devices, and more. It will be available on Coursera in early December. I'm excited to see what you do with these new resources. Keep learning. [END PLAYBACK] KEMAL EL MOUJAHID: All right. This is really cool. Since we started working on these programs, it's been pretty amazing to see hundreds of thousands of people take those courses. And the goal of these educational resources is to let everyone participate in the ML revolution, regardless of what your experience with machine learning is. And now, Contribute. So a great way to get involved is to connect with your GDE. We now have 126 machine learning GDEs globally. We love our GDEs. They're amazing. They do amazing things for the community. This year alone, they gave over 400 tech talks, 250 workshops. They wrote 221 articles reaching tens of thousands of developers. And one thing that was new this year is that they helped with doc sprints. So docs are really important. They're critical, right? You really need good quality docs to work on machine learning, and often the documentation is not available in people's native languages. And so this is why when we partnered with our GDEs, we launched the doc sprints. Over 9,000 API docs were updated by members of the TensorFlow community in over 15 countries. We heard amazing stories of power outage, and power running out, and people coming back later to finish a doc sprint, and actually writing docs on their phones. So if you've been helping with docs, thank you, if you're in the room, if you're over livestream, thank you so much. If you're interested in helping translate documentation in your native language, please reach out, and we'll help you organize a doc sprint. Another thing that the GDEs help with is experimenting with the latest features. So I want to call out Sam Witteveen, an ML GDE from Singapore, who's already experimenting with 2.x TPUs, and you can hear him talk later today to hear about his experience. So if you want to get involved, please reach out to your GDE and start working on TensorFlow. Another really great way to help is to join a SIG. A SIG is a Special Interest Group, and it helps you work on the things that you're the most excited about on TensorFlow. We have, now, 11 SIGs available. Addons, IO, and Networking, in particular, really supported the transition to 2.0 by embracing the parts of contrib and putting them into 2.0. And SIG Build ensures that TF runs well everywhere on any OS, any architecture, and plays well with the Python library. And we have many other really exciting SIGs, so I really encourage you to join one. Another really great way to contribute is through competition. And for those of you who were there at the Dev Summit back in March, we launched our 2.0 challenge on DevPost. And the grand prize was an invitation to this event, TensorFlow World. And so we would like to honor our 2.0 Challenge winners, and I think we are lucky to have two of them in the room-- Victor and Kyle, if you're here. [APPLAUSE] So Victor worked on Handtrack.js, a library for prototyping hand gesture in the browser. And then Kyle worked on a Python 3 package to simulate N-body, to generate N-body simulations. So one thing we heard, too, during our travels is, oh, that hackathon was great, but I totally missed it. Can we have another one? Well, yes. Let's do another one. So if you go on tfworld.devpost.com, we're launching a new challenge. You can apply your 2.0 skills and share the latest and greatest, and win cool prizes. So we're really excited to see what you're going to build. Another great community that we're very excited to partner with is Kaggle. So we've launched a contest on Kaggle to challenge you with question answering model based on Wikipedia articles. You can put your natural language processing skills to the test and earn $50,000 in prizes. It's open for entry until January 22, so best of luck. So we have a few action items for you, and they're listed on this slide. But remember, we created TensorFlow World for you, to help you connect and share what you've been working on. So our main action item for you in the next two days is really to get to know the community better. And with that, I'd like to thank you, and I hope you enjoy the rest of TF World. Thank you. [APPLAUSE]
B1 中級 TensorFlow World 2019主題演講 (TensorFlow World 2019 Keynote) 3 0 林宜悉 發佈於 2021 年 01 月 14 日 更多分享 分享 收藏 回報 影片單字