字幕列表 影片播放 列印英文字幕 (bell dings) - Hello, and welcome to another Coding Train video about neuroevolution. So this video, what I'm going to do is, I'm going to take my previous, finished, sort of finished, version of neuroevolution, the process of evolving the optimal weights of a neural network to solve some kind of, perform some kind of machine learning task. In this case, I'm making a guess based on the environment of this very crude Flappy Bird game. I'm not making this, the neural network is making a guess whether it should jump or not. So this is a very simple, actually, scenario to solve. And so it's a nice demonstration of the idea. I've been meaning to return to this topic for quite a while so I can try to look at some more complex scenarios. So neuroevolution can be applied to a lot of problems, at different, reinforcement learning algorithms are also applied too, but it's a pretty different technique. There's like five, or six, or seven videos all about neuroevolution that I would recommend you check out, or you can start right here. Because what I'm going to do in this video is, I am going to take the existing version that I made, which used a toy neural network JavaScript library that I implemented in yet another video series to, kind of, learn more about neural networks and spin up a very, very basic one in JavaScript without other dependencies. But I'm going to replace that now with TensorFlow.js. What you're seeing here is the last example from Part Five, where I saved a trained model and I'm loading it into a version of the game and watching it play out. What I did previous to that, is what you're seeing right here, which is, this is launching 500 random neural networks to try to make guesses. I can use this slider to sort of speed up the system, and over time those neural networks are going to perform the processes of crossover and mutation. It's not actually doing crossover, it's doing crossover which would be the act of combining two agents' genetic material together into one. I'm just taking one and copying it, but then I am mutating it to optimally search for the configuration of weights that will work the best. And just to remind you, when I'm talking about the weights, this is what I'm talking about. So we have the Flappy Bird game, this is the agent that needs to make a decision. In a way it's a classification problem. Should I go up or should I not go up, right? These are the only two possibilities, a very simple classification problem. Two categories. I have done, I'm the human being, have basically done the feature extraction. I could use the environment, the input to the neural network as this image, like all the pixels of the game, that would be a wonderful thing to try and I would love to do that, especially once I have convolutional layers with TensorFlow.js, which is not something I have in my toy neural network library. You might not know what a convolutional layer is. Don't worry, if I use it, I will explain it. But, I have done that feature extraction. So I have decided that the features that I want to use as inputs into my neural network are, I think it was like, the bird y position, the bird, y velocity, the top pipe location, the bottom pipe location. And then, I'll call this x, the distance to the nearest pipe. So these are what I've decided might be all the important values to use from this game, to feed into the neural network. So that means the neural network has one, two, three, four, five inputs. These all get normalized into a value with a range between zero and one, and fed into the neural network. Then the outputs is a classification problem. So there are just two outputs, and they would each output a number. If I got something like 0.8 here and 0.2 here, that means there is an 80%, basically, an 80% probability confidence score that I should jump, so I will jump. Now I could actually pick random numbers and that kind of thing, but I'm just going to take the highest one, the arg max, so to speak, and go. So, neural networks are able to learn a sophisticated amount of information through hidden layers. So the inputs don't pass directly to the outputs, but the inputs pass through a hidden layer. And I don't recall what I picked as the configuration of the hidden layer, but I think it was something like eight. So if there were eight, I think that's what it is. We'll look. One, two, three, four, five, six, seven, eight, these are what are known as dense layers. Meaning, every single node is connected to every node. I will not sit here and draw all of them, but I will start that process. And then all of these, sorry, are connected to the outputs, like this. So a neural network's core data, the configuration of that data is in a matrix. 'Cause every one of these connections has a weight. So if there are five here and eight here, that is 40 weights. If there are eight here and two here, that's 16 weights. Now the truth of the matter is, there's also a bias. So there's a bias for each one of these and a bias for each one of these. So the configuration is all of the weights, all of the biases for each of these nodes. But these are the details that I cover a lot more in the series about building a neural network from scratch in JavaScript. So if you want to learn about that, you can go back. But all you need to know for here, is that I have a way of creating this neural network. I have a way of feeding these values into it and looking at the stuff that comes out. So let's go look at that code. So if I go look at the code, we can see that here, there's this idea of a bird object. And the bird object makes a neural network with five inputs, eight hidden nodes, and two outputs. I don't know what I've said here, but this is known as a feedforward neural network. Multilayer Perceptron, it's two layers. The inputs, it looks like a layer, but these are just numbers coming in. There is a hidden layer in the output layer. That's important terminology. So that's happening here, and you can see in the bird, the bird sort of like, takes all of those properties, its y position, the closest pipe, the closest pipe's x position, its velocity, makes that into an array, feeds that into a predict function, and then I just sort of like, if the first output is greater then the second output, jump. So this is that process. And all of this happens in this neural network library. So there is neuralnetwork.js and matrix.js. This is like if TensorFlow.js was made by one person who really didn't know what they were doing (laughs). That's what's in here. So what I want to do, is I am going to delete this folder. Right, I'm deleting it, it's gone. Gone, move to trash, boom! So now when I go over here and I hit refresh, of course, it doesn't work. Because we have no neuralnetwork.js. Can I get this, I don't have a watch on, in however long I have right now, in the course of this video, working again without my library, but with TensorFlow.js instead. That's the thought experiment here. So here is my html file, and you can see that it has the P5 libraries, it's loading my neural network library, and then these are all the files for the Flappy Bird game itself, as well as the file that includes some information about how the genetic algorithm works. So what I want to do, I need to take these out, and then I want to import TensorFlow.js. So I have the TensorFlow.js website up and this is the script tag for importing TensorFlow.js. So I'm going to add that here, and actually I am pretty sure that the most current version is 1.0.4, so let's add that. And now I'm going to go back here and hit refresh. Now, of course, ah, NeuralNetwork is not defined, so we have fewer error messages here now, which is good. I just want to make sure tf is loaded, so, for example, I can call tf memory, and there is no memory being used, but I can see that tf js is loaded, 'cause I can call tf functions in the console now, which I'm going to need to do to figure all this out. All right, so now the thing I think I want to do is that, actually, I don't need to do this, but I think it would be nice for me to create, I'm going to create a file called nn.js, and I'm going to make a little wrapper for a neural network. And actually one of the reasons why I want to do this is, as you know, I've made a lot of videos using a library called ML5, which is a wrap around TensorFlow.js, which allows you to work with some of the machine learning algorithms without having to manipulate the lower level details of TensorFlow.js, and so this is, kind of, a little bit, of maybe a preview of that. Because this ultimately, this idea of neuroevolution is something that I would like to work on putting into the ML5 library. So if I make a neural network class, and I write a constructor, we know here, that in the bird object what it's doing is, it's making a neural network with five inputs, eight hidden nodes, and two outputs. So there's no reason why I can't actually just keep that same structure, and I'm going to have three arguments, I'm just going to call them a, b, and c, because at some point I might need to call the constructor in different ways, and I'm going to then say, this.input_nodes equals a. this.hidden_nodes equals b, and this.output_nodes equals c. Then I'm going to put, I already, by the way, I did this in a coding class that I'm teaching, last week, so I kind of have quite a bit of the sense of the plan here. So I'm also going to write a function called this.createModel 'cause I might need to do that in different places and then I'm going to make a separate function called createModel. So this now is the function, where I want to create my neural network, using TensorFlow.js. And I'm going to use the layers API. The truth of the matter is, this particular architecture, right, this is perhaps one of the simplest, the most basic, vanilla, so to speak, neural network architectures. It's got two layers, one hidden layer, very few nodes, there's no convolutions, no fancy stuff, just the very basic. So I could probably do this with just simple TensorFlow operations itselves, but I'm going to use something called tf.layers which is actually based on Keras, which is a Python library for TensorFlow, and Python, so many things, I've talked about these in videos. But basically tf.layers is an API that allows me to create a model from a slightly higher level perspective. So one of the key functions in tf.layers, and I'm going to create a variable called model, is tf.sequential. So if I say tf.sequential, that should create for me a sequential model. And we can just take a look at that here on the console, and see, there we go. It made something that works, you can see all these parameters that it's training. Now interestingly enough, what I'm doing in this video is not really what TensorFlow.js is designed for. Almost everything in TensorFlow.js has to do with optimization functions and algorithms for learning and tweaking weights, and back propagation. I'm not doing any of that. This neural evolution thing is quirky and weird and all I actually need is a bunch of neural networks and the ability to feed the data through them and then I can just delete that and mix them up. So actually most of this stuff will not get used. But what I want to create then, is my first layer. So the first layer, I'm going to call hidden. So I'm going to say, const hidden equals tf, I think it's dense. So I think I better look up the documentation at this point, I certainly don't have the stuff memorized, so I'm going to go to the API, tf.layer's dense. This is what I'm looking to do. And I can see a nice little example of it here. This. So I'm going to create a tf.layers dense. Dot dense. Then I need a little object which is going to contain all of the configuration information for that dense layer. And what are some things I can configure? I can give it the number of units, that's my hidden nodes, which is the hidden nodes. I also want to tell it how many things are coming in, like what is the input shape, so this is the layer, I need to tell it about the input shape, I forget what that's called, somewhere in here I'll find, inputDim, which would be the input dimensions, I guess, which would be this.input_nodes, I'm forgetting my this dots, as always, I always forget. So there we go, what else do I need? Ah, I think an activation function I definitely need. So activation. Now this merits quite a bit of discussion. An activation function is how the data gets processed as it's fed forward. So all of these inputs, the numbers out of these values come in here, they get multiplied by those weight values, summed together and passed through an activation function. The activation function's job is basically to squash the values, to enforce non-linearity, to allow for the stuff to go to the next layer. And there are all sorts of reasons why you might pick one or the other, there's more recent research and change over the years, and I've talked about this in other videos. I'm going to pick sigmoid as a kind of default activation function that squashes all values between zero and one. Now I'll include some links about sigmoid in the video's description, if you're curious. So I forget exactly what I do here, but I think I can just write sigmoid as a string. And somewhere in the documentation there's a list of all the activation functions. Do I need anything else? I think that's good. So then I just want to say this.model.addlayer, maybe? So tf.Sequential is what I'm doing here, which would be here, and oh, input shape is probably what I want, not input dimensions, we'll see. I'm noticing that there's input shape, actually I think that's what I want. That's certainly what I used previously. So input shape, and it would be just that number of nodes, but inside an array. A shape is a way of defining the dimensionality of data. So if it's just a list of stuff, like a list of numbers, it's an array with five things in it. So that's how I define a shape. So then I would say, model.add so this.model.add the hidden layer. Now I want to make an output layer which is going to be tf.layers also dense, and I want to, what do I want? The units is this.output_nodes, and then, I actually do not have to define the input shape for this layer, because it can be inferred from the previous one. So I couldn't infer the number of inputs from the hidden, because hidden is first. But the last layer here, these outputs, I know that I just added this layer, if it's dense, this is the shape of the input. So all I need to do is that, and I also need an activation function, I'm going to make it softmax. So softmax is a activation function which writes sigmoid, squashes all the values to arrange between zero and one, but it does something even more. It basically enforces that all of those values also add up to one. It basically turns them into probability or confidence scores and all of them add up to 100% or 1.0. And there's more to it than that, it's a kind of interval mathematics involved, but that's the basic gist of it. And that's what I want, I want the probability of whether I should jump or not jump. So now, I should be able to say, this.model.add output and then if I go back to the documentation, I don't need to do this fit stuff, but do I need to compile it? So interestingly enough, usually we need to compile, we need to give a loss function. Traditional machine learning you're going to have a loss function, an optimization function, you're going to train it with data, I'm not doing any of that in neuroevolution. Remember, if you watched my other videos on neuroevolution, I'm doing this weird thing, I just need a random neural network, that's all I need. So I don't know if I need that compile step, I'm guessing I kind of do, so I'm going to say this.compile, but I'm just going to give it an empty object. 'Cause I don't care about those other things. Let's try this.model.compile So now that I have that, I should be able to, with my code, I'm kind of getting a little bit further. Like look at that. Cannot read property 'minimize' of undefined. Because I didn't give it any object. Let's get rid of the compile. Let's see if I get it to work without the step model.compile Maybe compile only has to do with the learning process that I'm not using. Great, this.brain.predict is not a function. So now I have the next thing that I need to implement. So where is that? That's in bird.js and what is it saying? So I'm able to make the neural network, and then I am calling predict and I'm giving it an array. Perfect. So what I need to do now in my class here is I'm going to write a function called predict, which will receive an array. Now what I would essentially be doing is now calling this.model.predict with that array. Only here's the complexity of using TensorFlow.js. TensorFlow doesn't work with traditional job script arrays. It works with something called tensors, which are multidimensional arrays, they are multidimensional things, collections of numbers, that are living in the GPU's memory. And that's for optimization purposes. The operations can happen faster on the GPU which is optimized for a lot of matrix map. I don't really need that optimization, to be clear, because I'm working with the tiniest thing here, but I still need to convert this stuff to tensor. So I'm going to need to say, the xs, I'm going to call this, the inputs are often referred to as xs, equals tf.tensor, and what do I have, what is the shape of my array? It's one-dimensional, right, it's just an array. But the thing is that what TensorFlow.js expects, so this is my data, it comes in as an array of numbers. One, two, three, four, five, right? All of this stuff, that's only four, whatever, you get the idea, there's another one there. The thing is TensorFlow.js thinks like, you would never be so crazy as to give me just one thing. Usually you're going to give me a thousand rows or data samples. So it's actually looking for a shape that is like this, that would have a bunch of these. So I just need to take my inputs and put them in a 2D tensor. Because it's an array of arrays. So if I do that, tensor2d xs. I think that's all I need to do. Then I need to call predict with the xs. Now that's happening synchronously, I need to get the data off of, oh, and this would be ys, sorry. The ys would be the result of feeding it forward. So that's exactly this process, predict is the function to feed this stuff forward and now I have ys. I'm going to do something that is a little bit ill-advised, I want to be able to look at those values. So right now the ys are a tensor and that's living on the GPU. In order for me to pull it off the GPU to use it, I need to call a function called data, which is getting the actual data and that's not going to happen asynchronously. So I need to deal with a promise or a callback, but because I'm working with such tiny bits of data here, it's going to be much simpler for me right now to get this working, to just say, basically, the outputs equals ys.dataSync So the dataSync function gives me those outputs. Let's console.log them, and let's say return outputs also, just so we see what's going on here. So I'm now going to run this, NeuralNetwork.predict at Bird.think is not defined. What did I miss? (bell dings) Sorry, there's a major error in here that you probably all noticed, but I didn't. I'm taking that array and converting it to tensor. Arr is the array, and maybe it would make sense to me to call these inputs. I'm taking the inputs as a raw array, converting them into a tensor, passing them through the model, pulling out the outputs and returning it. Okay, let's see how that goes. Tensor2d requires shape to be provided when values are flat. Oh, guess what, this has to be inside an array, right? Because the whole point is it wants the 2d thing, so it wants that array inside of that array, okay? So that should fix that. Great, oh look, I'm getting stuff. I'm getting things, oh, look how slow it is. It's so slow. So it's kind of ironic in a sense. Like I'm using a much more powerful job script engine for deep learning, that does all the operations on the GPU with WebGL to optimize it and yet it runs super slow. This is really because I'm a crazy person and kind of doing this in a weird way that TensorFlow.js is not really designed for. I have all these different models, and I'm copying the data with dataSync. So there's a couple of things I could do. Number one is, let's just for now, and I'll have to come back to this or think how to make this run faster or be more thoughtful or maybe do this stuff asynchronously. But what I'm going to right now, is I am going to just reduce the number of birds to half, and I'm also going to something a little bit silly, which I'm going to inset tf.setBackend cpu. Because actually the amount of data that I'm working with is so, so, so, so small, I don't really need to use the GPU, I don't need to copy stuff back and forth, let's just use the CPU, it's perfectly fast enough. Remember how well it worked when I had my own silly little JavaScript library? So let's use the CPU and I think you'll find that now it's still running kind of slow. I have a feeling this console.log is definitely making things kind of unhappy. So let's take out that console.log and let's refresh and there we go. So now it's performing pretty well. Is it as fast as it was before? I don't know. Can I get it back to 500? But it's better. Ah, but now we have a problem. Copy. So copy is not a function. Because I didn't write a function to copy a neural network. So what do I mean by copy a neural network? So normally this isn't something you would necessarily do, but it is fundamental to neuroevolution. And by copying it, we can say like, this was a good one, let's make a copy of it, to keep it. And I actually want to copy it, not keep a pointer to it. Because I could also want to mutate it, and I want to mutate it a bunch of times. So I have to figure out, what is the way to copy a model in TensorFlow.js. And ultimately if you want to try, the proper way to do this, in terms of neuroevolution, would not be to just make a copy to go to the next generation, but to take two of them, and make a new one that's kind of a mixture. When I say mixture, I mean mixture of all these weights. But I'm just going to copy it. So what I need to do, is I need to go back to this class again, and I need to say copy. So if I say copy, I could just say, return new NeuralNetwork, and I could just put like, inputs_nodes, hidden_nodes, output_nodes, but it wouldn't have the right weights. So let's think about this. Let's do, I'm going to say modelCopy equals this.CreateModel. So I'm creating a model, which is the same, ah, and I need to create model with, oh, it has that stuff. So I can call CreateModel, it's going to do the right inputs_nodes, hidden_nodes, output_nodes, okay. So it's not a copy yet, it's another random neural network. But really, the architecture's the same. So in a sense I've copied the architecture, I've made a new neural network with the same architecture, now I need to get those weights. So luckily for me, there is a ntf.js, there are two functions called getWeights and setWeights. So let's take a look at those functions. So for example, if I go to the console here, and I would say something like, can I say, let n equal new NeuralNetwork, what's it, 8.5.2, right? No, 5.2.8, 5.8.2. So I'm making one of those. And so n.model is that model itself. And you can see, it's got all that data in there. So what I want to do is, say n.model.getweights. No? Oh, I said mode. I want to say n.model, can we just click here, model.getWeights. These are all the weight matrices, why are there four? Shouldn't there just be two? Well, there's four because there are the weights and the biases, one, two, the weights and the biases, three, four. So that is what is here. So all I need to do is take these weights and copy them. So let's see. I could say, constant weights equals this.model.getWeights and then I could say modelCopy.setWeights Then I could say return new NeuralNetwork with modelCopy. Now there's a bit of a problem here that I think I'm going to have to do, but this is pretty decent start. Right, make a model of the same architecture, get the weights from the other one, and put those weights in the new model, and then return. I think there's going to be a problem here, but this is kind of the idea. So now, if I were to ... Ah, what I need to do here is, now I'm making a neural network without giving it the input nodes, I need to basically say, if a is an instance of tf.sequential, is that right? Let's see, so then if I go back to the console, n.model instanceof tf.sequential false, tf.sequential with a capital, true! Okay, I need to check if it's tf.sequential object, then this.model equals a. So, in other words, I have two ways of creating a neural network. If I pass the constructor an existing model, then I just assign it. Otherwise, I create a new one. However, I've really always got to keep track of what, it's a little, this is some redundancy here, 'cause the input nodes, hidden nodes, and output nodes are in the model itself. But let's just add another argument here. And then I can say, also, this is awkward, I could refactor this later. So now I'm also going to always require that I also give it the shape basically, and so here when I'm making this new neural network with model.Copy, I'm also going to say this dot, what is it, input_nodes, this.hidden_nodes, this.output_nodes. Okay, so this is me making that new neural network with modelCopy and I'm also just giving it the shape also so that it retains that information. So now, I should be able to go back, and we've got a copy function. Let's speed this up. Ah, can't read! Cannot read property 'setWeights' of undefined. That's not good. (bell dings) Okay, I've sort of made a mistake in the way that I've designed this program. What I want to say is this.model equals this.createModel and I don't actually want to set it here, I'm going to just make in the createModel function, which is where I'm going to make this a variable, const model, and then I'm going to say model.add, model.add, and then I'm going to say return model. So this function, this could be leading somewhere else, a static function or something, but it's making the model and it is in setup, sorry, in the constructor. It is creating it and returning it and placing it in the instance variable, or here it's creating it and returning it and putting it in these copy variables. So that should work better. So let's see. Great, so now I don't have a mutate function. No problem, no problem, this is the way to deal with that. Bird.js line 33, Bird.js line 33, let's just not bother at mutate. No worries, don't mutate. Let's speed this up. (upbeat electronic music) All right, let it run for a bit, we can sort of see, did it get better? Kind of, but there's only one, they're all making the exact same decisions. So two things are wrong here. Number one is, without mutation I can't get any variety. But there's something else that I'm pretty sure is wrong, not a 100% sure, but I'm pretty sure is wrong, so I want to protect myself against this, is I haven't really made a copy. So I think in the case of my copy function, which is here, this is the weights as tensors. And I am assigning them to the copy. But they are all being assigned the same tensors. Even though they are multiple models, they're all using the same weight table. So if I were to mutate them, they're all going to mutate in exactly the same way. So I need to really make sure I'm copying them. And there is a function in ts.js called clone, which is to make a new tensor that is a copy of an existing one. So I think what I should do is, I should say const weightCopies is a new array, and then for let i equals zero, i is less than weights index i, i plus plus No, sorry, weights.length, so I'm going to iterate over all the weights, and I'm going to say weightCopies index i equals weights index i dot clone. So this is kind of the idea, like deep copying, I guess, I'm not sure if I'm 100% accurate about that, but the idea of actually copying them into a new tensor, so it's not the previous one, and then I should be able to set the weights to weightCopies. So this should guarantee that I really got it copied correctly. So let's do that. I'll run it again, speed it up a little bit. So it's still converging, kind of, into it alternately being just one of them, that the one that happened to do the best. But it's not doing that because I haven't properly copied, it's doing that because there's no mutation. And you can see actually, that it seems to be taking a little longer to get there. So I think maybe that copying is better now. Okay, so I do need that mutate function. So what do I need? Where was mutate? So I need a function called mutate, and it's getting argument 0.1, which is presumably the mutation rate. So the process of mutation in a genetic algorithm is to say, look at the genetic information, which is basically all of these weights and biases, and one by one examine them, and at some probability make it sum adjusted, make it random, change it, mutate it. So that's what I need to do. So it's more involved than just making a copy of the weights. I need to look at every single weight individually and then mutate it 10% of the time. I'm not sure how to do this, but let's see, okay. So if I go into neural network and I go into, I made a copy function, now let's look at, sorry, let's look at mutate. So I want to do the same thing. The first thing that I want to do, is I want to get the weights. What I call this, this.model.getWeights. Then I want to maybe have mutatedWeights, is a new, just an array. So I get the current weight tables and then, weight tensors, sorry, and then I'm going to go through and make mutated weights. So let me once again loop over this array, and I'm going to say, tensor equals weights index i, 'cause it's a tensor, and then what I want to do, I want to get the values, so I'm going to use dataSync again. I want the shape, 'cause I'm going to need that when I convert it back, I think I can just do this. So I need the tensor, which is the data, I'm going to need to retain the shape, because I need to put it back into a tensor, but I want to mutate them. I don't think I can mutate the tensor without looking at the values. Who knows? So now I'm going to go through and I'm going to say, let j equals zero, j is less than values.length, j plus plus. So the actual weight which I call w is values index j. And what I want to do is say values index j equals a new random weight. So I could do something like this, right, just give it, and I don't even actually need the current weight, I could just make a random one, right? I'm mutating, oh, but only 10% of the time. So this gets a rate, so I could say, if random one is less than that rate, then make a new weight. I do think this is a case where there might be some benefit to rather than picking a totally new weight, and I don't know what initialization was used to make the initial random weights, that's probably somewhere in TensorFlow.js, so there might be some operation I can call here, but I'm just going to tweak it. So this is why I want to grab that weight, and I'm going to say, weight plus, and I'm going to use P5's random Gaussian function which basically gives me a random number that's kind of near, with a mean of zero and a standard deviation of one, so it's just going to adjust that weight a tiny bit. So it's like, kind of, what you might see with graded descent, dial up, dial down, but I'm just making a total guess, just dial it, some direction. So it's going to equal that, this loop is looking at individual set of weights, it might be between inputs and hidden or hidden and outputs or might be one of the biases, either of those weights. And I have it as a tensor. And then I have its values, now I mutated those values. Aha, I need to make a new tensor, newTensor equals tf.tensor, and give it those values again, right, the mutated values, with the shape. So this is what's nice. I don't have to worry about what the original shape was, dataSync is just going to flatten everything and give me all the numbers, pretty sure, I think. And then, what I can do, is then mutate those, it's flat, and I can put it back into a tensor with the same shape and it should be back just like it was before. I can say mutatedWeights index i, is now that new tensor. So I grabbed the tensor, I got its shape, I got the values, I mutated them, I made a new tensor, and it's going to go in my new array, of mutated weights. And then when that's done, I can say, this.model.setWeights mutatedWeights. I think this is mutation. There is one mistake here, which is that tensor, sorry, which is that dataSync is actually not making a copy of the values. So I'm still going to be stuck with a lot of things pointing to the same exact data, so if I mutate one and it's copied for a bunch, all of those are going to get mutated in the same way, and I want to have a bunch of different mutations happening in parallel. I discovered this issue around dataSync in this discussion, and thank you so much to one of the creators of TensorFlow.js, Nikhil Thorat, who answered my question about this, so dataSync is mutating, I'm mutating the underlying values in the tensor. Apparently there is an arraySync function that I could use instead of dataSync, that doesn't actually copy, but me, according to a Coding Train viewer actually suggested that I just call .slice, which is a job script function for copying the array. So this I really really need in there. Because I really have to make sure that I'm not mutating the same underlying information. So I should add .slice here. And then I think I have this done, we're about to find out? The only way to find out, there are other ways to find out, I'm sure, is to let this run for a while and sort of see if it gets anywhere. So I'll let that do that for a little while. (upbeat electronic music) So I would say that this works. Because, there we go. I have an agent that learned basically, again this is a very simple problem for it to solve, but this technique that I've now done, there's no reason why it couldn't be applied to a more complex scenario. So I now have neuroevolution with TensorFlow.js, but I'm not finished. I could do this in the next part, but there's a huge problem here. I haven't thought at all about memory management. And if I do tf.memory in the console, ah, I have 128,870 tensors and I'm using all these bytes, oops sorry, if I start from the beginning I have that many tensors, and look how it's going up. So I'm leaking memory like crazy. I'm not cleaning up any of the tensors that I'm using. And so the way to do that, I don't know why I'm coming over here, but it's very important. In addition to the getWeights and setWeights functions, there are two functions that I really need to be conscientious about when working with TensorFlow.js. One is tf.tidy, and this is a function that you can basically give it a callback, and any code that you put up in there, it will do the cleanup for you. So it's kind of like TensorFlow.js's garbage collection feature. It's like just put the coding here and don't worry, we'll take care of disposing memory that's not used. So this is probably one I want to use most of the time, but any tensor I could also individually call dispose on, when I'm not using it anymore. So let me go and find everywhere in my neural network code that I'm using tensors, and make sure I'm cleaning up memory that's not being used anymore. So I'm in my code, scrolling, scrolling, scrolling, this should be fine, I don't think I need to dispose models, maybe I do, but here, okay, oh, look at this mutation. So I'm creating new tensors, I'm doing stuff, this is definitely a place where I want to say tf.tidy, and the nice thing is it's literally as simple as this, I just put everything inside of tf.tidy. So maybe there's a slightly more optimal, better way of doing this, but this should work. In predict, this one's a little easier, I could say like, I don't need the xs anymore, so dispose of those. And now once I have the outputs, I can dispose of the ys. So I can put manual dispose in here, like that's me disposing, but I probably will also just use tf.tidy, same exact thing. But there's a little something else I need to do here. Because this code is returning something, I need to return the result of tf.tidy. So that should be fine. I feel like the model stuff is fine, this should be all the cleanup that I need. But let's see. Oh, copy, oh, yeah. I guess let's do tf.tidy in here also, tf.tidy, let's do this. I also need to say return, since I'm returning a new model, I mean I could put tf.tidy in this, let's see if I need to. Let's look at the memory now. So that's a 1000 tensors, that makes sense, because I have 250 birds and there's four tensors, four weight matrices, that's a thousand. So 3000, oh, I missed something. So every generation, I'm keeping some of the old tensors that I don't need anymore. So let's try putting, I suppose, tf.tidy here, and also return, and let's see what happens. Ah wait, something undefined. What did I mess up? (bell dings) That didn't seem to work. But I have a better idea, I don't know why that didn't work, there's probably a very good reason, and I will hear from somebody about it, but I think I know what I need to do. Where am I copying? Here, but I have tf.tidy here. I was thinking that once I make this new neural network I would call dispose on the existing model. (bell dings) Okay, after examining this for a bit, with a little bit of help from the chat, it looks like the issue that I have is that, it's a little bit weird, what I'm doing here, I have a bunch of different arrays, I have birds, saved and savedBirds. So savedBirds is kind of like the previous generation, and I make all these new birds for the next generation, put them in the birds array. But the models, the savedBirds models still exist. So I need to dispose those. So the way I could do that is find the place where I'm done making the next generation, and iterate over the savedBirds. The way I could do that is, I could also, in the bird object right now, I could add a function called dispose, little silly to do this, but why not, which just says this.brain.dispose and then in the neural network class, I now have a function called dispose, which I would say this.model.dispose. So this is sort of a manual disposal of all of the memory that that particular model is using. And then somewhere in the sketch, when I call next generation, that makes the next generation and then I go on, that's in my genetic algorithm, which is here. So this is making the next generation, and then clearing the savedBirds. So before I clear the savedBirds, I think it's, I could just use this total value, what I can do now is, I can say savedBirds index i dot dispose. So that should call the dispose function which calls the dispose function which calls the dispose function. So now, I'm looking at a 1000 tensors, and then let me speed this up, so it gets to the next generation faster. Wow, it just got like a good one. Next generation and now, still a 1000 tensors. (whistles) So this is really complete. I mean there's so much more that I could do, but this is complete in terms of the idea of neuroevolution with TensorFlow.js and this particular Flappy Bird game. All right, so what's coming next, what could you do, what could I do next? So one thing is, in my previous coding challenge I saved the model to reload later, and TensorFlow.js has functions for doing that. Save and load layers, models, I forgot what those functions are called, but they're in there. So I could come back and add that to this. The other thing would be to apply this idea to different scenarios, different games, to different environments. One of the things that I would like to do is work with steering agents, whether they're finding food or avoiding predators, that would be an interesting thing to try. So hopefully I'll come back and do this same idea, but with a different environment that has maybe more complexity to it, with a less obvious result. Because I have a feeling that I could have gotten a bird to solve this game with just some if statements. Like I could probably guess like, you should always jump, if you're too high, you shouldn't jump if you're higher than where the pipe opening is, and if you're lower, you should jump. So I could write an if statement instead of using neural network. Thank you for watching. If you make your own version of this, if you have ideas for how to make this better, please let me know in the comments and also over at thecodingtrain.com, you can submit a link to your version of this running in the browser, and I'll take a look at those and hopefully share them in a future video or a livestream. Thanks for watching and goodbye. (whistles) (upbeat electronic music)
B1 中級 11.3 使用TensorFlow.js的神經進化Flappy Bird (11.3 Neuroevolution Flappy Bird with TensorFlow.js) 2 0 林宜悉 發佈於 2021 年 01 月 14 日 更多分享 分享 收藏 回報 影片單字