video_id
stringlengths 11
11
| text
stringlengths 361
490
| start_second
int64 0
11.3k
| end_second
int64 18
11.3k
| url
stringlengths 48
52
| title
stringlengths 0
100
| thumbnail
stringlengths 0
52
|
---|---|---|---|---|---|---|
Cs_j-oNwGgg | right so it will usually do this around the data point so the data points form the valleys right here and then the generator will basically take that discriminator function and will try to infer points that are also in these valleys to produce points that are also in the valleys and then you basically have an energy learning competition the discriminator now tries to push down on | 318 | 346 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=318s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | the energy where the true data is and push up on the energy where the generated data is and that will give you basically a steeper energy based function in the future I hope so in this case the discriminator neural network is the energy function and the D generator just tries to produce data that is compatible with that energy function so I hope that concept of what an energy | 346 | 376 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=346s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | function is is a bit clearer and in any again any machine learning problem can be formulated in terms of an energy function now what is not done so far is what we alluded to a little bit before in the classifier example and also here so right now when we want to Train again we simply take the generator to whose data now what's the generator skull the generators goal is to hit | 376 | 404 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=376s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | those valleys in the energy function and we produce a generator into in one shot produce this data but what we could also do is of course we could just start somewhere let's say here we pick a random data point and then we use gradient descent because the energy function in this case is smooth we use gradient descent to just drop down this valley and then find ourselves in this | 404 | 430 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=404s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | valley so without ever training a generator we can use this methods to produce points that are in the valley of the energy function right and this I don't know if people I can guess people have trained gams like this the reason why it doesn't work let's say in the real world is because that procedure will just produce adversarial examples for the discriminator and those usually look | 430 | 456 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=430s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | like nothing like data because if you keep the discriminator just stable and gradient descent against it what you'll get isn't really qualitatively good but in principle if the discriminator was a good energy function for the date to describe the data we could use gradient descent the same up here in order to find a good label for an image given that we have a good energy function | 456 | 484 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=456s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | right so this is that we could simply gradient descent on the label in order to find a better in order to find a better label so in this paper we're going to have a situation where we say we're given an energy function and we're given a bunch of inputs they are then called X a and W and if I have my energy function already if I have given my energy function and I have given two of | 484 | 519 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=484s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | those three things any two right I can infer the last thing simply by gradient descent on my energy function because I know the energy function is zero when all of these when the energy function is happy with the input so when all of these things agree basically the energy function is happy it will output zero otherwise it will output a high value therefore if I given any of those | 519 | 548 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=519s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | two any two of those three things I can find a compatible third thing by descending and then of course over here in these machine learning problems the task was always actually to learn an energy function right so usually in the training dates that we are given images and labels and we want to learn this energy function which would be parameterized so we want to learn the | 548 | 575 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=548s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | parameters and the same here in our general case if we are now given three things but we are not given the parameters of the energy function we don't know what those are as long as we're given all of the inputs and our training data set and our training data set guarantees these are actually you know these are inputs that are compatible with each other the energy function should below we can | 575 | 599 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=575s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | simply gradient descent on the parameters of the energy function so in a sense there are four things right there are these three inputs and then there are the parameters of the energy function if we're given any three of those four we can gradient descent on the rest and that's going to be the basis so the X here is going to be the so-called state and the state in this | 599 | 626 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=599s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | paper is going to be images of entities the entities sorry is not going to be images but the entities are these little circles that you're going to see and each of those entities can have an exposition a Y position and I believe a color so our G and B so each of those can have that and then the concatenation of all of those attributes is a one big vector and that is your X that's your | 626 | 657 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=626s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | state so state is number of entities and their attributes a is going to be an attention mask over the state so a is going to be here you have four entities so a will have four entries telling you which of these entities you should pay attention to right now and W is going to be a concept vector so called so W is going to be the embedding of a concept now what a concept is in this case is very | 657 | 697 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=657s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | general I can give you an example one a concept is do any of do the entities that the a pays attention to are they close to each other so in this case you see we have two entities that a has a high value one and this is this all up here and this ball down here now if the concept vector is the embedding for the concept of being close to each other then the energy function would be very | 697 | 733 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=697s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | happy if those two things are close to each other and it would be very unhappy if those two things aren't close to each other but in the very same situations of the same X the same attention mask but a different concept so a different W vector right here then the the energy function would be maybe very happy if the two things are far apart and maybe unhappy if the two things are close so | 733 | 759 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=733s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | the question is always how are the three things that you put into the energy function compatible with each other and given all but one of these things you can infer the other so let's say you have a perfect energy function for this this all of these for this situation you're just given the energy function you can trust it and you are given let's make an example you are given the X so | 759 | 790 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=759s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | you're given the state I'm gonna draw the state down here right okay this is the state and you were given the W and the W is the embedding it's a vector but the embedding space but the embedding is for a line right so the geometric the geometric unit of a line now your task is to find a the attention mask that will make the energy function happy and as you can see right here | 790 | 827 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=790s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | while you would do is you would put a lot of weight on this this this and this ball and no weight on that ball because those make a line and since everything here is differentiable so the state is differentiable the attentions differentiable and the concepts or vectors that are differentiable you can use gradient descent to find that another example if you're given again | 827 | 852 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=827s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | the same W so line and you were given this following thing and you were given now you're given the attention on these three and you say please find the X please find the X the states that makes this energy function happy now this here you would call the starting state the x0 your ear task is going to be find the x1 find the state how do you have to change this state such that the energy function | 852 | 887 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=852s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | is happy and of course the answer is going to be is to push this ball here inward until it is in the middle of the two others so the three form a line right these three form a line you don't you don't have to do anything to this ball up here because there is no attention on it and the attention it's only is the concept for the things that you put attention on yeah and the state | 887 | 912 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=887s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | are those three in agreement and the energy function is happy okay we have covered the basics now let's dive into the paper I think this is the longest introduction ever but I think it will pay off once see so they they specifically or this this author I think it's a single author identifies two different things that you can do with an energy function here of | 912 | 943 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=912s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | course you can do more as we saw but they identify two so here is where you have given the initial state and an attention mask and you want to find the x1 the state that satisfies the concept and the tension the most this the author calls generation as you can see here these four things that you have the attention on are pushed around until they make a square because the concept | 943 | 971 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=943s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | right now is square and in the other case where you are given this x0 and x1 just call this X right here just call this thing X if you're given those two and you are given the concept Square and you're tasked with finding a the attention mask of course you're going to put the attention on these right here and that is going to happen through gradient descent again we're not | 971 | 1,002 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=971s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | learning a model to give you that attention like in again we're learning a generator to just one shot give it to you right now what we're going to do is we're going to gradient descent optimize on our smooth energy function to give us that perfect attention mask that satisfies the energy function all right so this is the difference right here gradient descent is part of the output | 1,002 | 1,025 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1002s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | procedure of the model usually we just use it to learn and we learn a one-shot model but here gradient descent is part of the model so they introduce energy functions here and they say okay we can have a policy on X so if we're given a concept W and if we're given an A we can have a policy over X which basically means we can find X's that are compatible with that by running gradient | 1,025 | 1,055 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1025s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | descent here you see there is an XK minus 1 and we are running gradient descent on the energy function with respect to X to find a better X that satisfies the energy function given those inputs and the same if we want to find an attention mask we are running gradient descent on the attention mask again in order to satisfy the same energy function so you see the inputs | 1,055 | 1,088 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1055s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | are both times the same the concept here we can input square here we can input square but the difference is what we're running gradient descent on and what we keep constant and I would get I would add a third line here actually because we can also if we're given an X and an A we can also infer a W and that's going to be an integral part so if I have this right here and this situation and I have | 1,088 | 1,122 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1088s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | say I have a tension on these for now I can ask the model so I'm given X and I'm given a I can ask the model to infer W and the model should ideally output ha this is square now the model isn't going to output square the model is going to output a vector representation of square right so the model is going to output square but as a vector of numbers because that's how we've trained it W | 1,122 | 1,155 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1122s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | isn't it is the embedding but what we can then do later is we can say okay I'm not going to tell you it's a square you just come up with a vector W it describes this situation and now I'm going to take that vector W that you came up with miss mister or missus model and I'm going to take tell you a new situation this situation right here and I'm going to now give you X and I'm | 1,155 | 1,186 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1155s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | going to give you the W that you yourself have output and now please tell me what's the a and then the model is of course supposed to tell you oh these four here or the a so without without ever telling that it should be a square what you can do is you can let the model infer a W from one example situation and then transfer that W to a new situation so it can identify you can just say | 1,186 | 1,216 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1186s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | whatever concept I have up here please apply that same concept which is the W down here and this is the entire paper now this is the concept learning through energy based models okay so that is kind of a third line I would add down here you can infer a concept vector if you're given the X and the a so in order to do all this their energy function is going to be a so called relational neural | 1,216 | 1,247 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1216s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | network so what you'll have is you'll have a simple neural network a multi-layer perceptron that always connects two entities to each other with the concept vector and then this is a belief a sigmoid that connects the attention masks of the two and then you simply sum over all pairs of two entries in your model and then you send that through an MLP sorry through an MLP | 1,247 | 1,275 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1247s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | again this I believe is not so important it's just important that they can feed this entire situation the X the a and the W they can basically feed into a neural network in the neural network comes up with a number of how well those three things fit together and then you can transfer these concepts that's pretty cool now the only question is of course and we've always said we're given | 1,275 | 1,302 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1275s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | an energy function or just we just have it but of course this is a neural network and the neural network has parameters and the parameters we don't know what good parameters are at the beginning so we need to train this thing and again the reason why these are toy problems right here is I mean we'll get to why it's computational but this is kind of a new field | 1,302 | 1,327 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1302s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | I believe in machine learning at least I come from classical machine learning and we only ever have used like SGD to train and we only have her have produced models that one shot produce something and here we this is a I believe there's a new concept where you use gradient descent as part of the output and that makes a lot of trouble and so that's why we work | 1,327 | 1,353 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1327s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | in toy problems so what this this here is the situation I described you have a demo event where you're given the X and the a and you're supposed to infer the W so the question here is what's the W and the model will come up with a W and you're not gonna do anything you're not right now you're simply gonna take that W and tell it oh well here is a so called test event so please apply the W you | 1,353 | 1,385 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1353s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | came up with in this test event and please find me the the a in this case that satisfies the W and the X I give you here and of course the a right here is as you can see even you don't know that it's a square and the actual concept here is move the grey ball to the middle of the square right that that is it here but no one has told me this I just looked at the picture so the the | 1,385 | 1,413 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1385s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | correct answer here would be to place attention on those four things and then to take this thing and move it to the middle right here in in the in this over here so that would be the correct answer now the question is how do you train something like this and they they show that they so this is the loss function right here the loss function is they give you a concept and an initial | 1,413 | 1,445 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1413s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | situation and you're supposed to infer the x1 and the a and the loss function is simply the negative log likelihood of that but what does that mean so will will make it easier if if you have this this procedure right here where you have demo event this up here this is demo and this is a test event how are you going this entire procedure how are you going to learn the energy function well in | 1,445 | 1,478 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1445s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | this case this entire procedure this entire thing is one training sample sample but usually we have input and label and now here it's much more complicated because so we have input okay that's this X and this a cool but then we have SGD as integral part of the procedure to determine the W and now what we could do is just apply a loss to the W but we don't because we don't know | 1,478 | 1,512 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1478s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | what the embedding space for the concepts is we could maybe train a classifier but in this case we want to train the ability to transfer these concepts so our training sample needs to be one time transferring a concept so SGD for one is part of our process here and not only that but then this this X here of course is also part of our training sample write this up here as X | 1,512 | 1,540 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1512s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | 0 and this here is X 1 and now we need to find this a this attention mask and that is an SGD again remember inferring anything through the energy function is a gradient descent process so ultimately our one training example consists of X 0 a at the beginning so let's call that a zero it consists of the SGD procedure to find W it consists of X 1 and they consist of the SGD procedure to find a | 1,540 | 1,575 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1540s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | the a 1 the output a and then that will give us the output a the a 1 so this here is our input in the classical machine and this would be our X and this here would be our label Y and that's what we trained on we trained so it such that the output right here the this is of course sorry this is of course the Y hat this is what we predict and in the training sample we just write | 1,575 | 1,608 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1575s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | a little generator that will you know make this situation that knows what the concept is right it will say okay I'm gonna make an example for a square then it make this will make the attention mask for a square and then it will make the new situation again with a square but not tell us the attention mask there and it will make the attention mask into the true Y so at the end we can compare | 1,608 | 1,638 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1608s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | what our model output the attention mask we output here without ever knowing that it should be a square and we have the true label which comes out of the generator that at the beginning decided that it should be a square and then the loss in the distance between those two that's our loss this is an in this is an enormous procedure to get a loss and most crucially you have to back | 1,638 | 1,669 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1638s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | propagate through optimization procedures and this is something that we just can't do yet in our models if you take an image a resonate 50 right right now we do one forward propagation to get a label in this procedure if you had two back propagate through the optimization procedure for each sample you would need to basically back propagate through 50 forward passes of the resonate if you if | 1,669 | 1,697 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1669s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | your optimization procedure is 50 steps long and that is just not feasible right now so that's why we don't do it but I believe maybe once we find a smart way of back propping through optimization procedures a whole lot of these things will become the new and new wave and machine learning I really I'm excited - I'm pretty sure it doesn't work yet and this is very figley fiddly work but I'm | 1,697 | 1,727 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1697s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | excited by the prospect that we can do this so this is the training procedure right you've given X 0 x1 and a and you Optima is in order to infer the concept behind it right the generator that your level generator of your training data it knows the concept it has a concept in mind when it generated this but you're not telling your model what the concept is it needs | 1,727 | 1,753 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1727s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | to infer that and then using the model thing that the model inferred you can either give it x0 and x1 and infer a or you can give it the X and the a and in forex you can do either of those right these are called identification or generation respectively and then you compare the output here to what the generator at the beginning thought again it's not telling you it's that's because | 1,753 | 1,781 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1753s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | that's the label and you compare this to that and that will be your loss to train your energy function parameters so your training samples if you think of this entire thing as one forward pass of the model then it's just classic machine learning right you have a training sample which is one forward pass and you have a corresponding label that you infirm so let's jump to the experiments | 1,781 | 1,807 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1781s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | right here experiments are actually pretty cool so what they've done is for example have taken the concept of being far apart from something now being far apart so that the little X needs to be as far away as possible from the ball that has the attention on it so if you do generation and you start the little X right here and you ask the model where please infer the next state of the world | 1,807 | 1,841 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1807s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | it will push that little X away right here and in color you can see the energy function valleys of the position of the X so it pushes it away from this thing but if you take the same concept embedding the concept embedding of being far away but you don't do generation you do identification which means you infer the a then it will simply tell you that this ball right here is the furthest | 1,841 | 1,871 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1841s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | away from the X right so you can do all sorts of things like this and transferring concepts I find this here pretty interesting so they had to have two different concepts one concept is read as an identification you need to identify the red ball but the other concept is you need to turn something red right you need to take a ball that is maybe now blue and of course the | 1,871 | 1,901 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1871s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | color you can gradient descent on the colors you'd need to make it red and since the energy function it just takes three input X a and W it doesn't you you you're not going to tell it right now in which situation you are it has to create create this W embedding space through learning and if you do it with those two concepts then it will put the make something red concept and the | 1,901 | 1,933 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1901s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | is something red concepts in the same places so this is a PCA and then blue I think these blue is the attention codes for identify the red things and in red or the generation code for make something red and they will be put in the same place which is pretty cool it means that the energy function really learns the feature of something being red I find this pretty pretty neat and then | 1,933 | 1,961 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1933s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | here they they have some experiments where they basically show we need that gradient descent optimization procedure because only after many steps will will the energy function basically be aligned with the concept that you want so if you have a zero shot model like just one forward pass as we do here you'll see that the energy function that is supposed to make a circle from samples | 1,961 | 1,988 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1961s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | right this is the example concept right here it if you just have a one shot model it will it cannot but in this case at least it doesn't learn to one shot produce only if you opt in for a few steps will it get this so you optimize at inference time and that seems to be very important you can see again here demonstrations of this so the example is this and then the model as | 1,988 | 2,017 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=1988s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | you can see after 20 steps learn optimizes the points to go to these locations whereas after only one step it didn't do that yet so there are complex things at work here and this column here is where you don't have a relational or neural network so you can't basically capture dependencies between things so you you have no chance of making a square because you don't know where the | 2,017 | 2,042 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2017s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | things are in relation to each other but that's more of an engineering question their point is basically that if you have models that do an optimization at inference time they are much more powerful than models that just do a one-shot forward pass it's sort of like an auto regressive model in NLP versus a non auto regressive model that produces all words at once if you produce all | 2,042 | 2,069 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2042s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | words of a sentence at once no word can depend on any other word and you can just come loose independent or you can just produce independent things which will make the sentence often not make any sense they also have this KL objective which is a regularizer which I believe that's just a trial and error they built it in because but it is a regularizer I don't want to really go into that and | 2,069 | 2,097 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2069s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | then they they do demonstration in and they re-enacted on a robot the demonstration here is that there is a situation where two things have a tension on and you're supposed to move something into the middle of the two things so that's the content you don't tell the robot the concept it needs to learn that from data and then infer that this is the concept that you want and | 2,097 | 2,120 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2097s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | then transfer that to the other environment now you know this you look you know there's this robot environment but ultimately they still encode the positions of these things and the position of that and really all you have to do different here is that instead of moving this actuator directly you need to like calculate what you need to do to the individual joints in the robot so I | 2,120 | 2,150 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2120s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | think this is maybe because it's open AI and it needs to you know look robot II and stuff but the problem here is not really different it's it's not even it's not real-world transfer or anything so yeah let's let go through some of the things they can learn with this so you can see here they can learn these regional geometric shapes and so on the left is the example | 2,150 | 2,175 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2150s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | event that the model needs to take the concept from now this is this is I believe very much identification so what they did is they trained with a data set where all of these appear right so this there are squares there are lines there are circles so this is maybe my criticism here that it is not so much to generally infer a concept it is more like identify the concept so the model | 2,175 | 2,203 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2175s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | basically just needs to decide is this line is the circle or is this square because that was those things were in the training data set it would be nice to see how this generalizes to general concepts or if we can even make that if we can have a zero shot concept inference and then transfer those concepts to other things maybe that's already happening I don't I don't know so here the spatial | 2,203 | 2,230 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2203s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | arrangement is to either be close to something or to be between two things so if the attention is on two things you want in between so you see the top ones are the demonstrations it needs to recognize the concept and it needs to basically optimize to fulfill that concept shapes so to make shapes is mmm oh yeah there's a triangle right again this this this just very much I believe | 2,230 | 2,266 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2230s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | relies on recognition and not actual understanding of what a triangle here you have proximity being closer being far apart what else is cool oh yeah if the recognition for the same task right you need to identify the ball that is closer for and here you really also see the optimization procedure in action where for example at the beginning of each flicker you kind of | 2,266 | 2,293 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2266s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | see the attention being everywhere and then stabilizing to one or two points so if two points are equally close or far apart you'll see the attention being on multiple points which is pretty cool right so that means the model really learns this this is concept here's the count quantity so you can either have one two or larger than three or something yeah that seems like they | 2,293 | 2,322 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2293s | Concept Learning with Energy-Based Models (Paper Explained) | |
Cs_j-oNwGgg | tried three and four and didn't work so they just said I will just do larger than three and here is this robot thing where it also always needs to move in between now this this is the part that I'm not really impressed with but you know whatever whatever you want okay I hope this was a good introduction to energy functions what you can do with them what I think of them end of this | 2,322 | 2,345 | https://www.youtube.com/watch?v=Cs_j-oNwGgg&t=2322s | Concept Learning with Energy-Based Models (Paper Explained) | |
IRVdiHu1VCc | The "Dirty Jobs" crew and I were called to a little town in Colorado, called Craig. It's only a couple dozen square miles. It's in the Rockies. And the job in question was sheep rancher. My role on the show, for those of you who haven't seen it -- it's pretty simple. I'm an apprentice, and I work with the people who do the jobs in question. And my responsibilities are to simply try and keep up, | 0 | 38 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=0s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and give an honest account of what it's like to be these people for one day in their life. The job in question: herding sheep. Great. We go to Craig and we check into a hotel, and I realize the next day that castration is going to be an absolute part of this work. Normally, I never do any research at all. But this is a touchy subject, and I work for the Discovery Channel, | 38 | 62 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=38s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and we want to portray accurately whatever it is we do. And we certainly want to do it with a lot of respect for the animals. So I call the Humane Society and I say, "Look, I'm going to be castrating some lambs. Can you tell me the deal?" And they're like, "Yeah, it's pretty straightforward." They use a band, basically, a rubber band, like this, only a little smaller. | 62 | 86 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=62s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | This one was actually around the playing cards I got yesterday -- (Laughter) But it had a certain familiarity to it. And I said, "Well, what exactly is the process?" And they said, "The band is applied to the tail, tightly. And then another band is applied to the scrotum, tightly. Blood flow is slowly retarded; a week later the parts in question fall off. "Great -- got it." | 86 | 108 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=86s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | OK, I call the SPCA to confirm this. They confirm it. I also call PETA just for fun, and they don't like it, but they confirm it. OK, that's basically how you do it. So the next day I go out. And I'm given a horse and we go get the lambs and we take them to a pen that we built, and we go about the business of animal husbandry. Melanie is the wife of Albert. Albert is the shepherd in question. | 108 | 130 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=108s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | Melanie picks up the lamb, one hand on both legs on the right, likewise on the left. Lamb goes on the post, she opens it up. Alright. Great. Albert goes in, I follow Albert, the crew is around. I always watch the process done the first time before I try it. Being an apprentice, you know, you do that. Albert reaches in his pocket to pull out, you know, this black rubber band, | 130 | 151 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=130s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | but what comes out instead is a knife. And I'm like, "Hmm, that's not rubber at all," you know? (Laughter) And he kind of flicked it open in a way that caught the sun that was just coming over the Rockies, it was very -- (Laughter) It was ... it was impressive. In the space of about two seconds, Albert had the knife between the cartilage of the tail, right next to the butt of the lamb, | 151 | 176 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=151s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and very quickly, the tail was gone and in the bucket that I was holding. A second later, with a big thumb and a well-calloused forefinger, he had the scrotum firmly in his grasp. And he pulled it toward him, like so, and he took the knife and he put it on the tip. "Now, you think you know what's coming, Michael, You don't, OK?" (Laughter) He snips it, throws the tip over his shoulder, | 176 | 198 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=176s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and then grabs the scrotum and pushes it upward, and then his head dips down, obscuring my view. But what I hear is a slurping sound, and a noise that sounds like Velcro being yanked off a sticky wall, and I am not even kidding. Can we roll the video? No, I'm kidding, we don't -- (Laughter) I thought it best to talk in pictures. I do something now I've never, ever done on a "Dirty Jobs" shoot, ever. | 198 | 225 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=198s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | I say, "Time out. Stop." You guys know the show, we use take one; we don't do take two. There's no writing, there's no scripting, there's no nonsense. We don't fool around, we don't rehearse -- we shoot what we get! I said, "Stop. This is nuts." I mean -- (Laughter) "This is crazy. We can't do this." And Albert's like, "What?" And I'm like, "I don't know what just happened, | 225 | 248 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=225s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | but there are testicles in this bucket, and that's not how we do it." He said "Well, that's how we do it." I said, "Why would you do it this way?" And before I even let him explain, I said, "I want to do it the right way, with the rubber bands." And he says, "Like the Humane Society?" I said, "Yes, like the Humane Society. Let's do something that doesn't make the lamb squeal and bleed. | 248 | 268 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=248s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | We're on in five continents, dude! We're on twice a day on the Discovery -- we can't do this." He says, "OK." He goes to his box and pulls out a bag of these little rubber bands. Melanie picks up another lamb, puts it on the post, band goes on the tail, band goes on the scrotum. Lamb goes on the ground, lamb takes two steps, falls down, gets up, shakes a little, | 268 | 288 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=268s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | takes another couple steps, falls down. I'm like, this is not a good sign for this lamb, at all. Gets up, walks to the corner. It's quivering, and it lies down and it's in obvious distress. And I'm looking at the lamb and I say, "Albert, how long? When does he get up?" He's like, "A day?" I said, "A day! How long does it take them to fall off?" "A week." Meanwhile, the lamb that he had just done his little procedure on | 288 | 316 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=288s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | is, you know, he's just prancing around, bleeding stopped. He's, you know, nibbling on some grass, frolicking. And I was just so blown away at how completely wrong I was, in that second. And I was reminded how utterly wrong I am, so much of the time. (Laughter) And I was especially reminded of what a ridiculously short straw I had that day, because now I had to do what Albert had just done, | 316 | 343 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=316s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and there are like 100 of these lambs in the pen. And suddenly, this whole thing's starting to feel like a German porno, and I'm like -- (Laughter) Melanie picks up the lamb, puts it on the post, opens it up. Albert hands me the knife. I go in, tail comes off. I go in, I grab the scrotum, tip comes off. Albert instructs, "Push it way up there." I do. "Push it further." | 343 | 369 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=343s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | I do. The testicles emerge. They look like thumbs, coming right at you. And he says, "Bite 'em. Just bite 'em off." (Laughter) And I heard him, I heard all the words -- (Laughter) Like, how did I get here? How did -- I mean -- how did I get here? It's just -- it's one of those moments where the brain goes off on its own, and suddenly, I'm standing there in the Rockies, | 369 | 399 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=369s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and all I can think of is the Aristotelian definition of a tragedy. You know, Aristotle says a tragedy is that moment when the hero comes face to face with his true identity. (Laughter) And I'm like, "What is this jacked-up metaphor? I don't like what I'm thinking right now." And I can't get this thought out of my head, and I can't get that vision out of my sight, | 399 | 421 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=399s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | so I did what I had to do. I went in and I took them. I took them like this, and I yanked my head back. And I'm standing there with two testicles on my chin. (Laughter) And now I can't get -- I can't shake the metaphor. I'm still in "Poetics," in Aristotle, and I'm thinking -- out of nowhere, two terms come crashing into my head, that I hadn't heard since my classics professor in college drilled them there. | 421 | 449 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=421s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | And they are "anagnorisis" and "peripeteia." Anagnorisis and peripeteia. Anagnorisis is the Greek word for discovery. Literally, the transition from ignorance to knowledge is anagnorisis. It's what our network does; it's what "Dirty Jobs" is. And I'm up to my neck in anagnorises every single day. Great. The other word, peripeteia, that's the moment in the great tragedies -- | 449 | 480 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=449s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | Euripides and Sophocles. That's the moment where Oedipus has his moment, where he suddenly realizes that hot chick he's been sleeping with and having babies with is his mother. That's peripety, or peripeteia. And this metaphor in my head -- I've got anagnorisis and peripeteia on my chin -- (Laughter) I've got to tell you, it's such a great device, though. When you start to look for peripeteia, | 480 | 508 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=480s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | you find it everywhere. I mean, Bruce Willis in "The Sixth Sense," right? Spends the whole movie trying to help the little kid who sees dead people, and then -- boom! -- "Oh, I'm dead." Peripeteia. You know? It's crushing when the audience sees it the right way. Neo in "The Matrix," you know? "Oh, I'm living in a computer program. That's weird." These discoveries that lead to sudden realizations. | 508 | 532 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=508s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | And I've been having them, over 200 dirty jobs, I have them all the time, but that one -- that one drilled something home in a way that I just wasn't prepared for. And, as I stood there, looking at the happy lamb that I had just defiled -- but it looked OK; looking at that poor other little thing that I'd done it the right way on, and I just was struck by -- | 532 | 557 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=532s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | if I'm wrong about that, and if I'm wrong so often, in a literal way, what other peripatetic misconceptions might I be able to comment upon? Because, look -- I'm not a social anthropologist, but I have a friend who is. And I talk to him. (Laughter) And he says, "Hey Mike, look. I don't know if your brain is interested in this sort of thing or not, but do you realize you've shot in every state? | 557 | 584 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=557s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | You've worked in mining, you've worked in fishing, you've worked in steel, you've worked in every major industry. You've had your back shoulder to shoulder with these guys that our politicians are desperate to relate to every four years, right?" I can still see Hillary doing the shots of rye, dribbling down her chin, with the steel workers. I mean, these are the people that I work with every single day. | 584 | 606 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=584s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | "And if you have something to say about their thoughts, collectively, it might be time to think about it. Because, dude, you know, four years." So, that's in my head, testicles are on my chin, thoughts are bouncing around. And, after that shoot, "Dirty Jobs" really didn't change, in terms of what the show is, but it changed for me, personally. And now, when I talk about the show, | 606 | 633 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=606s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | I no longer just tell the story you heard and 190 like it. I do, but I also start to talk about some of the other things I got wrong; some of the other notions of work that I've just been assuming are sacrosanct, and they're not. People with dirty jobs are happier than you think. As a group, they're the happiest people I know. And I don't want to start whistling "Look for the Union Label," | 633 | 660 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=633s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and all that happy-worker crap. I'm just telling you that these are balanced people who do unthinkable work. Roadkill picker-uppers whistle while they work, I swear to God -- I did it with them. They've got this amazing sort of symmetry to their life. And I see it over and over and over again. So I started to wonder what would happen if we challenged some of these sacred cows? | 660 | 681 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=660s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | Follow your passion -- we've been talking about it here for the last 36 hours. Follow your passion -- what could possibly be wrong with that? It's probably the worst advice I ever got. (Laughter) Follow your dreams and go broke, right? I mean, that's all I heard growing up. I didn't know what to do with my life, but I was told if you follow your passion, it's going to work out. | 681 | 701 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=681s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | I can give you 30 examples right now. Bob Combs, the pig farmer in Las Vegas who collects the uneaten scraps of food from the casinos and feeds them to his swine. Why? Because there's so much protein in the stuff we don't eat, his pigs grow at twice the normal speed, and he's one rich pig farmer. He's good for the environment, he spends his days doing this incredible service, | 701 | 721 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=701s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | and he smells like hell, but God bless him. He's making a great living. You ask him, "Did you follow your passion here?" and he'd laugh at you. The guy's worth -- he just got offered like 60 million dollars for his farm and turned it down, outside of Vegas. He didn't follow his passion. He stepped back and he watched where everybody was going, and he went the other way. | 721 | 741 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=721s | Learning from dirty jobs | Mike Rowe | |
IRVdiHu1VCc | And I hear that story over and over. Matt Freund, a dairy farmer in New Canaan, Connecticut, who woke up one day and realized the crap from his cows was worth more than their milk, if he could use it to make these biodegradable flowerpots. Now he's selling them to Walmart, right? Follow his passion? The guy's -- come on. So I started to look at passion, I started to look at efficiency vs. effectiveness. | 741 | 767 | https://www.youtube.com/watch?v=IRVdiHu1VCc&t=741s | Learning from dirty jobs | Mike Rowe |