title
stringlengths
12
112
published
stringlengths
19
23
url
stringlengths
28
28
video_id
stringlengths
11
11
channel_id
stringclasses
5 values
id
stringlengths
16
31
text
stringlengths
0
596
start
float64
0
37.8k
end
float64
2.18
37.8k
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1702.82
the whole parse tree here and the island on the top layer suggests that these two things
1,702.82
1,712.2
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1707.4199999999998
should be parsed independently from each other and therefore also processed independently
1,707.42
1,720.84
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1712.2
from each other. So here is my suggestion to to extend this and maybe Hinton's already
1,712.2
1,728.84
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1720.8400000000001
thought of this, but I would suggest that the this attention mechanism here is modulated
1,720.84
1,737.88
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1728.8400000000001
by how close two things are in the parse tree. Okay, so what would that be? So for a given
1,728.84
1,743.6
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1737.88
a given vector, it would be how much do you attend to this vector right here? Well, a
1,737.88
1,749.64
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1743.6000000000001
lot because it agrees with you, right? It you know, this the softmax of the inner product
1,743.6
1,756.04
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1749.64
would be high, it agrees with you. And also, it is in the same, it is same branch of the
1,749.64
1,760.76
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1756.0400000000002
parse tree. So that's perfect, right? This one right here doesn't agree with you, but
1,756.04
1,766
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1760.7600000000002
is in the same branch. So it could potentially later agree with you through a consensus algorithm.
1,760.76
1,771.96
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1766.0
However, this one over here, I, you probably shouldn't attend to that too much, even though
1,766
1,778.12
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1771.96
it points in the same direction, because it's in a different branch of the parse tree, you
1,771.96
1,783.84
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1778.12
shouldn't attend zero to it, like because these branches on top, they could change.
1,778.12
1,789.88
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1783.84
And you know, by you sending information there, this one could change the the top structure
1,783.84
1,795.02
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1789.88
here that could agree more with your branch of the parse tree and so on. So my suggestion
1,789.88
1,804.48
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1795.02
would be that let's not only get the softmax of the, let's not only get the softmax of
1,795.02
1,812.28
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1804.48
the current layer things, but let's do x times and here we're going to have a sum. So this
1,804.48
1,819.76
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1812.28
is going to be k. And let's say we're at we're at layer L. And this is layer one, this is
1,812.28
1,826.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1819.76
layer two, this is layer three, going to number them from the top, actually from the bottom,
1,819.76
1,834.04
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1826.56
layer m, layer m minus one, and this is layer L, I'm I suck at this. So from the current
1,826.56
1,843.12
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1834.04
layer, I want to go up the hierarchy until layer one. And I'm going to take the softmax
1,834.04
1,856.04
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1843.12
of the representation at layer L, at layer k, where I'm at x k transposed, like this.
1,843.12
1,862.48
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1856.04
What we aggregate is still the the values on the current layer, but how much we should
1,856.04
1,867.48
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1862.4799999999998
attend to that should be dependent on the parse tree. And we do that like this. And
1,862.48
1,877.64
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1867.48
maybe we have like a kind of a lambda k, L minus k, L minus k, I hope you get what I
1,867.48
1,884.88
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1877.64
mean. So how much how much you aggregate this sum here, the sum here is weird. This should
1,877.64
1,895.4
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1884.88
go probably. Hi, it's future Yannick. And I just wanted to write that down again. So
1,884.88
1,902
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1895.4
because I've made some mistakes. Obviously, the sum here should be within the softmax
1,895.4
1,907.84
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1902.0
because you want to have aggregate the distributions in log space. And the softmax should still
1,902
1,915.88
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1907.8400000000001
be valid, you know, distribution. And then the lambda is exponentiated by k and k now
1,907.84
1,925.48
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1915.88
properly runs from the zero to all the way up the stacks. So big L would be the total
1,915.88
1,931.84
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1925.48
number of layers and little L would be the layer where you're currently at. And you can
1,925.48
1,939.16
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1931.8400000000001
clearly see that the contribution of these attention matrices, it is so lambda would
1,931.84
1,945.6
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1939.16
be something smaller than one. And therefore, the contribution is in the current layer is
1,939.16
1,951.48
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1945.6
the strongest, but also in the next one up is a bit weaker than one more up is even a
1,945.6
1,956.68
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1951.48
bit weaker, and so on. So you'd still have essentially the same mechanism as Hinton is
1,951.48
1,962.92
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1956.6799999999998
suggesting controlling for the fact that things are in different branches of the parse tree.
1,956.68
1,970.92
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1962.9199999999998
All right, back to classic Yannick, who is thoroughly confused by these things. Yeah,
1,962.92
1,975.84
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1970.92
I'm not good at I'm not good at coming up with math on the spot. But I hope you can
1,970.92
1,982.08
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1975.8400000000001
see what it's doing. So it is if if you simply take the first K, you would simply stay at
1,975.84
1,987.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1982.0800000000002
that layer and it would be what Hinton said. But what I'm saying is you should also consider
1,982.08
1,996.52
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1987.5600000000002
how much your top your higher layer, one layer up from you agrees with one layer up from
1,987.56
2,002.72
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t1996.52
the thing you want to attend to. So you also compute that inner product between between
1,996.52
2,008.36
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2002.72
the embeddings, and you add that to the softmax distribution. So initially, the softmax distribution
2,002.72
2,014.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2008.36
would be like you should attend to this thing and this thing, and this thing a lot. But
2,008.36
2,021.4
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2014.56
then the next up hierarchy would maybe say, Well, we agree, because you know, these are
2,014.56
2,026.4
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2021.4
in the same thing. But this one, maybe not so much. And you would add those together,
2,021.4
2,030.92
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2026.4
maybe with a lambda factor in here, and then you go one layer up, and it would say, Well,
2,026.4
2,037.08
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2030.92
okay, everything over here basically agrees, right? And here, but everything over here
2,030.92
2,041.64
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2037.0800000000002
basically doesn't agree. So you would add that maybe with a lambda squared, as you go
2,037.08
2,049.42
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2041.64
up the layers, it would be less and less important, but still, you'd consider it. All right. Now,
2,041.64
2,056.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2049.42
if this is going to work out, cite the channel. Now back to what Hinton says, this, this is
2,049.42
2,064.44
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2056.56
actually the system. This is the system. As in a nutshell, you're going to input the image
2,056.56
2,068.8
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2064.44
at the bottom. And Hinton says you could use like a conv net at the very bottom to get
2,064.44
2,074.72
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2068.8
it into the columns. But then you're going to every time step pass information up the
2,068.8
2,082.2
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2074.72
columns down the columns, and between the same layer of the different columns. And that's
2,074.72
2,088.12
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2082.2
going to, in some point, this is going to stabilize, I don't know if it has cycles,
2,082.2
2,095.12
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2088.12
it probably doesn't have cycles, this probably does not have cycles. So at some point, this
2,088.12
2,102.42
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2095.12
comes to an end. And if that comes to an end, it should be that the object level embeddings
2,095.12
2,108
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2102.42
agree on an object, the part level embeddings agree on what parts there are the sub parts
2,102.42
2,113.2
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2108.0
agree, and so on. And they form these islands, these islands give rise to a parse tree. And
2,108
2,117.6
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2113.2000000000003
the parse tree can tell you what object is there, what is it made of? And where are these
2,113.2
2,128.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2117.6
parts in the image and so on. So exactly, that is it. And now, we're going to look at
2,117.6
2,135.52
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2128.56
what Hinton calls some design decisions. How many levels are there? About five. Okay, we
2,128.56
2,142.2
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2135.52
can skip that. How fine grained are the locations? Hinton says, you could be as fine grained
2,135.52
2,148.8
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2142.2
as pixels, or they could correspond to larger image patches. You and he says you could do
2,142.2
2,155.96
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2148.7999999999997
convolutional neural network to get it in there. Does the bottom op net look at nearby
2,148.8
2,162.76
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2155.96
locations? He says, yes, the bottom op net. So this this is not the attention network.
2,155.96
2,168.76
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2162.76
That's the bottom op network, it could look at nearby locations. But Hinton imagines that
2,162.76
2,174.44
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2168.76
if you have bottom up, top down, and if you have attention drawing in for information,
2,168.76
2,182.12
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2174.44
and if you maybe limit that attention to a neighborhood, then then the the attention
2,174.44
2,186.36
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2182.12
will do the job because you can have instead of looking at neighboring locations in the
2,182.12
2,192.04
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2186.3599999999997
bottom up network, you can simply in two time steps, aggregate that information. So you
2,186.36
2,197.68
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2192.04
can do bottom up here, bottom up here, and then using the attention, the lateral mechanism,
2,192.04
2,205
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2197.68
you can pass that information around this way. And also, it is not as biasing the network
2,197.68
2,212.02
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2205.0
to the immediate neighborhood. So the attention mechanism can sort of look farther, which
2,205
2,217.2
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2212.02
conflicts with what he's saying on top that the attention mechanism might only be looking
2,212.02
2,224.16
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2217.2
at the neighbors, I think there are different possibilities here. And only looking at neighbors
2,217.2
2,230.52
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2224.16
is actually one of the solution to the problem of having, you know, kind of similar vectors
2,224.16
2,236.64
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2230.52
at very distant locations at down the levels. But I think it's not as as good a solutions
2,230.52
2,241
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2236.64
to simply look at how close things are in pixel space, because even though things are
2,236.64
2,248.2
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2241.0
close in pixel space, they might be far away in the parse tree space. How does the attention
2,241
2,255.62
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2248.2
work? We've already looked at this. So the way that one location attends to another location
2,248.2
2,262.38
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2255.62
is going to be the softmax of the inner product between the embeddings here. And the values
2,255.62
2,271.6
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2262.38
are also going to be just the embeddings at layer at that layer. The visual input, he
2,262.38
2,281.58
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2271.6
says convolutional net could be used. Color and texture. He says, he makes he gives this
2,271.6
2,288.72
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2281.58
example, like if you know, if an object is entirely pale, or entirely green, or entirely,
2,281.58
2,293.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2288.72
I don't even know how to pronounce this, the color of a part is straightforward. But what
2,288.72
2,300.36
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2293.56
color is the whole object. So this entire notion of capsules, by the way, Hinton imagines
2,293.56
2,310.58
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2300.3599999999997
this as these embeddings represent kind of properties of the object so that the cat ear
2,300.36
2,316.24
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2310.58
embedding represents not only the fact that it is a cat ear, but also different properties
2,310.58
2,322.62
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2316.24
about the cat ear and even its location in the image is in the embedding. And, you know,
2,316.24
2,328.76
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2322.62
we know that transformers, they must be doing something like this, because we feed in positional
2,322.62
2,333.48
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2328.7599999999998
embeddings, for example, at the very bottom, and it can still, you know, compute things
2,328.76
2,341.1
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2333.4799999999996
in terms of positions. So that's the there's an intrinsic connection between kind of capsules
2,333.48
2,348.5
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2341.1
and the kind of transformer architecture. He says, one of the motivations of glom was
2,341.1
2,354
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2348.5
the idea that the whole object has a compound color, which might be called pale green or
2,348.5
2,361.68
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2354.0
move. And at the object level, every location belonging to the object has exactly the same
2,354
2,369.12
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2361.68
compound color. So the object is whatever this all over. When deciding which other locations
2,361.68
2,372.92
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t2369.12
the object level attend to preference would be given two locations with a similar compound
2,369.12
2,380.72