title
stringlengths 12
112
| published
stringlengths 19
23
| url
stringlengths 28
28
| video_id
stringlengths 11
11
| channel_id
stringclasses 5
values | id
stringlengths 16
31
| text
stringlengths 0
596
| start
float64 0
37.8k
| end
float64 2.18
37.8k
|
---|---|---|---|---|---|---|---|---|
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2372.92
|
color. So what he's saying right here is that, you know, you could give preference to two
| 2,372.92 | 2,387.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2380.72
|
similar color locations when you decide what you want to attend to. But the color isn't
| 2,380.72 | 2,394.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2387.8399999999997
|
as easy as simply saying what color is there in the location that you are at. But you could
| 2,387.84 | 2,403.36 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2394.76
|
be so if this is green, and this here is blue, then the bottom layer would say yes, I'm green
| 2,394.76 | 2,409 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2403.36
|
and yes, I'm blue. But they could also be saying, well, I am part of a green blue object,
| 2,403.36 | 2,415.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2409.0
|
right? And then the the higher layer here, you know, attending or caring about multiple
| 2,409 | 2,421.18 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2415.76
|
or bigger region, its color would then be, you know, green, blue, and the consensus could
| 2,415.76 | 2,427.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2421.18
|
reach on, well, we are a green blue object, even though the object isn't a pure green
| 2,421.18 | 2,436 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2427.16
|
or pure blue all throughout. So he, I think, yeah, it's it's, I think it's a side suggestion.
| 2,427.16 | 2,443 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2436.0
|
Maybe he has this as a core motivation between the system. But it's just interesting to see
| 2,436 | 2,450.6 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2443.0
|
how he thinks of things and he extends the color here to textures and even shapes. The
| 2,443 | 2,455.04 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2450.6
|
individual texture elements have their own shapes and poses in spatial relationships,
| 2,450.6 | 2,459.2 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2455.04
|
but an object with a textured surface has exactly the same texture everywhere at the
| 2,455.04 | 2,466.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2459.2
|
object level. glom extends this ideas to shapes, an object may have parts that are very different
| 2,459.2 | 2,471.18 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2466.4
|
from one another. But at the object level, it has exactly the same compound shape in
| 2,466.4 | 2,476.28 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2471.18
|
all of the location that it occupies, basically saying that, okay, every pixel that's part
| 2,471.18 | 2,481.72 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2476.28
|
of a cat head is a is a cat head has the shape of a cat head, even though the individual
| 2,476.28 | 2,487.66 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2481.7200000000003
|
locations might not recognize that. And that information could be passed around through
| 2,481.72 | 2,495.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2487.6600000000003
|
this consensus mechanism over time. So the cluster discovery versus cluster formation,
| 2,487.66 | 2,502.04 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2495.1600000000003
|
we've seen that and he makes a lot of, he makes a lot of analogies to face recognition.
| 2,495.16 | 2,507.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2502.04
|
But yeah, the clusters are not the islands of similar embedding vectors at a level can
| 2,502.04 | 2,512.1 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2507.16
|
be viewed as clusters, but these clusters are not discovered in immutable data. They
| 2,507.16 | 2,517.82 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2512.1
|
are formed by the interaction between the intra level process that favors islands of
| 2,512.1 | 2,523.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2517.82
|
similarity and dynamically changing suggestions coming from the locations embedding at adjacent
| 2,517.82 | 2,531.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2523.4
|
levels. So the core here is really this consensus algorithm that creates these clusters. And
| 2,523.4 | 2,535.72 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2531.88
|
yeah, the clustering algorithm doesn't work by simply looking at embeddings and deciding
| 2,531.88 | 2,540.68 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2535.7200000000003
|
which ones go together, but the embeddings themselves update themselves in order to form
| 2,535.72 | 2,548.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2540.6800000000003
|
clusters. And yeah, this is a replicating embedding vectors. This is a response to a
| 2,540.68 | 2,555.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2548.76
|
criticism that I guess he got where someone said, Well, why don't Why do you represent
| 2,548.76 | 2,559.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2555.2400000000002
|
if you have these, you know, these columns at the bottom, it makes sense, you have all
| 2,555.24 | 2,562.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2559.12
|
the different vectors, but then as you go up, you know, you have that kind of the same
| 2,559.12 | 2,568.6 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2562.8399999999997
|
vector for all locations, because it's the same object, why does it make sense to replicate
| 2,562.84 | 2,575.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2568.6
|
that everywhere, and not just have one, because, you know, in a database, we just have one.
| 2,568.6 | 2,581.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2575.88
|
And it basically says that, in order to reach the consensus, first of all, it's important
| 2,575.88 | 2,585.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2581.12
|
to have different vectors, they might be slightly different. So they might have some nuance
| 2,581.12 | 2,590.56 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2585.4
|
in them, because, you know, they might get pulled into different directions from the
| 2,585.4 | 2,597.6 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2590.56
|
sign of bottom up signal, then from the consensus algorithm on the same layer. So I, you know,
| 2,590.56 | 2,603.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2597.6
|
I believe that it is that is important. Here, I think it's just this is a criticism he got.
| 2,597.6 | 2,611 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2603.84
|
And then he decided to put this in here, learning islands. So what we haven't discussed about
| 2,603.84 | 2,618.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2611.0
|
this yet is how this is trained. And Hinton says this is trained as a denoising auto encoder.
| 2,611 | 2,623.98 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2618.48
|
Let us assume that glom is trained to reconstruct at its output, the uncorrupted version of
| 2,618.48 | 2,632.2 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2623.98
|
an image from which some region has been have been removed. So he goes into self supervised
| 2,623.98 | 2,638.36 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2632.2
|
learning with this system. This objective should ensure that information about the input
| 2,632.2 | 2,643.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2638.36
|
is preserved during the forward pass. And if the regions are sufficiently large, it
| 2,638.36 | 2,648.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2643.52
|
should also ensure that identifying familiar objects will be helpful for filling in the
| 2,643.52 | 2,656.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2648.48
|
missing regions. To encourage islands of near identity, we need to add a regularizer. And
| 2,648.48 | 2,660.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2656.84
|
experience shows that a regularizer that simply encourages similarity between the embeddings
| 2,656.84 | 2,667.28 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2660.88
|
of nearby locations can cause representations to collapse. All the embedding vectors may
| 2,660.88 | 2,672.32 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2667.28
|
become very small, so that they are all very similar. And the reconstruction will then
| 2,667.28 | 2,676.92 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2672.32
|
use very large weights to deal with the very small scale to prevent collapse. And then
| 2,672.32 | 2,683.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2676.92
|
he says contrastive learning is the answer to this. So how do you regularize the model
| 2,676.92 | 2,691.92 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2683.84
|
such that this consensus is formed? He says contrastive learning might be useful, but
| 2,683.84 | 2,698.02 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2691.92
|
you can't simply apply it straight out. So it learns to make representations of two different
| 2,691.92 | 2,703.18 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2698.02
|
crops of the same image agree, and the representations of two crops from different images disagree.
| 2,698.02 | 2,709.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2703.1800000000003
|
But this is not a sensible thing to do. If our aim is to recognize objects, if crop one
| 2,703.18 | 2,714.96 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2709.2400000000002
|
contains objects A and B and crop two from the same image contains objects B and C, it
| 2,709.24 | 2,720.34 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2714.96
|
does not make sense to demand that the representation of the two crops is the same at the object
| 2,714.96 | 2,727.56 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2720.34
|
level. Okay, so he says that contrastive learning is good, but you have to pay very careful
| 2,720.34 | 2,736.88 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2727.56
|
attention at which layer you employ it. Because, you know, if you go down far enough, then
| 2,727.56 | 2,742.36 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2736.88
|
contrastive learning, especially, you know, this this type where you crop the image into
| 2,736.88 | 2,746.28 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2742.36
|
different parts, and you say, well, since it's the same image, the representations should
| 2,742.36 | 2,751.48 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2746.28
|
agree in would say, well, at the top layer, yes, but at the bottom layer, certainly not
| 2,746.28 | 2,759.72 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2751.48
|
because they display different things, right. So you have to be careful where you apply
| 2,751.48 | 2,767.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2759.7200000000003
|
this contrastive learning. And he gives a bunch of suggestions on how to solve that.
| 2,759.72 | 2,773.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2767.4
|
He says things like, well, negative examples, for example, might not might not even be needed.
| 2,767.4 | 2,778.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2773.12
|
Well, that's it. Sorry, that's a different thing. So the obvious solution is to regularize
| 2,773.12 | 2,783.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2778.8399999999997
|
the bottom up and top down neural networks by encouraging each of them to predict the
| 2,778.84 | 2,793.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2783.16
|
consensus option. Yeah, this is the way to geometric mean of the predictions coming from
| 2,783.16 | 2,798.16 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2793.24
|
the top down and bottom up networks, the attention weighted average of the embeddings at nearby
| 2,793.24 | 2,804.6 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2798.16
|
locations at the previous time step, the previous state of end, I guess, and there should be
| 2,798.16 | 2,810.92 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2804.6
|
an end, and the previous state of the embedding, training the inter level prediction to agree
| 2,804.6 | 2,815.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2810.92
|
with the consensus will clearly make the islands found during feed forward inference be more
| 2,810.92 | 2,825.74 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2815.52
|
coherent. So he says you could regularize the model to, to regress to the consensus
| 2,815.52 | 2,833.96 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2825.74
|
option. So it's sort of like a self a self regression. And he asks whether or not that
| 2,825.74 | 2,840.9 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2833.9599999999996
|
will lead to a collapse. Because if you don't have negative examples and contrastive learning,
| 2,833.96 | 2,848.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2840.8999999999996
|
this could lead to simply a collapse. An important question is whether this type of training
| 2,840.9 | 2,853.08 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2848.24
|
will necessarily cause collapse if it is not accompanied by training the inter level predictions
| 2,848.24 | 2,858.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2853.08
|
to be different for negative examples that use the consensus options for unrelated spatial
| 2,853.08 | 2,865.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2858.4
|
contexts. So here is that problem, right? If you use the consensus opinion for unrelated
| 2,858.4 | 2,875.32 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2865.84
|
spatial context, that might be a problem. He says using layer or batch norm should reduce
| 2,865.84 | 2,880.24 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2875.3199999999997
|
the tendency to collapse, but a more important consideration may be the achievability of
| 2,875.32 | 2,888.9 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2880.24
|
the goal. It goes into why regularization could help. And he says, if however, an embedding
| 2,880.24 | 2,893.84 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2888.8999999999996
|
at one location is free to choose which embeddings at other locations, it should resemble, the
| 2,888.9 | 2,898.52 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2893.8399999999997
|
goal can be achieved almost perfectly by learning to form islands of identical vectors and attending
| 2,893.84 | 2,905.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2898.52
|
almost entirely to other locations that are in the same island. And I don't know, I don't
| 2,898.52 | 2,913.4 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2905.4
|
know if this is what I suggested. So I guess this is kind of a convoluted paragraph. And
| 2,905.4 | 2,918.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2913.4
|
I had to also read it multiple times. And I still don't exactly know what he's trying
| 2,913.4 | 2,925.8 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2918.76
|
to say right here. But I think what he's saying is that what we want to do is we want to sort
| 2,918.76 | 2,933.38 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2925.8
|
of regularize the network to produce this consensus, right? So we have a bottom up signal,
| 2,925.8 | 2,939.08 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2933.38
|
a top down signal, we have a current value, and we have the signal from the attention
| 2,933.38 | 2,945.8 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2939.08
|
mechanism. Now, what we want to do is we want to reach a consensus such that these islands
| 2,939.08 | 2,953.46 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2945.8
|
form. However, if you attend to any sort of things here that have nothing to do with you,
| 2,945.8 | 2,959.1 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2953.46
|
you might not be able to reach this consensus, right? That's I think that's the problem I
| 2,953.46 | 2,966.2 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2959.1
|
think he's touching on the problem that I said before. So what he says is, you know,
| 2,959.1 | 2,974.44 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2966.2
|
what you should do is you should simply attend to things that are in the same islands already.
| 2,966.2 | 2,979.12 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2974.44
|
So if an embedding at one location is free to choose which embedding at other locations,
| 2,974.44 | 2,985.28 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2979.12
|
it should resemble the goal can be achieved by learning to form islands of identical vectors
| 2,979.12 | 2,992.08 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2985.28
|
and attending almost entirely to other locations that are in the same island. Now, I think
| 2,985.28 | 2,998.76 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2992.0800000000004
|
here, what he's doing, he makes the case for the attention mechanism itself, right? So
| 2,992.08 | 3,005.74 |
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
|
2021-02-27 15:47:03
|
https://youtu.be/cllFzkvrYmE
|
cllFzkvrYmE
|
UCZHmQk67mSJgfCCTn7xBfew
|
cllFzkvrYmE-t2998.76
|
he says, if if we simply draw in information from the same layer here, you know, anything,
| 2,998.76 | 3,010.8 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.