title
stringlengths
12
112
published
stringlengths
19
23
url
stringlengths
28
28
video_id
stringlengths
11
11
channel_id
stringclasses
5 values
id
stringlengths
16
31
text
stringlengths
0
596
start
float64
0
37.8k
end
float64
2.18
37.8k
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3420.2799999999997
entries. So there's this entries, there is color and then link and then here, the URL would go on,
3,420.28
3,433.16
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3426.44
right? And it is the, in fact, the the only document in the internet, at least these these
3,426.44
3,442.68
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3433.16
authors claim that contains these URLs, but many of the URLs are repeated many times. In fact, here
3,433.16
3,448.36
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3442.68
you can see that these are the continuations of the URLs, right? This one, even though it's contained
3,442.68
3,455.48
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3448.36
in one document, it's actually repeated 359 times, and so on. So this is the the the the
3,448.36
3,462.68
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3455.48
one. So this is a playground. They say, okay, this document was in the training data of GBT two.
3,455.48
3,471.08
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3463.8
Here, we know how often each of these strings appeared in the document. So they can directly
3,463.8
3,479.16
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3471.08
make an experiment. How often does a string need to be present for the model to memorize it? They
3,471.08
3,486.68
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3479.16
simply order by the number of total occurrences right here, as you can see, and they ask each of
3,479.16
3,494.92
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3486.68
these models whether or not it has memorized the string. And they do this by inputting this. So this
3,486.68
3,501.56
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3494.92
is the input. And they simply sample if the model manages to output any of these URLs, they consider
3,494.92
3,508.68
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3501.56
that to be memorized. If not, then not. If it doesn't memorize it, they have a second trick that
3,501.56
3,516.2
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3508.68
if model can get half a point, if they input this first random sequence, I think they put six tokens
3,508.68
3,522.84
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3516.2
of this random sequence. And if then the model completes, then they say, ah, it has memorized it,
3,516.2
3,531.72
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3522.8399999999997
right. So you can see right here, it appears that the this large language model needs this needs a
3,522.84
3,539.24
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3531.72
string, let's say 20 times or higher for it to memorize it. And you can also see the trend right
3,531.72
3,545.72
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3539.24
here that if you go to the smaller models, they need a lot more in order to memorize them because
3,539.24
3,553.24
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3545.72
they have less weights, they can't afford to memorize stuff easily, right? They need to extract
3,545.72
3,559
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3553.24
the pattern. So they'd rather forget about the string incur a loss and focus on other training
3,553.24
3,567.32
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3559.0
examples. So yeah, two things in this direction, smaller models in this direction, larger models.
3,559
3,574.2
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3567.32
So that means that something like GPT-3 will have this problem much more pronounced. So that's the
3,567.32
3,581.8
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3574.2
bad news about this result. The good news about this result is that this is the case where you
3,574.2
3,588.92
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3581.8
have fairly random sequences, right? These, even you know, that if tokenizing this is not going to
3,581.8
3,593.48
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3588.92
be natural text, and there are these, you know, random, these Reddit URLs have these random
3,588.92
3,601
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3593.48
prefixes. So this is very much this sort of outlier case. It's a pretty clever case study to
3,593.48
3,610.28
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3601.0
find this document, I have to say, but it is sort of good news that this is not the usual case,
3,601
3,616.28
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3610.28
this is really the case that this data is very, very prone to being memorized, right? Because it's
3,610.28
3,626.52
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3616.28
not patternable. And it's very random. And yeah, so okay, so that was that was that.
3,616.28
3,637.88
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3628.6000000000004
As I said, the amount of hedging right here is is really, really, like, it's a lot. They discuss
3,628.6
3,643.56
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3637.88
what you can do with it, you can train with differential privacy, though that doesn't really
3,637.88
3,649.96
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3643.56
help, as we said, because some of these strings are included in, you know, more than one time.
3,643.56
3,658.6
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3651.88
You can curate the training data, which doesn't really help because the training data is too large.
3,651.88
3,664.2
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3659.6400000000003
You can limit impact of memorization on downstream applications. So if you fine tune,
3,659.64
3,670.92
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3664.2
but we don't know exactly what fine tuned models forget, and what they retain, or you can audit,
3,664.2
3,675.96
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3670.9199999999996
which is essentially what this paper paper right here does. And that seems like a that seems like
3,670.92
3,683.96
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3677.24
seems like a good, you know, the best strategy we have so far is is to audit these models. And
3,677.24
3,692.2
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3684.52
yeah, so I wanted to quickly check out also the appendix, the appendix here shows sort of these
3,684.52
3,698.28
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3692.2
graphs for the other methods. And it is very cool, if you want to, if you want to check that out,
3,692.2
3,705.4
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3698.2799999999997
and it has sort of categorization of what they find as these memorized pieces of text. But what
3,698.28
3,713.4
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3705.3999999999996
my main point was right here is that this paper shows a problem, let's say, with these large
3,705.4
3,720.12
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3713.3999999999996
language models, namely that they memorize certain pieces of training data. While that sounds scary,
3,713.4
3,726.2
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3720.12
I feel that the nature of the data that it remembers is very particular. So not you cannot
3,720.12
3,731.4
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3726.2
extract any piece of training data, the nature is very particular. It's the sort of outlier ish
3,726.2
3,742.04
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3732.12
training data points. And also, it very, very, very often, it isn't enough that it just is there
3,732.12
3,749.24
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3742.04
one time. So even when they say, this piece of information is only in one document, very often,
3,742.04
3,757.56
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3749.24
it appears many times in that document, that together with the sort of non pattern ability
3,749.24
3,764.04
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3757.56
of the data that it memorizes right here, actually makes me fairly, fairly optimistic,
3,757.56
3,768.52
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3764.6
more optimistic than I would have thought, honestly, about these language models.
3,764.6
3,777.16
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3770.3599999999997
Yes, so we'll see what the future brings. As I said, this is going to be more pronounced
3,770.36
3,787.08
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3777.16
in larger models. And this is not the only problem with these models, as my GPT-3 Google search in
3,777.16
3,795.24
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3787.08
that video shows. Alright, I hope this was enjoyable. Let me know what you think and maybe
3,787.08
3,807.8
Extracting Training Data from Large Language Models (Paper Explained)
2020-12-26 19:42:56
https://youtu.be/plK2WVdLTOY
plK2WVdLTOY
UCZHmQk67mSJgfCCTn7xBfew
plK2WVdLTOY-t3795.24
3,795.24
3,807.8
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t0.0
Hi, there. Today, we'll look at how to represent part-whole hierarchies in a neural network
0
13.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t6.4
by the legend himself, Jeffrey Hinton. He describes a system also known as GLOM, that
6.4
22.14
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t13.56
is a new approach to processing visual information using neural networks. And interestingly,
13.56
29.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t22.14
the paper starts off by saying, this paper does not describe a working system. So this
22.14
36.52
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t29.56
is an idea paper, Jeffrey Hinton's suggestion of how we should go about solving vision or
29.56
44.14
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t36.519999999999996
furthering vision in the AI community. He says openly, these are just ideas. Please
36.52
50.62
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t44.14
prove me right, prove me wrong, try them out, and so on. And I absolutely welcome this.
44.14
55.32
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t50.62
Idea papers is a thing that I think we have lost as a community because everything needs
50.62
61.28
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t55.32
to be state of the art and so on. This is super cool. And I encourage more people to
55.32
65.7
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t61.28
do it. I'm not saying you're going to have the same kind of success with an idea paper
61.28
72.78
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t65.7
as Jeff Hinton. He is banking on his name in large part with this. But nevertheless,
65.7
77.24
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t72.78
it's just an archive paper. Like I see people complaining, this would never be possible
72.78
81.32
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t77.24000000000001
if it wasn't. Yeah, it wouldn't like people wouldn't pay attention, but you're welcome
77.24
88.08
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t81.32
to write your ideas and post them on archive, like, or write a blog post, make a YouTube
81.32
95.88
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t88.08
video, anyone has opinions. So, you know, go ahead. Yeah, so to the paper itself, glom,
88.08
105.36
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t95.88
glom, as you can see here, glom comes from the stems from agglomeration is a system that
95.88
112.28
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t105.36
instead it presents a single idea about representation, which allows advances made by several different
105.36
118.32
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t112.28
groups to be combined into a an imaginary system called glom. The advances include transformers,
112.28
125.08
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t118.32
neural field, contrastive representation learning, distillation and capsules. glom answers the
118.32
131.6
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t125.08
question, how can a neural network with fixed architecture parse an image into a part whole
125.08
138.88
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t131.6
hierarchy, which has different structure for each image. The idea is simply to use islands
131.6
145.04
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t138.88
of identical vectors to represent the nodes in the parse tree. If glom can be made to
138.88
149.64
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t145.04
work, it should significantly improve the interpretability of the representations produced
145.04
155.56
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t149.64
by transformer like systems when applied to vision or language. That's the abstract, we'll
149.64
161.48
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t155.56
dive into the system, we'll see what it's about. I think I can actually make a suggestion
155.56
170.36
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t161.48
to improve it. But maybe I'm way behind other folks. So what is the glom system? And what
161.48
175.28
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t170.35999999999999
are these parse tree about? And why does it come combine all of these things? And for
170.36
182.72
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t175.28
that, we look at so it has two core diagrams here. This is the first diagram. This is the
175.28
188.16
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t182.72
second diagram. And at first sight, they have little to do with each other. So let me try
182.72
194.98
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t188.16
to go about it like this, if you have an image, and it looks at vision very much in terms
188.16
204.36
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t194.98
of you have an image or a video, and you want to parse the image into kind of a tree. And
194.98
210.96
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t204.35999999999999
the tree should be sort of like a tree of objects and their parts. So let's say it's
204.36
218.12
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t210.96
an image of a car. So the whole notion is very, very object centric. So this is like
210.96
226.84
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t218.12
my best attempt at a car. And a parse tree for this image would look something like this.
218.12
232.16
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t226.84
All right, so this whole thing here is a car. So that's going to be your top node in the
226.84
240.64
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t232.16
parse tree. The car has different parts, namely, it has this cabin, it has a motor, and it
232.16
247.2
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t240.64
has wheels. So that is going to be those are going to be kind of downstream of that parse
240.64
255.08
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t247.2
tree. Then the cabin itself is going to have two segments here, windows, and maybe here
247.2
261.6
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t255.07999999999998
is the door area. So that is going to be window, window, door, and so on. So you get that we
255.08
265.88
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t261.59999999999997
what we want to do is we want to look at an image, sort of create this parse tree over
261.6
272.76
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t265.88
here, this is very much into the into the area of go fi, good old fashioned AI people
265.88
280.04
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t272.76
that want to understand a the world in terms of their symbolic representations and relation
272.76
286
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t280.04
of the symbols to each other. However, what Hinton is saying is that if you simply do
280.04
290.64
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t286.0
this, it's it's, you know, you can't really do this with neural networks, neural networks
286
296.96
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t290.64
are continuous and so on. So what would you have to do? In addition, we know that the
290.64
304.68
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t296.96
brain doesn't reconfigure itself every single time you get a new input. So the brain, even
296.96
310.32
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t304.68
though it has some neuroplasticity, while you look at the world and do inference in
304.68
315.72
GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton's Paper Explained)
2021-02-27 15:47:03
https://youtu.be/cllFzkvrYmE
cllFzkvrYmE
UCZHmQk67mSJgfCCTn7xBfew
cllFzkvrYmE-t310.32
the world, the connection stay the same. So what we need to do is we need to come up with
310.32
321.36