ahmedelsayed's picture
commit files to HF hub
2ffb90d
WEBVTT
00:00:00.120 --> 00:00:04.880
everyone I today I'd like to talk about
00:00:02.760 --> 00:00:07.399
uh learning from knowledge bases uh
00:00:04.880 --> 00:00:11.440
learning from in for knowledge bases
00:00:07.399 --> 00:00:14.799
this is kind of a a shift uh from a lot
00:00:11.440 --> 00:00:16.480
of the stuff that we've done so far uh
00:00:14.799 --> 00:00:18.439
and I'm going to be talking about like a
00:00:16.480 --> 00:00:20.480
different information Source some
00:00:18.439 --> 00:00:21.960
relatively different algorithms compared
00:00:20.480 --> 00:00:26.080
to the stuff that we talked about up
00:00:21.960 --> 00:00:28.880
until this point so um you know it might
00:00:26.080 --> 00:00:32.360
be uh interesting it might be different
00:00:28.880 --> 00:00:35.640
so uh get started with
00:00:32.360 --> 00:00:37.360
that so I'm going to be talking about
00:00:35.640 --> 00:00:40.000
knowledge bases and knowledge bases are
00:00:37.360 --> 00:00:43.039
basically a structured databases of
00:00:40.000 --> 00:00:46.079
knowledge and they can contain a lot of
00:00:43.039 --> 00:00:48.559
things but most commonly when people are
00:00:46.079 --> 00:00:50.600
talking about them they are talking
00:00:48.559 --> 00:00:53.160
about relational knowledge bases that
00:00:50.600 --> 00:00:55.559
include things like entities which are
00:00:53.160 --> 00:00:57.399
nodes in a graph and relations which are
00:00:55.559 --> 00:01:00.239
edges between
00:00:57.399 --> 00:01:02.079
nodes and
00:01:00.239 --> 00:01:03.879
I'll I'll talk about some examples of
00:01:02.079 --> 00:01:05.479
this in a little bit to make that a
00:01:03.879 --> 00:01:08.040
little bit more concrete and then some
00:01:05.479 --> 00:01:11.240
of the questions that we ask about these
00:01:08.040 --> 00:01:14.400
are how can we learn to create and
00:01:11.240 --> 00:01:16.799
expand knowledge bases with uh you know
00:01:14.400 --> 00:01:18.439
neural network based methods and then
00:01:16.799 --> 00:01:20.200
the second question is how can we learn
00:01:18.439 --> 00:01:22.600
from the information in knowledge bases
00:01:20.200 --> 00:01:24.720
to improve like neural network models or
00:01:22.600 --> 00:01:27.560
uh use them in effective
00:01:24.720 --> 00:01:31.479
ways and how can we use uh structured
00:01:27.560 --> 00:01:31.479
knowledge to answer questions
00:01:32.200 --> 00:01:37.159
so the first uh thing I'd like to talk
00:01:35.000 --> 00:01:40.960
about a little bit is types of knowledge
00:01:37.159 --> 00:01:43.079
bases and they come in several different
00:01:40.960 --> 00:01:46.119
varieties the first one I'd like to talk
00:01:43.079 --> 00:01:48.560
about is a very uh classical one called
00:01:46.119 --> 00:01:50.960
wordnet has anyone actually ever used
00:01:48.560 --> 00:01:53.479
wordnet
00:01:50.960 --> 00:01:55.520
before I see at least one person raising
00:01:53.479 --> 00:01:57.640
their hand so it's not entirely uh
00:01:55.520 --> 00:02:00.119
hasn't entirely disappeared has anyone
00:01:57.640 --> 00:02:03.240
heard of wordnet before
00:02:00.119 --> 00:02:05.079
okay more more people um so basically
00:02:03.240 --> 00:02:06.960
this used to be a really big thing in in
00:02:05.079 --> 00:02:10.440
natural language processing it's not So
00:02:06.960 --> 00:02:12.319
Much Anymore um but I I want to explain
00:02:10.440 --> 00:02:14.800
about it because I want to explain why
00:02:12.319 --> 00:02:17.360
this is maybe like less necessary to use
00:02:14.800 --> 00:02:19.599
but actual knowledge bases are still
00:02:17.360 --> 00:02:23.160
more necessary to
00:02:19.599 --> 00:02:26.280
use and so wordnet is a large database
00:02:23.160 --> 00:02:29.560
of words and specifically what it does
00:02:26.280 --> 00:02:32.720
is each word or something they call a
00:02:29.560 --> 00:02:37.120
syn set is a node and then there are
00:02:32.720 --> 00:02:42.560
relationships between nodes and the
00:02:37.120 --> 00:02:44.319
nodes can correspond to nouns um and or
00:02:42.560 --> 00:02:45.920
verbs or
00:02:44.319 --> 00:02:48.360
adjectives
00:02:45.920 --> 00:02:49.959
and nouns have different types of
00:02:48.360 --> 00:02:53.360
relations between them so they have
00:02:49.959 --> 00:02:56.280
things like an is a relation so like a
00:02:53.360 --> 00:03:00.040
hatchback is a type of car they are part
00:02:56.280 --> 00:03:02.840
of relations uh where a wheel is a part
00:03:00.040 --> 00:03:05.720
of a car um and they also make
00:03:02.840 --> 00:03:09.799
distinctions between types and instances
00:03:05.720 --> 00:03:12.400
so like Joe Biden is an instance of a
00:03:09.799 --> 00:03:16.560
president and president is the
00:03:12.400 --> 00:03:19.239
type so um verb relations are ordered by
00:03:16.560 --> 00:03:22.680
specificity so like communicate is more
00:03:19.239 --> 00:03:25.799
broad than talk so talk is you know
00:03:22.680 --> 00:03:27.519
generally a sub class of communicate and
00:03:25.799 --> 00:03:30.720
then whisper is generally a subass of
00:03:27.519 --> 00:03:33.159
talk so it's ordered in this way
00:03:30.720 --> 00:03:35.920
and then adjective relations are mostly
00:03:33.159 --> 00:03:37.720
antonyms so like wet and wet versus dry
00:03:35.920 --> 00:03:43.599
and other things like
00:03:37.720 --> 00:03:47.080
this um when I said sinets uh actually
00:03:43.599 --> 00:03:50.239
the each node is not a word despite the
00:03:47.080 --> 00:03:53.239
name word net it's a set of words that
00:03:50.239 --> 00:03:56.200
all have the same meaning so you might
00:03:53.239 --> 00:03:59.120
have artifact and thing would both
00:03:56.200 --> 00:04:00.879
correspond to this um node because they
00:03:59.120 --> 00:04:02.599
both mean basically the same thing so
00:04:00.879 --> 00:04:04.159
it's like sets of synonyms and this is
00:04:02.599 --> 00:04:07.599
also important when we talk about other
00:04:04.159 --> 00:04:09.920
types of uh knowledge bases as well and
00:04:07.599 --> 00:04:13.920
so what was this used for um this was
00:04:09.920 --> 00:04:17.160
used for for example uh trying to figure
00:04:13.920 --> 00:04:22.400
out whether trying to find all the cars
00:04:17.160 --> 00:04:24.440
that were mentioned in like a in a large
00:04:22.400 --> 00:04:27.440
set of text so you would go through you
00:04:24.440 --> 00:04:30.280
would identify all
00:04:27.440 --> 00:04:32.120
sinets or you would identify all words
00:04:30.280 --> 00:04:34.120
that corresponded to these sunsets and
00:04:32.120 --> 00:04:35.720
then you would take a step up and find
00:04:34.120 --> 00:04:38.800
motor car and you would know that like
00:04:35.720 --> 00:04:42.320
all of those were mentions of cars so
00:04:38.800 --> 00:04:45.520
like why don't we use wordnet very much
00:04:42.320 --> 00:04:45.520
anymore any
00:04:49.160 --> 00:04:52.840
ideas what would what would you do
00:04:51.080 --> 00:04:55.560
instead if I told you find all the cars
00:04:52.840 --> 00:04:55.560
in a big piece of
00:04:55.960 --> 00:05:00.160
text yeah just do something with the
00:04:58.280 --> 00:05:02.880
embeding just do something with
00:05:00.160 --> 00:05:04.560
embeddings yeah so you might get um you
00:05:02.880 --> 00:05:06.720
might get something and find all things
00:05:04.560 --> 00:05:10.360
that were close in embedding space to a
00:05:06.720 --> 00:05:10.360
car what what's another thing you might
00:05:11.560 --> 00:05:15.520
do like what I would do is I would
00:05:13.639 --> 00:05:17.080
download mistol and say does this
00:05:15.520 --> 00:05:19.880
sentence talk about a car and it would
00:05:17.080 --> 00:05:22.199
say yes or no and I I would you know or
00:05:19.880 --> 00:05:23.479
I would say find all the cars in this uh
00:05:22.199 --> 00:05:25.319
that are mentioned in the sentence and
00:05:23.479 --> 00:05:28.720
it would get them and sure that's like
00:05:25.319 --> 00:05:31.319
expensive but it's really easy so um you
00:05:28.720 --> 00:05:32.919
know there are other options that might
00:05:31.319 --> 00:05:36.720
be less expensive but that could solve a
00:05:32.919 --> 00:05:39.520
lot of the things so word not you know
00:05:36.720 --> 00:05:41.039
started out with more and more it it
00:05:39.520 --> 00:05:42.600
started out being very popular in
00:05:41.039 --> 00:05:44.039
natural language processing but now it's
00:05:42.600 --> 00:05:45.440
less so because we can get a lot of it
00:05:44.039 --> 00:05:47.639
from embeddings we can get a lot of it
00:05:45.440 --> 00:05:50.520
from language models
00:05:47.639 --> 00:05:52.759
itself um another thing that started
00:05:50.520 --> 00:05:55.759
maybe before wordnet or even around the
00:05:52.759 --> 00:05:58.840
same time as wordnet was this uh data
00:05:55.759 --> 00:06:00.800
base called psych and it was a manually
00:05:58.840 --> 00:06:04.160
curated database attempting to encode
00:06:00.800 --> 00:06:06.280
all common sense knowledge um and the
00:06:04.160 --> 00:06:08.759
project itself lasted for about 30 to 40
00:06:06.280 --> 00:06:11.840
years it might even still
00:06:08.759 --> 00:06:13.319
exist um and so they had this huge uh
00:06:11.840 --> 00:06:15.199
like hierarchy of all the different
00:06:13.319 --> 00:06:17.680
types of knowledge you could have it
00:06:15.199 --> 00:06:19.680
encoded knowledge about like events and
00:06:17.680 --> 00:06:21.479
like which events happened before other
00:06:19.680 --> 00:06:26.840
events and all these other stuff like
00:06:21.479 --> 00:06:29.039
this um but the problem with this is uh
00:06:26.840 --> 00:06:31.000
this was just too ambitious basically it
00:06:29.039 --> 00:06:35.680
was not possible to encode all of this
00:06:31.000 --> 00:06:37.440
manually by hand so people um like it it
00:06:35.680 --> 00:06:38.840
did it got part of the way there but
00:06:37.440 --> 00:06:40.240
that part of the way there was not
00:06:38.840 --> 00:06:42.560
enough for it to be really useful in
00:06:40.240 --> 00:06:45.199
Practical systems so it isn't this sort
00:06:42.560 --> 00:06:47.800
of method is not used as frequently
00:06:45.199 --> 00:06:51.240
now
00:06:47.800 --> 00:06:56.000
um a a followup one
00:06:51.240 --> 00:06:57.479
um which is it's successor is now uh the
00:06:56.000 --> 00:06:59.879
the most widely used knowledge Bas is
00:06:57.479 --> 00:07:03.240
something called dbpedia and the basic
00:06:59.879 --> 00:07:06.120
idea behind dbpedia is that while Psych
00:07:03.240 --> 00:07:07.840
is too difficult because they had people
00:07:06.120 --> 00:07:12.400
on the psych project who would go in and
00:07:07.840 --> 00:07:12.400
curate rules um for
00:07:13.280 --> 00:07:19.080
machines Wikipedia basically they have a
00:07:17.160 --> 00:07:21.080
very very large number of humans
00:07:19.080 --> 00:07:23.639
curating this structured data about
00:07:21.080 --> 00:07:25.199
entities in the world for humans they're
00:07:23.639 --> 00:07:27.879
creating it for humans because then you
00:07:25.199 --> 00:07:29.599
can put it on a Wikipedia page and you
00:07:27.879 --> 00:07:31.440
can look and see it says cardig melan
00:07:29.599 --> 00:07:34.160
University it has the former names of
00:07:31.440 --> 00:07:36.919
Carnegie melon um it has the motto of
00:07:34.160 --> 00:07:38.759
Carnegie melon the type of entity who it
00:07:36.919 --> 00:07:41.360
was established by and when and other
00:07:38.759 --> 00:07:42.840
stuff like that and because people are
00:07:41.360 --> 00:07:44.280
no longer creating it for machines
00:07:42.840 --> 00:07:46.280
they're creating it for humans people
00:07:44.280 --> 00:07:47.840
are like motivated to do this so like
00:07:46.280 --> 00:07:49.960
lots of people will do it for free so
00:07:47.840 --> 00:07:51.960
you can actually get a reasonably sized
00:07:49.960 --> 00:07:53.639
amount of data from this and actually
00:07:51.960 --> 00:07:55.720
cover you know like most of the entities
00:07:53.639 --> 00:07:57.080
in the world or not most of the entities
00:07:55.720 --> 00:08:00.120
in the world but most of the notable
00:07:57.080 --> 00:08:03.319
entities in uh part of the world that
00:08:00.120 --> 00:08:03.319
have high participation in
00:08:03.479 --> 00:08:09.800
Wikipedia um so now the the thing that a
00:08:08.039 --> 00:08:13.319
lot of people use is something called
00:08:09.800 --> 00:08:14.919
Wiki data this is not this name is a
00:08:13.319 --> 00:08:17.039
little bit of a misnomer because it's
00:08:14.919 --> 00:08:18.960
not actually that closely connected to
00:08:17.039 --> 00:08:20.639
Wikipedia they extract data from
00:08:18.960 --> 00:08:21.720
Wikipedia but they also extract it from
00:08:20.639 --> 00:08:24.400
lots of other
00:08:21.720 --> 00:08:27.520
sources and this is a curated database
00:08:24.400 --> 00:08:30.360
of entities um it's linked it's
00:08:27.520 --> 00:08:33.959
extremely large scale and it's
00:08:30.360 --> 00:08:38.080
multilingual and um this is an example
00:08:33.959 --> 00:08:39.680
of a thing from Richard fean um where
00:08:38.080 --> 00:08:42.680
people can go in and they can actually
00:08:39.680 --> 00:08:45.320
like add information and stuff like that
00:08:42.680 --> 00:08:47.440
um and you know it gives information
00:08:45.320 --> 00:08:50.959
about education and all kinds of other
00:08:47.440 --> 00:08:52.600
stuff so um for fun I can go to the wiki
00:08:50.959 --> 00:08:55.040
data
00:08:52.600 --> 00:08:59.360
site does anyone have an entity they'd
00:08:55.040 --> 00:08:59.360
like to know more about
00:09:01.640 --> 00:09:07.320
any any ideas maybe something that has
00:09:03.959 --> 00:09:07.320
been in the news recently
00:09:10.680 --> 00:09:16.160
or nobody brave enough to come up with
00:09:13.040 --> 00:09:18.360
an entity yeah
00:09:16.160 --> 00:09:20.640
Mamba that's a good one I'm actually not
00:09:18.360 --> 00:09:23.800
sure if that one's going to be in here
00:09:20.640 --> 00:09:27.720
um there's lots of mambas but I don't
00:09:23.800 --> 00:09:27.720
know about that particular Mamba let me
00:09:27.839 --> 00:09:31.200
see do you want to know about a
00:09:29.720 --> 00:09:33.399
different Mamba do you want about know
00:09:31.200 --> 00:09:36.040
about Mamba the research
00:09:33.399 --> 00:09:38.399
group so Mamba is a research group it's
00:09:36.040 --> 00:09:41.800
the modeling and Analysis for medicine
00:09:38.399 --> 00:09:44.800
research group um it focuses on
00:09:41.800 --> 00:09:48.000
mathematical biology and it's in the uh
00:09:44.800 --> 00:09:51.120
in this National Center for scientific
00:09:48.000 --> 00:09:52.519
research in France um the chairperson is
00:09:51.120 --> 00:09:55.360
this person and stuff like that so you
00:09:52.519 --> 00:10:00.200
can see it has all of these things so
00:09:55.360 --> 00:10:03.920
Mamba this Mamba is a node in the graph
00:10:00.200 --> 00:10:06.839
and then the edges are pointing um the
00:10:03.920 --> 00:10:09.440
edges are labeled with like instance of
00:10:06.839 --> 00:10:11.200
and then the next note is research group
00:10:09.440 --> 00:10:13.000
so research group is like another note
00:10:11.200 --> 00:10:17.120
in the graph and so you can click
00:10:13.000 --> 00:10:18.680
through this and it has its own ID and
00:10:17.120 --> 00:10:21.200
other things like
00:10:18.680 --> 00:10:22.839
this also you'll notice that research
00:10:21.200 --> 00:10:24.160
group is translated into lots of
00:10:22.839 --> 00:10:27.440
different languages in the world so you
00:10:24.160 --> 00:10:30.120
can use it multi multilingually and um
00:10:27.440 --> 00:10:33.880
and other things like that
00:10:30.120 --> 00:10:37.000
um even minor entities like Graham
00:10:33.880 --> 00:10:40.160
nuig are included in this and it has a
00:10:37.000 --> 00:10:42.240
little bit of um like information about
00:10:40.160 --> 00:10:45.480
me like my PhD was in Kyoto University
00:10:42.240 --> 00:10:45.480
in 2012 I am a
00:10:45.600 --> 00:10:52.079
human I I am male uh and first name last
00:10:50.519 --> 00:10:53.720
name University teacher computer
00:10:52.079 --> 00:10:56.279
scientist natural language processing
00:10:53.720 --> 00:10:58.639
this is all right um because this is
00:10:56.279 --> 00:11:00.240
mostly hand curated it even has the IDS
00:10:58.639 --> 00:11:04.240
of my advisor
00:11:00.240 --> 00:11:06.519
advisers um the reason why it has all of
00:11:04.240 --> 00:11:09.839
this stuff actually is because like 15
00:11:06.519 --> 00:11:12.160
years ago or like 10 years ago I entered
00:11:09.839 --> 00:11:14.399
in my uh my information into the
00:11:12.160 --> 00:11:16.240
mathematical genealogy project uh which
00:11:14.399 --> 00:11:18.880
is this project about who your advisers
00:11:16.240 --> 00:11:20.680
were because I wanted to see like who my
00:11:18.880 --> 00:11:22.800
mathematical like siblings were and
00:11:20.680 --> 00:11:24.519
stuff like that and uh somehow they
00:11:22.800 --> 00:11:27.360
managed to pull that out and keep this
00:11:24.519 --> 00:11:28.760
like 10 years later so um basically
00:11:27.360 --> 00:11:30.519
they're pulling information from like
00:11:28.760 --> 00:11:32.800
many many different structured data
00:11:30.519 --> 00:11:34.160
sources that they can use so uh they can
00:11:32.800 --> 00:11:37.480
pull it in there I don't know where they
00:11:34.160 --> 00:11:39.440
got that I'm human uh but maybe that was
00:11:37.480 --> 00:11:43.240
inferred from some piece of data
00:11:39.440 --> 00:11:44.760
somewhere online or something cool um
00:11:43.240 --> 00:11:46.839
another good thing about this that
00:11:44.760 --> 00:11:52.680
actually I didn't mention directly in
00:11:46.839 --> 00:11:52.680
the um in the lecture note or
00:11:54.680 --> 00:12:01.120
slides is that there's a query language
00:11:57.360 --> 00:12:04.320
for this yeah and a query language this
00:12:01.120 --> 00:12:06.839
query language is called Sparkle so
00:12:04.320 --> 00:12:10.680
there's a sequel for querying relational
00:12:06.839 --> 00:12:14.399
databases and Sparkle is for querying
00:12:10.680 --> 00:12:15.240
these uh knowledge bases and let me see
00:12:14.399 --> 00:12:18.279
if I
00:12:15.240 --> 00:12:22.560
can I asked chat
00:12:18.279 --> 00:12:24.560
GPT to write me a sparkle query to find
00:12:22.560 --> 00:12:26.839
all presidents of Carnegie melon
00:12:24.560 --> 00:12:31.160
University so let's see if Chad GPT is
00:12:26.839 --> 00:12:31.160
capable of doing that um
00:12:35.639 --> 00:12:39.680
okay that's a problem let me
00:12:41.279 --> 00:12:47.000
see okay there's there's an errand there
00:12:43.880 --> 00:12:48.360
but like if uh uh if I could find a I
00:12:47.000 --> 00:12:50.160
don't want to waste time in class like
00:12:48.360 --> 00:12:52.079
finding a working query but basically
00:12:50.160 --> 00:12:53.399
you can put it in a query and it allows
00:12:52.079 --> 00:12:56.120
you to do a lot of things that are
00:12:53.399 --> 00:13:00.519
similar to what you can do in SQL so you
00:12:56.120 --> 00:13:02.720
can find like all of the edges of nodes
00:13:00.519 --> 00:13:05.279
that satisfy a particular relation so
00:13:02.720 --> 00:13:07.360
you could say I want for Carnegie melon
00:13:05.279 --> 00:13:10.160
University to find all things that
00:13:07.360 --> 00:13:13.519
followed the like president of relation
00:13:10.160 --> 00:13:14.959
and that would give me all um you know
00:13:13.519 --> 00:13:18.680
all presidents of Carnegie melon
00:13:14.959 --> 00:13:20.440
University you can also like filter um
00:13:18.680 --> 00:13:22.160
filter by their start date and end date
00:13:20.440 --> 00:13:24.120
so find all of the preceden between a
00:13:22.160 --> 00:13:25.839
certain time and a another time or
00:13:24.120 --> 00:13:30.480
things like
00:13:25.839 --> 00:13:34.199
that so this is good if you want to get
00:13:30.480 --> 00:13:36.600
like high reli high reliability data um
00:13:34.199 --> 00:13:39.839
in a scalable way because like if I ask
00:13:36.600 --> 00:13:41.920
chat GPT like one of my favorite um one
00:13:39.839 --> 00:13:45.720
of my favorite queries for chat GPT is
00:13:41.920 --> 00:13:48.600
like name all of the name all of the
00:13:45.720 --> 00:13:51.959
presidents that were born uh east of the
00:13:48.600 --> 00:13:53.880
Mississippi River um and I've never
00:13:51.959 --> 00:13:56.519
successfully gotten chat GPT to be able
00:13:53.880 --> 00:13:57.800
to do this um because there's lots of
00:13:56.519 --> 00:13:59.560
presidents who were born east of the
00:13:57.800 --> 00:14:02.320
Mississippi River and it starts counting
00:13:59.560 --> 00:14:04.079
them it can't distinguish what position
00:14:02.320 --> 00:14:05.639
is east of the Mississippi and what
00:14:04.079 --> 00:14:09.120
position is the west west of the
00:14:05.639 --> 00:14:11.279
Mississippi but if you write a uh like a
00:14:09.120 --> 00:14:14.759
sparkle query it's not that hard to do
00:14:11.279 --> 00:14:16.480
that so there are um you know there are
00:14:14.759 --> 00:14:18.639
certain types of questions especially
00:14:16.480 --> 00:14:20.399
information aggregation and complex
00:14:18.639 --> 00:14:22.839
relations and stuff that uh language
00:14:20.399 --> 00:14:26.600
models are not very good
00:14:22.839 --> 00:14:28.120
at cool um so that's kind of an intro to
00:14:26.600 --> 00:14:31.240
knowledge bases why you might want to
00:14:28.120 --> 00:14:33.759
think about them any questions so far
00:14:31.240 --> 00:14:33.759
for
00:14:34.759 --> 00:14:39.720
discussion okay um I will move on next
00:14:38.320 --> 00:14:41.199
so the next thing I'd like to talk about
00:14:39.720 --> 00:14:43.839
is learning representations for
00:14:41.199 --> 00:14:45.519
knowledge bases um so knowledge bases
00:14:43.839 --> 00:14:48.000
are great but one problem is they're
00:14:45.519 --> 00:14:51.040
like inherently
00:14:48.000 --> 00:14:55.040
incomplete and even with extremely large
00:14:51.040 --> 00:14:58.279
scale uh it becomes impossible to have
00:14:55.040 --> 00:15:00.360
them be complete and the reason why is
00:14:58.279 --> 00:15:03.639
uh for examp example in Freebase which
00:15:00.360 --> 00:15:05.480
was the predecessor to Wiki data um 71%
00:15:03.639 --> 00:15:08.560
of humans didn't have a date of
00:15:05.480 --> 00:15:10.560
birth um and probably every human
00:15:08.560 --> 00:15:12.079
actually has a date of birth right um
00:15:10.560 --> 00:15:15.880
you know we're pretty much guaranteed
00:15:12.079 --> 00:15:17.639
for that to be the case so the issue is
00:15:15.880 --> 00:15:19.160
like for very famous entities you want
00:15:17.639 --> 00:15:21.040
lots of detailed information like you
00:15:19.160 --> 00:15:24.000
can know absolutely everything about Joe
00:15:21.040 --> 00:15:25.759
Biden or Barack Obama but you know at
00:15:24.000 --> 00:15:26.880
the same time for Less major entities
00:15:25.759 --> 00:15:28.079
you still want them in the knowledge
00:15:26.880 --> 00:15:30.079
base but you're not going to be able to
00:15:28.079 --> 00:15:31.519
get all that information or should you
00:15:30.079 --> 00:15:35.600
for privacy
00:15:31.519 --> 00:15:36.680
purposes and so the idea is um for
00:15:35.600 --> 00:15:38.079
information that's written on the
00:15:36.680 --> 00:15:40.600
internet somewhere can you perform
00:15:38.079 --> 00:15:42.759
relation extraction which essentially
00:15:40.600 --> 00:15:44.600
allows you to extract this information
00:15:42.759 --> 00:15:46.360
and create your own knowledge bases and
00:15:44.600 --> 00:15:47.680
stuff like this and this can also be
00:15:46.360 --> 00:15:50.079
useful if you want to create it for like
00:15:47.680 --> 00:15:52.199
a specialized domain or um or other
00:15:50.079 --> 00:15:55.000
stuff like
00:15:52.199 --> 00:15:59.519
that so there's a bunch of ways that
00:15:55.000 --> 00:16:03.079
people do this um and one kind of
00:15:59.519 --> 00:16:06.120
popular way that people have tried to do
00:16:03.079 --> 00:16:09.199
relation extraction is through uh
00:16:06.120 --> 00:16:12.560
leveraging consistency in embedding
00:16:09.199 --> 00:16:15.319
space and so this is the most famous
00:16:12.560 --> 00:16:17.959
example from word de uh what seems like
00:16:15.319 --> 00:16:21.880
ages ago uh in
00:16:17.959 --> 00:16:23.920
2013 and in the word Toc paper one of
00:16:21.880 --> 00:16:26.279
the big you know exciting things was
00:16:23.920 --> 00:16:28.639
essentially they demonstrated that
00:16:26.279 --> 00:16:30.120
vectors in embedding space had kind of
00:16:28.639 --> 00:16:31.839
in
00:16:30.120 --> 00:16:33.160
you know meaning and actually the
00:16:31.839 --> 00:16:34.600
vectors in embedding space could
00:16:33.160 --> 00:16:37.639
correspond to relations between
00:16:34.600 --> 00:16:39.480
embeddings so like uh we would have man
00:16:37.639 --> 00:16:41.000
pointing to woman in approximately the
00:16:39.480 --> 00:16:42.920
same direction that we had Uncle
00:16:41.000 --> 00:16:46.600
pointing to Aunt and King pointing to
00:16:42.920 --> 00:16:49.680
Queen and so um then you could do things
00:16:46.600 --> 00:16:51.440
like you could take Kings subtract out
00:16:49.680 --> 00:16:53.560
the vector that corresponded to
00:16:51.440 --> 00:16:58.360
plurality uh add the vector that
00:16:53.560 --> 00:17:00.839
corresponded to um you know uh to going
00:16:58.360 --> 00:17:04.319
from masculine to feminine words and
00:17:00.839 --> 00:17:05.559
then um like read the vector to that
00:17:04.319 --> 00:17:07.160
were plural and you'd be able to
00:17:05.559 --> 00:17:09.439
identify the plural by just knowing
00:17:07.160 --> 00:17:11.000
these two uh vectors the plural of green
00:17:09.439 --> 00:17:14.000
by just knowing those two
00:17:11.000 --> 00:17:14.000
vectors
00:17:14.160 --> 00:17:21.880
um but it turns out that you can either
00:17:18.199 --> 00:17:21.880
learn embeddings
00:17:22.720 --> 00:17:28.240
from like uh you can either learn
00:17:25.000 --> 00:17:30.400
embeddings from text or you can use the
00:17:28.240 --> 00:17:32.039
fact that you have a big knowledge base
00:17:30.400 --> 00:17:34.880
that was curated by humans like Wiki
00:17:32.039 --> 00:17:36.120
data to improve the embeddings of a
00:17:34.880 --> 00:17:39.559
neural model
00:17:36.120 --> 00:17:41.799
itself and so another pretty large uh
00:17:39.559 --> 00:17:43.600
research area that a lot of people have
00:17:41.799 --> 00:17:47.120
focused on is how do you get good
00:17:43.600 --> 00:17:48.720
embeddings of a Knowledge Graph and this
00:17:47.120 --> 00:17:50.600
is important if you want to do any sort
00:17:48.720 --> 00:17:52.799
of like Knowledge Graph Search or other
00:17:50.600 --> 00:17:54.160
things like this like for example one of
00:17:52.799 --> 00:17:56.799
the really nice things about knowledge
00:17:54.160 --> 00:17:58.880
graphs is they have information about a
00:17:56.799 --> 00:18:00.200
whole bunch of really sparse entities
00:17:58.880 --> 00:18:03.240
that aren't mentioned very much on the
00:18:00.200 --> 00:18:05.679
internet for example and so because of
00:18:03.240 --> 00:18:07.440
that you can um you can leverage the
00:18:05.679 --> 00:18:10.720
knowledge graph structure together with
00:18:07.440 --> 00:18:10.720
text to learn better embeddings
00:18:11.240 --> 00:18:18.520
overall and so this particular paper is
00:18:15.280 --> 00:18:20.960
one example of it um and the way they do
00:18:18.520 --> 00:18:23.280
this is they express uh Knowledge Graph
00:18:20.960 --> 00:18:25.919
triples is additive
00:18:23.280 --> 00:18:28.480
Transformations and they minimize the
00:18:25.919 --> 00:18:31.640
distance uh of existing triples with a
00:18:28.480 --> 00:18:35.039
margin based loss so the way they do
00:18:31.640 --> 00:18:38.240
this is they have the head um in the
00:18:35.039 --> 00:18:40.799
tail and L is the vector corresponding
00:18:38.240 --> 00:18:42.679
to like the link between the things that
00:18:40.799 --> 00:18:47.960
corresponds to a
00:18:42.679 --> 00:18:52.159
relation and so you go uh you have H and
00:18:47.960 --> 00:18:53.559
T and here um like this is L but here
00:18:52.159 --> 00:18:55.640
it's written as are because I got this
00:18:53.559 --> 00:18:58.120
from a different paper and basically you
00:18:55.640 --> 00:18:59.480
you try to go from H to T um according
00:18:58.120 --> 00:19:00.919
to the relation
00:18:59.480 --> 00:19:05.120
uh Vector
00:19:00.919 --> 00:19:07.200
are and you use a hinge loss where um
00:19:05.120 --> 00:19:10.039
for the hinge loss you you have a hinge
00:19:07.200 --> 00:19:12.640
parameter and then you try to upweight
00:19:10.039 --> 00:19:15.760
the example of a true triple and
00:19:12.640 --> 00:19:17.960
downweight the example of a of a false
00:19:15.760 --> 00:19:19.880
triple so this could be one that was
00:19:17.960 --> 00:19:22.080
like randomly sampled to be incorrect
00:19:19.880 --> 00:19:22.080
for
00:19:23.760 --> 00:19:29.080
example um one interesting thing about
00:19:26.880 --> 00:19:31.559
knowledge graph embeddings is like a lot
00:19:29.080 --> 00:19:33.600
of famous AI researchers got their start
00:19:31.559 --> 00:19:36.000
in Knowledge Graph embeddings and so
00:19:33.600 --> 00:19:39.760
Richard soer is one of them if you know
00:19:36.000 --> 00:19:44.320
he's the CEO of vi.com search engine now
00:19:39.760 --> 00:19:46.679
um and uh this was a first attempt at
00:19:44.320 --> 00:19:49.679
predicting relations they basically
00:19:46.679 --> 00:19:55.400
created a um MLP that tries to predict
00:19:49.679 --> 00:19:58.880
whether a relation exists so they have
00:19:55.400 --> 00:20:00.760
a matrix for the left side of the
00:19:58.880 --> 00:20:03.320
relation a matrix for the right side of
00:20:00.760 --> 00:20:05.080
the relation and then they feed in the
00:20:03.320 --> 00:20:07.559
embeddings of each of the entities in
00:20:05.080 --> 00:20:08.919
the relation they have a nonlinearity
00:20:07.559 --> 00:20:11.799
and then they have another Vector that
00:20:08.919 --> 00:20:14.720
tries to predict the um the probability
00:20:11.799 --> 00:20:16.679
of the uh actual relation being correct
00:20:14.720 --> 00:20:18.960
so you would run this through a sigmoid
00:20:16.679 --> 00:20:21.000
and then uh if it was one the relation
00:20:18.960 --> 00:20:24.039
was likely to exist if it was Zero then
00:20:21.000 --> 00:20:25.480
the relation was likely to not exist and
00:20:24.039 --> 00:20:27.799
then they also propos something called a
00:20:25.480 --> 00:20:31.480
neural tensor Network and this adds a
00:20:27.799 --> 00:20:34.000
bilinear feature extractor um and so
00:20:31.480 --> 00:20:37.440
basically what this is saying is we have
00:20:34.000 --> 00:20:40.000
the embedding here the embedding here we
00:20:37.440 --> 00:20:41.840
have a matrix and then we calculate the
00:20:40.000 --> 00:20:43.080
dot product between the embedding after
00:20:41.840 --> 00:20:45.799
transformation it looks a lot like
00:20:43.080 --> 00:20:47.720
attention actually in a way um because
00:20:45.799 --> 00:20:50.000
we had the bilinear attention so it's
00:20:47.720 --> 00:20:53.640
similar to that as well and then we also
00:20:50.000 --> 00:20:56.840
have the MLP so this part corresponds to
00:20:53.640 --> 00:21:00.320
MLP and then we have a bias
00:20:56.840 --> 00:21:02.200
term and um this is a powerful model but
00:21:00.320 --> 00:21:05.400
it's a bit overparameterized so we
00:21:02.200 --> 00:21:08.120
actually later um uh this kind of fell
00:21:05.400 --> 00:21:10.360
out of uh favor towards these more
00:21:08.120 --> 00:21:14.520
simple models that we're using uh kind
00:21:10.360 --> 00:21:14.520
of just linear projections between the
00:21:17.600 --> 00:21:22.279
two so there's um there's a lot of
00:21:20.120 --> 00:21:25.320
methods like this these methods are
00:21:22.279 --> 00:21:27.039
basically assuming that we have either
00:21:25.320 --> 00:21:29.080
Knowledge Graph
00:21:27.039 --> 00:21:30.799
embeddings um and we want to learn
00:21:29.080 --> 00:21:32.480
relations or they're assuming that we
00:21:30.799 --> 00:21:34.320
don't have any information at all about
00:21:32.480 --> 00:21:36.840
the knowledge graph and we want to learn
00:21:34.320 --> 00:21:40.039
the knowledge graph embedding themselves
00:21:36.840 --> 00:21:42.400
it's been used for both of them but um I
00:21:40.039 --> 00:21:44.000
I'd say now it's probably most useful
00:21:42.400 --> 00:21:45.520
for learning Knowledge Graph embeddings
00:21:44.000 --> 00:21:50.480
if you want to do any sort of Knowledge
00:21:45.520 --> 00:21:50.480
Graph based modeling uh which can be
00:21:51.240 --> 00:21:55.919
useful um cool any questions about these
00:21:57.360 --> 00:22:01.679
ones okay
00:21:59.520 --> 00:22:04.360
next um actually this part might be a
00:22:01.679 --> 00:22:06.600
little bit simpler than the uh than the
00:22:04.360 --> 00:22:09.000
like knowledge graft based approaches so
00:22:06.600 --> 00:22:10.960
another method for relations extraction
00:22:09.000 --> 00:22:13.440
is learning from text
00:22:10.960 --> 00:22:16.120
directly
00:22:13.440 --> 00:22:19.080
and the first question about this is how
00:22:16.120 --> 00:22:22.200
do you get training data to learn uh
00:22:19.080 --> 00:22:24.480
about relation learn relation extraction
00:22:22.200 --> 00:22:26.720
and so there was this very influential
00:22:24.480 --> 00:22:28.279
paper a distant supervision for relation
00:22:26.720 --> 00:22:31.120
extraction I would say it's almost one
00:22:28.279 --> 00:22:32.880
of the first or certainly one of the
00:22:31.120 --> 00:22:34.559
most influential papers on like data
00:22:32.880 --> 00:22:35.960
augmentation or synthetic data for
00:22:34.559 --> 00:22:38.400
natural language
00:22:35.960 --> 00:22:40.440
processing and basically the idea is you
00:22:38.400 --> 00:22:44.279
already have a knowledge base that has
00:22:40.440 --> 00:22:47.440
some entries in it like Wiki data and so
00:22:44.279 --> 00:22:50.919
then given in entity relation entity
00:22:47.440 --> 00:22:52.919
triples um can you extract all text that
00:22:50.919 --> 00:22:54.799
matches this particular relation type
00:22:52.919 --> 00:22:56.480
and use it to train a relation extractor
00:22:54.799 --> 00:22:59.640
a supervised relation
00:22:56.480 --> 00:23:01.880
extractor so the way this works
00:22:59.640 --> 00:23:04.039
is like let's say we have this is an old
00:23:01.880 --> 00:23:06.120
paper so the examples are also old but
00:23:04.039 --> 00:23:08.039
um let's say we have Steven Spielberg
00:23:06.120 --> 00:23:10.159
being a director of the film Saving
00:23:08.039 --> 00:23:12.840
Private Ryan and that's included in our
00:23:10.159 --> 00:23:14.840
uh our knowledge base so what it would
00:23:12.840 --> 00:23:17.080
do is it would find all sentences that
00:23:14.840 --> 00:23:19.400
have Steven Spielberg and Saving Private
00:23:17.080 --> 00:23:22.080
Ryan included in them and it would label
00:23:19.400 --> 00:23:24.159
this as like a positive example of that
00:23:22.080 --> 00:23:28.240
relation so this
00:23:24.159 --> 00:23:30.760
is in general often it's okay it it
00:23:28.240 --> 00:23:34.480
works reasonably well but the problem
00:23:30.760 --> 00:23:37.200
with this is there are also um negative
00:23:34.480 --> 00:23:38.840
examples of this so like for example
00:23:37.200 --> 00:23:40.480
here I think the first one is kind of a
00:23:38.840 --> 00:23:43.240
negative example for the director
00:23:40.480 --> 00:23:45.880
relation because Steven Spielberg's film
00:23:43.240 --> 00:23:48.120
Saving Private Ryan doesn't actually
00:23:45.880 --> 00:23:50.000
tell you he's the director it just tells
00:23:48.120 --> 00:23:52.520
you that he's somehow affiliated with it
00:23:50.000 --> 00:23:54.840
he could be the writer or he could be uh
00:23:52.520 --> 00:23:57.679
the actor or or something else like that
00:23:54.840 --> 00:24:00.440
so this is a nice way to create data for
00:23:57.679 --> 00:24:03.640
basically free but at the same time uh
00:24:00.440 --> 00:24:06.159
you can like create noisy examples and
00:24:03.640 --> 00:24:06.159
that can be a
00:24:07.159 --> 00:24:14.600
problem so um there's been a lot of work
00:24:11.400 --> 00:24:16.000
about this um relationship uh relation
00:24:14.600 --> 00:24:17.840
classification with neural networks
00:24:16.000 --> 00:24:20.840
there's a lot of uh different methods
00:24:17.840 --> 00:24:23.159
that could be uh doing this most of them
00:24:20.840 --> 00:24:24.919
work by extracting features and then
00:24:23.159 --> 00:24:27.039
classifying somehow although there are
00:24:24.919 --> 00:24:29.960
some uh large language model based
00:24:27.039 --> 00:24:33.120
methods now um one one thing about
00:24:29.960 --> 00:24:35.440
relation extraction or not kind of like
00:24:33.120 --> 00:24:36.799
information extraction in general is
00:24:35.440 --> 00:24:38.559
that very often you want to run this
00:24:36.799 --> 00:24:40.200
over like a huge Corpus you want to run
00:24:38.559 --> 00:24:42.320
it over the whole internet or other
00:24:40.200 --> 00:24:45.000
things like that so from that point of
00:24:42.320 --> 00:24:47.159
view like I I said I could just ask
00:24:45.000 --> 00:24:49.480
mistol to give me the answer about like
00:24:47.159 --> 00:24:52.440
whether cars are included in sentences
00:24:49.480 --> 00:24:55.120
but if you want to run you know gp4 over
00:24:52.440 --> 00:24:56.799
the whole internet that's a pretty big
00:24:55.120 --> 00:25:00.159
budget and you might want to reconsider
00:24:56.799 --> 00:25:02.440
that so there are so um there is also
00:25:00.159 --> 00:25:04.440
some you know benefit in having cheap
00:25:02.440 --> 00:25:07.200
and lightweight
00:25:04.440 --> 00:25:09.159
methods so basically what this
00:25:07.200 --> 00:25:11.279
particular paper did is it extracted
00:25:09.159 --> 00:25:12.760
features in in classified so it
00:25:11.279 --> 00:25:15.600
extracted lexical features of the
00:25:12.760 --> 00:25:20.240
entities themselves and features of the
00:25:15.600 --> 00:25:22.360
whole span and so like the way I uh most
00:25:20.240 --> 00:25:26.960
modern methods for this do this is they
00:25:22.360 --> 00:25:29.399
basically um extract features from the
00:25:26.960 --> 00:25:31.679
first part of the first entity the
00:25:29.399 --> 00:25:33.760
second part of the the first entity the
00:25:31.679 --> 00:25:36.360
first part of the second entity and the
00:25:33.760 --> 00:25:37.720
last part of the uh second entity and
00:25:36.360 --> 00:25:39.600
take all of those embeddings feed them
00:25:37.720 --> 00:25:41.440
into like an MLP or something like that
00:25:39.600 --> 00:25:44.039
and then make a prediction about whether
00:25:41.440 --> 00:25:45.760
that relation exists so if you have an
00:25:44.039 --> 00:25:47.840
embedding model this is relatively easy
00:25:45.760 --> 00:25:50.360
to do you feed it through like uh
00:25:47.840 --> 00:25:51.919
Roberta or you feed it through mistol
00:25:50.360 --> 00:25:54.559
and get the embeddings for each of the
00:25:51.919 --> 00:25:55.840
tokens and um and then you make a
00:25:54.559 --> 00:25:58.840
prediction based on those four
00:25:55.840 --> 00:25:58.840
embeddings
00:26:00.600 --> 00:26:04.840
um the details of that are like not
00:26:03.520 --> 00:26:07.320
super important unless you're going to
00:26:04.840 --> 00:26:09.279
go in and implement it yourself so you
00:26:07.320 --> 00:26:10.919
can um like if you're actually going to
00:26:09.279 --> 00:26:12.120
be doing relation extraction obviously
00:26:10.919 --> 00:26:14.279
the details are important but I'm
00:26:12.120 --> 00:26:16.000
assuming that most people won't be uh
00:26:14.279 --> 00:26:19.720
you know doing that as your final
00:26:16.000 --> 00:26:21.240
project but um one really interesting
00:26:19.720 --> 00:26:22.919
thing that is relevant even if you're
00:26:21.240 --> 00:26:26.360
not doing relationship relation
00:26:22.919 --> 00:26:29.360
extraction is how you can model noise
00:26:26.360 --> 00:26:32.600
because this um as I said they're
00:26:29.360 --> 00:26:35.720
creating lots of like semi noisy data
00:26:32.600 --> 00:26:38.919
and a lot of the work in getting good
00:26:35.720 --> 00:26:40.360
bottles for relation extraction has been
00:26:38.919 --> 00:26:41.799
how do we deal with this distant
00:26:40.360 --> 00:26:43.799
supervision noise and I'm just going to
00:26:41.799 --> 00:26:45.760
give one example here but there's like a
00:26:43.799 --> 00:26:49.120
series of papers after this that also
00:26:45.760 --> 00:26:50.600
tried to do similar things so the idea
00:26:49.120 --> 00:26:53.600
is that there's noise in the distant
00:26:50.600 --> 00:26:56.559
supervision labels um and so we want to
00:26:53.600 --> 00:27:01.039
model and mitigate that noise and the
00:26:56.559 --> 00:27:03.919
way this paper does this is they have an
00:27:01.039 --> 00:27:06.679
encoder and from the encoder you
00:27:03.919 --> 00:27:10.960
calculate embeddings and make
00:27:06.679 --> 00:27:14.279
predictions and so you have a small set
00:27:10.960 --> 00:27:16.080
of like very high quality data and this
00:27:14.279 --> 00:27:17.760
small set of very high quality data you
00:27:16.080 --> 00:27:19.880
can basically trust that all of the data
00:27:17.760 --> 00:27:22.320
is not noisy like maybe it's manually
00:27:19.880 --> 00:27:23.720
annotated data and you have like 5,000
00:27:22.320 --> 00:27:25.000
examples of it or something like that
00:27:23.720 --> 00:27:26.880
and then separately from that you have
00:27:25.000 --> 00:27:28.440
like 5 million examples of automatically
00:27:26.880 --> 00:27:30.799
labeled data that might be good might
00:27:28.440 --> 00:27:32.679
not be good and so what they do is
00:27:30.799 --> 00:27:34.200
essentially at the beginning they take
00:27:32.679 --> 00:27:36.520
this encoder get embeddings make
00:27:34.200 --> 00:27:38.000
predictions over the high quality data
00:27:36.520 --> 00:27:40.320
and then they have a separate noise
00:27:38.000 --> 00:27:43.440
modeling layer where what this noise
00:27:40.320 --> 00:27:46.919
modeling layer does is it has a
00:27:43.440 --> 00:27:50.039
transition Matrix which says given that
00:27:46.919 --> 00:27:53.279
this given that we made a particular
00:27:50.039 --> 00:27:55.159
prediction over classes because this is
00:27:53.279 --> 00:27:59.919
essentially a multiclass classification
00:27:55.159 --> 00:28:01.519
problem they transform the
00:27:59.919 --> 00:28:03.159
sorry I don't remember if they transform
00:28:01.519 --> 00:28:04.640
the probabilities or the low Jets I
00:28:03.159 --> 00:28:07.320
think it's the probabilities but they
00:28:04.640 --> 00:28:12.799
transform the probabilities and get a
00:28:07.320 --> 00:28:14.720
final uh distribution after noise and so
00:28:12.799 --> 00:28:17.399
that means that you can basically smooth
00:28:14.720 --> 00:28:19.240
out this uh distribution and account for
00:28:17.399 --> 00:28:20.880
the fact that the labels may be noisy or
00:28:19.240 --> 00:28:24.399
may may not be
00:28:20.880 --> 00:28:26.600
noisy um then they add additional
00:28:24.399 --> 00:28:28.559
normalization on this transition Matrix
00:28:26.600 --> 00:28:32.440
using something called Trace normal
00:28:28.559 --> 00:28:35.840
ization to move this Matrix closer to
00:28:32.440 --> 00:28:38.480
the identity function which says that
00:28:35.840 --> 00:28:40.720
the predictions are probably not wrong
00:28:38.480 --> 00:28:43.159
all the time uh the predictions are
00:28:40.720 --> 00:28:45.360
probably correct you know a lot of the
00:28:43.159 --> 00:28:46.600
time they're not correct all the time uh
00:28:45.360 --> 00:28:49.720
so then you have that Trace
00:28:46.600 --> 00:28:51.880
normalization competing with um this uh
00:28:49.720 --> 00:28:55.440
trying to give you like a more smooth
00:28:51.880 --> 00:28:58.760
distribution and and reduce your uh L
00:28:55.440 --> 00:29:00.320
like reduce your loss so um I I think
00:28:58.760 --> 00:29:02.559
this is actually a pretty interesting
00:29:00.320 --> 00:29:04.480
idea and it can be used not just for
00:29:02.559 --> 00:29:08.600
relation extraction but also in cases
00:29:04.480 --> 00:29:08.600
where um you might have noisy labels
00:29:08.799 --> 00:29:14.320
overall um so are there any questions
00:29:12.360 --> 00:29:15.720
about this or any of the things that are
00:29:14.320 --> 00:29:18.480
going on
00:29:15.720 --> 00:29:20.279
here um even if you're completely
00:29:18.480 --> 00:29:21.960
uninterested in relation extraction I'd
00:29:20.279 --> 00:29:23.720
encourage you to think about like what
00:29:21.960 --> 00:29:26.159
are
00:29:23.720 --> 00:29:27.360
some examples of things that you are
00:29:26.159 --> 00:29:29.519
interested in where you could get
00:29:27.360 --> 00:29:31.840
potentially labels and how could you for
00:29:29.519 --> 00:29:34.880
theise there like that might be uh you
00:29:31.840 --> 00:29:34.880
know a thing to
00:29:35.679 --> 00:29:39.919
about okay so this was a very very brief
00:29:38.320 --> 00:29:42.679
overview of how we create knowledge
00:29:39.919 --> 00:29:44.080
bases uh from textual data or from
00:29:42.679 --> 00:29:47.159
Knowledge Graph data structured
00:29:44.080 --> 00:29:48.840
Knowledge Graph data um so now I like to
00:29:47.159 --> 00:29:51.519
talk a little bit about how to use
00:29:48.840 --> 00:29:53.960
knowledge bases to inform neural
00:29:51.519 --> 00:29:56.159
models and there's a bunch of different
00:29:53.960 --> 00:29:59.519
ways to do this
00:29:56.159 --> 00:30:02.600
um the
00:29:59.519 --> 00:30:06.960
the first way um is to
00:30:02.600 --> 00:30:09.840
improve embeddings uh
00:30:06.960 --> 00:30:11.960
with existing lexicons and this example
00:30:09.840 --> 00:30:14.679
is using non-contextual embeddings like
00:30:11.960 --> 00:30:16.240
not the not the ones we get from neural
00:30:14.679 --> 00:30:17.919
language models but once we get from
00:30:16.240 --> 00:30:20.919
just running a embedding model like word
00:30:17.919 --> 00:30:22.960
toac or something like this um and what
00:30:20.919 --> 00:30:25.640
they did in this paper is they
00:30:22.960 --> 00:30:27.600
essentially um retrofitted embeddings to
00:30:25.640 --> 00:30:30.840
existing lexicons by doing post Hawk
00:30:27.600 --> 00:30:34.080
trans of the embeddings so that they
00:30:30.840 --> 00:30:36.840
matched the um the knowledge graph for
00:30:34.080 --> 00:30:39.080
lexon better and so the way they did
00:30:36.840 --> 00:30:41.880
this is
00:30:39.080 --> 00:30:43.720
um they started out with pre-trained
00:30:41.880 --> 00:30:45.399
embeddings and they had a double
00:30:43.720 --> 00:30:47.240
objective of making the transform
00:30:45.399 --> 00:30:49.120
embeddings close to the neighbors and
00:30:47.240 --> 00:30:52.519
close to the original
00:30:49.120 --> 00:30:58.840
embedding and the way they did this is
00:30:52.519 --> 00:30:58.840
they essentially had um this
00:30:59.799 --> 00:31:03.720
this regularization term over here so
00:31:01.880 --> 00:31:06.200
this regularization term is basically
00:31:03.720 --> 00:31:08.279
saying um I don't want you to move your
00:31:06.200 --> 00:31:09.360
embeddings too far away from how they
00:31:08.279 --> 00:31:11.679
were
00:31:09.360 --> 00:31:14.799
initialized and then at the same time I
00:31:11.679 --> 00:31:17.279
would like you to make these uh
00:31:14.799 --> 00:31:19.600
embeddings closer to each other if they
00:31:17.279 --> 00:31:21.240
are synonyms of each other so they did
00:31:19.600 --> 00:31:23.600
this using word net and they basically
00:31:21.240 --> 00:31:26.200
took the words uh that were synonyms to
00:31:23.600 --> 00:31:28.679
each other in sinets with each other and
00:31:26.200 --> 00:31:30.000
they tried to regularize the synonyms to
00:31:28.679 --> 00:31:32.120
be closer together but also the
00:31:30.000 --> 00:31:33.639
embeddings to be closer to how they
00:31:32.120 --> 00:31:35.960
started
00:31:33.639 --> 00:31:38.799
out and there were also examples of
00:31:35.960 --> 00:31:40.720
forcing anms away from each other so
00:31:38.799 --> 00:31:42.480
like if you're um this is a little bit
00:31:40.720 --> 00:31:44.799
of an older work so it was working on
00:31:42.480 --> 00:31:47.600
non-contextualized embeddings but we
00:31:44.799 --> 00:31:49.399
could do something very similar for um
00:31:47.600 --> 00:31:52.000
more modern models in like Knowledge
00:31:49.399 --> 00:31:55.320
Graph embeddings for example so let's
00:31:52.000 --> 00:31:58.960
say we had
00:31:55.320 --> 00:32:03.240
um a model that ident
00:31:58.960 --> 00:32:06.600
entities and then different examples of
00:32:03.240 --> 00:32:06.600
those entities across different
00:32:07.159 --> 00:32:11.480
contexts um let's go back to the wiki
00:32:20.639 --> 00:32:26.840
data and so um if we had lots of
00:32:23.960 --> 00:32:29.360
examples of Joe Biden um Joe Biden is
00:32:26.840 --> 00:32:35.159
referred to in a number ways like Joe
00:32:29.360 --> 00:32:44.440
Biden Joseph Biden Joseph R Biden um J
00:32:35.159 --> 00:32:47.880
jrb I guess um pus 48 46 sorry um and uh
00:32:44.440 --> 00:32:50.799
so you could find different examples of
00:32:47.880 --> 00:32:52.799
things that match these strings um and
00:32:50.799 --> 00:32:55.360
even do entity linking uh which I'll
00:32:52.799 --> 00:32:57.200
I'll talk about in a little bit and then
00:32:55.360 --> 00:32:58.760
encourag the embeddings for all of these
00:32:57.200 --> 00:33:01.360
different instances is to be closer
00:32:58.760 --> 00:33:04.039
together to make your model like disting
00:33:01.360 --> 00:33:06.799
uh distinguish them less and Ure that
00:33:04.039 --> 00:33:08.399
they uh they get closer edings and that
00:33:06.799 --> 00:33:11.639
could improve like question answering
00:33:08.399 --> 00:33:11.639
look up other stuff like
00:33:12.960 --> 00:33:19.880
that
00:33:14.919 --> 00:33:23.399
cool um yeah I have a question about
00:33:19.880 --> 00:33:25.399
this so what happens if you do like subw
00:33:23.399 --> 00:33:28.000
modeling and then you don't have like
00:33:25.399 --> 00:33:30.440
the embedment for that entire string
00:33:28.000 --> 00:33:32.320
that is supposed to be Clos yeah what
00:33:30.440 --> 00:33:34.279
happens if you do subword modeling and
00:33:32.320 --> 00:33:35.480
you don't have the embedding uh you
00:33:34.279 --> 00:33:37.159
don't have a single embedding that
00:33:35.480 --> 00:33:40.360
corresponds to an entity so that's a
00:33:37.159 --> 00:33:42.559
really good question um let me
00:33:40.360 --> 00:33:44.240
check I don't think I actually have
00:33:42.559 --> 00:33:46.600
these on the slide so I might have to
00:33:44.240 --> 00:33:46.600
open a
00:33:53.639 --> 00:33:59.720
paper yeah okay so there's a lot of
00:33:56.440 --> 00:33:59.720
different ways to handle this
00:34:11.520 --> 00:34:18.079
so there there's two papers um the first
00:34:14.879 --> 00:34:20.000
paper is uh a really nice paper very
00:34:18.079 --> 00:34:22.359
influential on the subject of
00:34:20.000 --> 00:34:25.359
co-reference resolution and co-reference
00:34:22.359 --> 00:34:27.240
resolution um is essentially trying to
00:34:25.359 --> 00:34:30.000
identify when two spans correspond to
00:34:27.240 --> 00:34:32.320
each other so like if I say Joe B Joe
00:34:30.000 --> 00:34:34.359
Biden early in a document and then later
00:34:32.320 --> 00:34:35.480
in a document it just says Biden we want
00:34:34.359 --> 00:34:38.839
to know that those two things are
00:34:35.480 --> 00:34:40.919
referring to each other and then um we
00:34:38.839 --> 00:34:42.839
had a paper later where we generalized
00:34:40.919 --> 00:34:44.839
this and applied you know very similar
00:34:42.839 --> 00:34:48.079
methodology to like lots and lots of
00:34:44.839 --> 00:34:50.760
different analysis tasks but I can um I
00:34:48.079 --> 00:34:53.839
can show the beginning here and
00:34:50.760 --> 00:34:59.320
basically the methodology that they use
00:34:53.839 --> 00:35:02.440
here um is they add
00:34:59.320 --> 00:35:04.440
a and this is specifically for modeling
00:35:02.440 --> 00:35:08.240
spans and getting embeddings out of
00:35:04.440 --> 00:35:09.040
spans of uh tokens and what they did is
00:35:08.240 --> 00:35:13.079
they
00:35:09.040 --> 00:35:14.920
essentially have a model where you take
00:35:13.079 --> 00:35:16.440
the thing from the beginning the
00:35:14.920 --> 00:35:18.760
embedding from the beginning of the span
00:35:16.440 --> 00:35:22.040
the embedding from the end of the span
00:35:18.760 --> 00:35:24.280
and the average embedding of all of the
00:35:22.040 --> 00:35:26.280
embeddings in the span and that gives
00:35:24.280 --> 00:35:27.480
you three vectors for any span right
00:35:26.280 --> 00:35:30.160
because you can always get the beginning
00:35:27.480 --> 00:35:33.280
that and in the mean and then based on
00:35:30.160 --> 00:35:36.560
that they feed that through um like a
00:35:33.280 --> 00:35:37.800
neural network and get a new edting so
00:35:36.560 --> 00:35:40.000
they feed that through a transformation
00:35:37.800 --> 00:35:42.520
and get a new edting and so that's the
00:35:40.000 --> 00:35:44.200
method that they used and I think our
00:35:42.520 --> 00:35:46.640
paper actually has a
00:35:44.200 --> 00:35:49.640
better
00:35:46.640 --> 00:35:52.640
um a better figure of how you can
00:35:49.640 --> 00:35:56.680
actually use that actually maybe it
00:35:52.640 --> 00:35:58.160
doesn't okay but anyway um yeah because
00:35:56.680 --> 00:36:00.240
uh yeah here's the figure
00:35:58.160 --> 00:36:01.520
so then you can use that for a number of
00:36:00.240 --> 00:36:03.040
things you could use that to like look
00:36:01.520 --> 00:36:06.359
up something in a knowledge base you
00:36:03.040 --> 00:36:08.599
could also use that to um decide whether
00:36:06.359 --> 00:36:10.440
two spans are co-referent by feeding in
00:36:08.599 --> 00:36:12.800
like the first span and the second Span
00:36:10.440 --> 00:36:14.960
in and then predicting whether those two
00:36:12.800 --> 00:36:19.640
spans cor correspond to each other or
00:36:14.960 --> 00:36:21.240
not so this general idea of modeling
00:36:19.640 --> 00:36:22.960
spans and then modeling relations
00:36:21.240 --> 00:36:24.520
between the spans allows you to solve
00:36:22.960 --> 00:36:26.119
like lots of different tasks like part
00:36:24.520 --> 00:36:27.920
of speech tagging or named entity
00:36:26.119 --> 00:36:30.319
recognition or relation extraction or
00:36:27.920 --> 00:36:31.920
other stuff like that so um yeah
00:36:30.319 --> 00:36:34.040
actually I realized now that I should
00:36:31.920 --> 00:36:35.079
have probably talked about these in the
00:36:34.040 --> 00:36:36.560
slides where I was talking about
00:36:35.079 --> 00:36:38.599
modeling but that that would be my
00:36:36.560 --> 00:36:42.319
recommended way of doing
00:36:38.599 --> 00:36:42.319
it cool any other
00:36:43.839 --> 00:36:49.480
questions nice okay
00:36:46.880 --> 00:36:52.880
um
00:36:49.480 --> 00:36:55.119
so another question is how can we inject
00:36:52.880 --> 00:36:56.640
knowledge into language models um
00:36:55.119 --> 00:36:58.720
there's a bunch of different ways to do
00:36:56.640 --> 00:37:03.079
this um
00:36:58.720 --> 00:37:05.000
one very easy way is to somehow look up
00:37:03.079 --> 00:37:09.640
relevant knowledge in your knowledge
00:37:05.000 --> 00:37:09.640
graph and um oh
00:37:10.280 --> 00:37:15.440
sorry I was presenting on my own screen
00:37:13.040 --> 00:37:18.240
not the screen that everybody can see so
00:37:15.440 --> 00:37:22.000
um to look up all of the uh knowledge in
00:37:18.240 --> 00:37:24.000
a Knowledge Graph and um somehow provide
00:37:22.000 --> 00:37:26.800
it to the model one way you can provide
00:37:24.000 --> 00:37:28.720
it to the model is through prompting um
00:37:26.800 --> 00:37:32.400
but the problem with with prompting is
00:37:28.720 --> 00:37:33.920
that you're not necessarily going to uh
00:37:32.400 --> 00:37:37.319
be able
00:37:33.920 --> 00:37:41.359
to utilize knowledge that is kind of
00:37:37.319 --> 00:37:43.920
like minority knowledge because the
00:37:41.359 --> 00:37:47.560
embeddings of the entities that you're
00:37:43.920 --> 00:37:49.440
presenting may not be you know like well
00:37:47.560 --> 00:37:51.839
learned so
00:37:49.440 --> 00:37:53.200
you're requiring essentially the model
00:37:51.839 --> 00:37:55.359
to be able to generalize from the
00:37:53.200 --> 00:37:57.880
knowledge you provide in
00:37:55.359 --> 00:38:00.839
the prompt despite the fact that the
00:37:57.880 --> 00:38:02.240
prompt is like minor entities or other
00:38:00.839 --> 00:38:07.040
things like that that are not as well
00:38:02.240 --> 00:38:10.400
learned so is another um method to
00:38:07.040 --> 00:38:13.440
handle this um we previously proposed a
00:38:10.400 --> 00:38:15.599
method that allows you
00:38:13.440 --> 00:38:18.319
to essentially
00:38:15.599 --> 00:38:21.319
predict instead of predicting directly
00:38:18.319 --> 00:38:24.920
the words here you can predict a tag
00:38:21.319 --> 00:38:27.200
that says birth name or a given name or
00:38:24.920 --> 00:38:31.480
family name or something like that and
00:38:27.200 --> 00:38:32.839
then post talk the model will fill in uh
00:38:31.480 --> 00:38:36.720
that like birth
00:38:32.839 --> 00:38:39.400
name text based on a knowledge base so
00:38:36.720 --> 00:38:41.079
um you know if you have a a Wikipedia
00:38:39.400 --> 00:38:44.240
article about Barack Obama that you're
00:38:41.079 --> 00:38:48.680
trying to write it could predict um
00:38:44.240 --> 00:38:52.040
birth name born uh birth name comma born
00:38:48.680 --> 00:38:55.359
in birth date and that's like a very
00:38:52.040 --> 00:38:56.880
very common thing in Wikipedia right so
00:38:55.359 --> 00:39:00.960
because of that it can predict it very
00:38:56.880 --> 00:39:03.160
consistently very uh formulaically and
00:39:00.960 --> 00:39:04.599
that allows you to um you know with high
00:39:03.160 --> 00:39:06.079
confidence get something that makes
00:39:04.599 --> 00:39:08.599
sense and is factual and reduce
00:39:06.079 --> 00:39:11.400
hallucination and other stuff like that
00:39:08.599 --> 00:39:12.599
so um basically how could you inject
00:39:11.400 --> 00:39:14.280
this into language models there's
00:39:12.599 --> 00:39:16.240
multiple ways one is prompting that's
00:39:14.280 --> 00:39:18.160
maybe the easier way another way is
00:39:16.240 --> 00:39:21.520
through like templatic generation like
00:39:18.160 --> 00:39:23.200
this where you generate placeholders uh
00:39:21.520 --> 00:39:25.200
for all the information you want to add
00:39:23.200 --> 00:39:26.480
and then you add the information uh
00:39:25.200 --> 00:39:29.359
directly from the knowledge base through
00:39:26.480 --> 00:39:29.359
the placeholders like
00:39:30.680 --> 00:39:36.800
cool um there there's details about this
00:39:34.240 --> 00:39:38.920
in the paper like how we um formulate a
00:39:36.800 --> 00:39:41.319
training objective for something like
00:39:38.920 --> 00:39:43.480
this and the difficulty in formulating a
00:39:41.319 --> 00:39:46.400
training objective is that you need to
00:39:43.480 --> 00:39:48.280
figure out when you want to replace
00:39:46.400 --> 00:39:49.720
things so like you might not always want
00:39:48.280 --> 00:39:51.000
to replace with birth name you might
00:39:49.720 --> 00:39:53.920
want to replace with given name and
00:39:51.000 --> 00:39:55.839
family name and we demonstrate that you
00:39:53.920 --> 00:39:58.400
can figure out how to do this by
00:39:55.839 --> 00:40:00.960
essentially like Mar iing over the
00:39:58.400 --> 00:40:03.520
various ways of uh of doing this but
00:40:00.960 --> 00:40:05.880
that's kind of more complex detail
00:40:03.520 --> 00:40:05.880
that's in the
00:40:08.440 --> 00:40:15.480
paper another really interesting
00:40:11.000 --> 00:40:17.319
question um that uh we this is a also a
00:40:15.480 --> 00:40:19.440
paper that I was involved in from uh
00:40:17.319 --> 00:40:22.040
four years ago but I feel like this is
00:40:19.440 --> 00:40:25.040
not entirely solved even in like modern
00:40:22.040 --> 00:40:26.920
rag systems uh today is how can we
00:40:25.040 --> 00:40:28.880
reason over a lot of text that's
00:40:26.920 --> 00:40:32.440
included in a knowledge
00:40:28.880 --> 00:40:35.839
base um oh sorry reason over Text corpus
00:40:32.440 --> 00:40:40.480
like we reason over knowledge bases
00:40:35.839 --> 00:40:43.280
and basically uh what we did was we
00:40:40.480 --> 00:40:44.960
answered questions using text corpora as
00:40:43.280 --> 00:40:48.680
a traceable knowledge
00:40:44.960 --> 00:40:52.800
bases and we did relevance matching over
00:40:48.680 --> 00:40:54.920
mentions um and the way we did this is
00:40:52.800 --> 00:40:57.440
we created mentioned
00:40:54.920 --> 00:40:59.480
vectors and the mentioned vectors
00:40:57.440 --> 00:41:01.720
vectors of all of the mentions in the
00:40:59.480 --> 00:41:04.920
knowledge base of particular
00:41:01.720 --> 00:41:05.920
entities um and then we retrieved
00:41:04.920 --> 00:41:09.599
relevant
00:41:05.920 --> 00:41:13.440
mentions um from pre-trained Models uh
00:41:09.599 --> 00:41:15.040
so we we ran embeddings and generated uh
00:41:13.440 --> 00:41:16.000
embeddings for each of the mentions in
00:41:15.040 --> 00:41:20.440
the whole
00:41:16.000 --> 00:41:25.440
Corpus and based on this let let
00:41:20.440 --> 00:41:29.119
me find the place over here so based on
00:41:25.440 --> 00:41:32.720
this we basically um encoded all of
00:41:29.119 --> 00:41:35.040
these uh in here and then we had a dense
00:41:32.720 --> 00:41:37.359
query vector and the dense query Vector
00:41:35.040 --> 00:41:41.640
was specifically trained so that it
00:41:37.359 --> 00:41:44.280
would be able to identify entity
00:41:41.640 --> 00:41:46.760
mentions that answered the problem so if
00:41:44.280 --> 00:41:50.240
we had like when was The Grateful Dead
00:41:46.760 --> 00:41:52.520
and uh Bob Dylan album released uh we
00:41:50.240 --> 00:41:54.760
would have Bob Dylan be one vector The
00:41:52.520 --> 00:41:56.560
Grateful Dead be another vector and the
00:41:54.760 --> 00:41:58.200
model would be specifically trained so
00:41:56.560 --> 00:42:00.040
that when you took took the entity
00:41:58.200 --> 00:42:03.319
embedding of this and matched it with an
00:42:00.040 --> 00:42:05.400
entity embedding in this big Corpus of
00:42:03.319 --> 00:42:07.920
encoded things here it would be most
00:42:05.400 --> 00:42:10.400
likely to return relevant information to
00:42:07.920 --> 00:42:13.160
answer these like entity relation
00:42:10.400 --> 00:42:14.680
questions so then the question is how do
00:42:13.160 --> 00:42:18.040
we train a model like this how do we
00:42:14.680 --> 00:42:20.280
train like a dense uh embedding model so
00:42:18.040 --> 00:42:21.520
that it gets relevant information for
00:42:20.280 --> 00:42:23.800
answering
00:42:21.520 --> 00:42:26.920
questions and basically the way we did
00:42:23.800 --> 00:42:29.280
this was through week supervision uh
00:42:26.920 --> 00:42:31.640
just like I talked about for relation
00:42:29.280 --> 00:42:33.599
extraction in relation extraction we can
00:42:31.640 --> 00:42:35.680
create weak supervision by taking a big
00:42:33.599 --> 00:42:37.960
existing knowledge base and identifying
00:42:35.680 --> 00:42:40.920
all of the sentences where the answer is
00:42:37.960 --> 00:42:43.319
included and so what we did is we took
00:42:40.920 --> 00:42:45.880
this big existing knowledge base and
00:42:43.319 --> 00:42:47.920
said okay what are some of the relations
00:42:45.880 --> 00:42:49.800
in the knowledge base one example of a
00:42:47.920 --> 00:42:51.559
relation in the knowledge base is Steven
00:42:49.800 --> 00:42:54.359
Spielberg is the director of Saving
00:42:51.559 --> 00:42:57.319
Private Ryan so we created questions
00:42:54.359 --> 00:42:59.119
that said um
00:42:57.319 --> 00:43:01.079
was the director of Saving Private Ryan
00:42:59.119 --> 00:43:03.920
we can create those with templates uh
00:43:01.079 --> 00:43:06.359
easily for many different relations and
00:43:03.920 --> 00:43:09.480
then we took the embedding for Saving
00:43:06.359 --> 00:43:10.760
Private Ryan in that question and we
00:43:09.480 --> 00:43:14.200
tried to
00:43:10.760 --> 00:43:17.119
upweight all of the Saving Private Ryan
00:43:14.200 --> 00:43:19.680
embeddings over all of Wikipedia where
00:43:17.119 --> 00:43:23.160
Steven Spielberg cooccurred in that
00:43:19.680 --> 00:43:25.640
sentence so that tries to match um you
00:43:23.160 --> 00:43:27.079
know artificially created questions with
00:43:25.640 --> 00:43:29.040
sentences that would be the answer
00:43:27.079 --> 00:43:31.040
answer to that question and so that
00:43:29.040 --> 00:43:32.480
gives you like supervision it gives you
00:43:31.040 --> 00:43:35.079
a lot of data to train over it gives you
00:43:32.480 --> 00:43:38.920
a good model so that that allowed us to
00:43:35.079 --> 00:43:41.319
learn this model well so um this is one
00:43:38.920 --> 00:43:43.160
example of how you can do like rag spe
00:43:41.319 --> 00:43:46.200
specifically like informed by knowledge
00:43:43.160 --> 00:43:46.200
bases and stuff like
00:43:47.280 --> 00:43:52.160
that um any any questions about this
00:43:53.480 --> 00:43:57.680
or
00:43:55.079 --> 00:44:00.079
okay so another thing that I I'd like to
00:43:57.680 --> 00:44:03.599
go into is uh something we call schema
00:44:00.079 --> 00:44:06.240
free extraction and so if I go back to
00:44:03.599 --> 00:44:09.960
the wiki Data
00:44:06.240 --> 00:44:10.760
Page um Wiki data has something we call
00:44:09.960 --> 00:44:13.599
a
00:44:10.760 --> 00:44:16.880
schema and the schema is basically like
00:44:13.599 --> 00:44:19.640
what are the relations that are included
00:44:16.880 --> 00:44:21.000
in the database so one of the relations
00:44:19.640 --> 00:44:25.079
that's included in the databas is
00:44:21.000 --> 00:44:25.079
instance of I guess also
00:44:25.200 --> 00:44:29.040
image lots of images
00:44:29.079 --> 00:44:33.880
um
00:44:30.440 --> 00:44:35.680
signature uh sex or gender country of
00:44:33.880 --> 00:44:38.319
citizenship and these relations are like
00:44:35.680 --> 00:44:41.079
decided a priori by the people who
00:44:38.319 --> 00:44:43.200
created Wiki data um and there's lots
00:44:41.079 --> 00:44:45.880
and lots of them but that doesn't
00:44:43.200 --> 00:44:48.880
necessarily mean
00:44:45.880 --> 00:44:50.400
that like similarly to the problem of
00:44:48.880 --> 00:44:51.839
not having all of the entities we can't
00:44:50.400 --> 00:44:55.119
have all of the relations and just to
00:44:51.839 --> 00:44:57.280
give one example I was um in preparation
00:44:55.119 --> 00:44:59.680
for our large language models lecture I
00:44:57.280 --> 00:45:02.640
actually created some structured data
00:44:59.680 --> 00:45:04.319
about large language models and some of
00:45:02.640 --> 00:45:06.119
the instru the structured data about
00:45:04.319 --> 00:45:09.319
large language models that I created was
00:45:06.119 --> 00:45:11.440
like what is the variety of positional
00:45:09.319 --> 00:45:13.079
embedding that they're using or
00:45:11.440 --> 00:45:15.800
positional embedding variety and
00:45:13.079 --> 00:45:18.720
positional embedding variety is not in
00:45:15.800 --> 00:45:20.359
Wiki data I think um I'd be surprised if
00:45:18.720 --> 00:45:23.200
it was in Wiki data but I think it's not
00:45:20.359 --> 00:45:25.760
in Wiki data um so like as you go down
00:45:23.200 --> 00:45:27.760
to like more esoteric Concepts or like
00:45:25.760 --> 00:45:29.599
specialized domains or stuff like that
00:45:27.760 --> 00:45:31.359
you're almost always guaranteed to not
00:45:29.599 --> 00:45:34.040
you know have all the entities you need
00:45:31.359 --> 00:45:36.680
or not have all the relations you need
00:45:34.040 --> 00:45:38.160
so that's the problem that schema free
00:45:36.680 --> 00:45:39.920
extraction is trying to solve it's
00:45:38.160 --> 00:45:41.680
trying to figure out how we can like
00:45:39.920 --> 00:45:45.920
jointly figure out the schema together
00:45:41.680 --> 00:45:45.920
with uh the information you want to
00:45:48.480 --> 00:45:54.040
extract and the um the most famous
00:45:52.319 --> 00:45:55.599
example of this is something called open
00:45:54.040 --> 00:45:57.200
information extraction in open
00:45:55.599 --> 00:46:01.160
information extraction basically what
00:45:57.200 --> 00:46:04.040
it's saying is um we don't need a schema
00:46:01.160 --> 00:46:06.359
uh there's no there's no schema um the
00:46:04.040 --> 00:46:08.720
only schema that we have is the actual
00:46:06.359 --> 00:46:12.200
text in the sentences that we're
00:46:08.720 --> 00:46:14.520
referring to um the entities so if we
00:46:12.200 --> 00:46:16.040
have United United has a Hub in Chicago
00:46:14.520 --> 00:46:17.359
which is the headquarters of United
00:46:16.040 --> 00:46:21.200
Continental
00:46:17.359 --> 00:46:25.880
Holdings um the relation is literally
00:46:21.200 --> 00:46:29.359
has a Hub in um that that's the relation
00:46:25.880 --> 00:46:33.359
um and then for this we have Chicago is
00:46:29.359 --> 00:46:35.559
the headquarters of um but the problem
00:46:33.359 --> 00:46:37.520
with this uh is that this cannot
00:46:35.559 --> 00:46:40.359
abstract away so if we had another
00:46:37.520 --> 00:46:42.000
sentence that said Chicago or United
00:46:40.359 --> 00:46:44.319
Continental Holdings has its
00:46:42.000 --> 00:46:45.720
headquarters in Chicago that would be
00:46:44.319 --> 00:46:49.800
treated as completely different you
00:46:45.720 --> 00:46:49.800
wouldn't be able to like group those two
00:46:51.119 --> 00:46:57.720
together so um in open information
00:46:55.000 --> 00:47:00.079
extraction actually a lot of the methods
00:46:57.720 --> 00:47:02.800
this is one of the few things where
00:47:00.079 --> 00:47:05.480
people still use rule-based systems as
00:47:02.800 --> 00:47:07.640
kind of like uh you know almost
00:47:05.480 --> 00:47:09.319
state-of-the-art systems but basically
00:47:07.640 --> 00:47:11.559
the reason why you're able to do this is
00:47:09.319 --> 00:47:14.440
it's not actually that hard to extract
00:47:11.559 --> 00:47:16.839
kind of the relevant strings between uh
00:47:14.440 --> 00:47:19.599
two entities and so the both the
00:47:16.839 --> 00:47:21.359
Precision and recall are pretty high and
00:47:19.599 --> 00:47:24.079
another reason why people use rule-based
00:47:21.359 --> 00:47:25.760
systems is because they um like you want
00:47:24.079 --> 00:47:27.440
to run it over the whole web and running
00:47:25.760 --> 00:47:29.079
a neural model over the whole web is
00:47:27.440 --> 00:47:32.000
expensive so you can use a role-based
00:47:29.079 --> 00:47:35.319
model so some examples of this include
00:47:32.000 --> 00:47:37.640
text Runner and Reverb um the basic
00:47:35.319 --> 00:47:41.000
ideas behind them is that you use a
00:47:37.640 --> 00:47:43.720
parser to extract um to do a syntactic
00:47:41.000 --> 00:47:45.760
analysis of the sentence um in extract
00:47:43.720 --> 00:47:47.640
during according to rules so for example
00:47:45.760 --> 00:47:50.160
the relation must contain a
00:47:47.640 --> 00:47:52.720
predicate um the subject and object must
00:47:50.160 --> 00:47:56.040
be noun phrases other things like
00:47:52.720 --> 00:47:57.640
this um and then what they did later is
00:47:56.040 --> 00:47:59.240
what they did in this this paper
00:47:57.640 --> 00:48:00.800
arguably this is maybe no longer
00:47:59.240 --> 00:48:02.280
necessary with the compute power we have
00:48:00.800 --> 00:48:04.000
now but they trained an even faster
00:48:02.280 --> 00:48:06.960
model to extract over large amounts of
00:48:04.000 --> 00:48:08.720
data so they basically um use this as a
00:48:06.960 --> 00:48:10.599
su weak supervision and then train a
00:48:08.720 --> 00:48:12.160
model that could do it even faster with
00:48:10.599 --> 00:48:14.680
the sequence base
00:48:12.160 --> 00:48:18.119
model
00:48:14.680 --> 00:48:19.880
um another thing that they did was um
00:48:18.119 --> 00:48:22.280
they aggregated multiple pieces of
00:48:19.880 --> 00:48:24.480
evidence heris to find common and
00:48:22.280 --> 00:48:28.760
therefore potentially reliable
00:48:24.480 --> 00:48:28.760
extractions so like
00:48:29.800 --> 00:48:36.960
any piece of text on the internet like
00:48:31.559 --> 00:48:40.200
could be a lie right so um you know
00:48:36.960 --> 00:48:43.400
if I I might write on my blog United has
00:48:40.200 --> 00:48:45.119
a Hub in like Denver or on the other
00:48:43.400 --> 00:48:48.240
hand
00:48:45.119 --> 00:48:50.839
um wait a set
00:48:48.240 --> 00:48:52.680
right some something has a Hub in Denver
00:48:50.839 --> 00:48:54.960
but United has a Hub in Pittsburgh is
00:48:52.680 --> 00:48:58.040
definitely wrong so let's uh let's go
00:48:54.960 --> 00:49:00.000
with that um uh so somebody could write
00:48:58.040 --> 00:49:02.359
that on the internet and in fact because
00:49:00.000 --> 00:49:06.440
I just said it it's probably in YouTube
00:49:02.359 --> 00:49:09.119
comments somewhere but um uh
00:49:06.440 --> 00:49:10.760
like any any piece of information on the
00:49:09.119 --> 00:49:13.079
internet could be wrong so basically
00:49:10.760 --> 00:49:16.680
they had um heuristic methods to filter
00:49:13.079 --> 00:49:19.559
these out and usually these were
00:49:16.680 --> 00:49:21.559
frequency based so it's like um if both
00:49:19.559 --> 00:49:23.520
United and Pittsburgh are very common
00:49:21.559 --> 00:49:26.000
but it's very rare for somebody to says
00:49:23.520 --> 00:49:27.799
say United has a Hub in Pittsburgh then
00:49:26.000 --> 00:49:29.200
that means it's statistically unlikely
00:49:27.799 --> 00:49:30.799
for this to be correct because if it
00:49:29.200 --> 00:49:33.280
were correct we'd expect to see it much
00:49:30.799 --> 00:49:36.799
more frequently so um those were the
00:49:33.280 --> 00:49:36.799
kind of things that they they did
00:49:37.520 --> 00:49:44.440
here there's also some neural models for
00:49:40.400 --> 00:49:46.839
open IE um I I think these are uh used
00:49:44.440 --> 00:49:48.440
maybe a little bit less often um but
00:49:46.839 --> 00:49:52.559
basically heuristics are still not
00:49:48.440 --> 00:49:55.280
perfect and so what they did the problem
00:49:52.559 --> 00:49:56.720
with um like not relying on heuristics
00:49:55.280 --> 00:49:58.880
is you need to get training data from
00:49:56.720 --> 00:50:01.880
somewhere so there's a rather clever
00:49:58.880 --> 00:50:03.599
paper um and again if you're not
00:50:01.880 --> 00:50:05.119
interested in relation extraction in
00:50:03.599 --> 00:50:07.559
particular I think this is one thing
00:50:05.119 --> 00:50:10.000
that's still worth paying attention to
00:50:07.559 --> 00:50:12.680
um which is
00:50:10.000 --> 00:50:14.559
they demonstrated that it's possible to
00:50:12.680 --> 00:50:16.319
create relatively large data sets by
00:50:14.559 --> 00:50:18.160
asking people simple
00:50:16.319 --> 00:50:21.440
questions
00:50:18.160 --> 00:50:24.480
and in particular they wanted to
00:50:21.440 --> 00:50:27.119
get relation extraction data sets that
00:50:24.480 --> 00:50:30.799
are like um
00:50:27.119 --> 00:50:34.200
who finished something like UCD finished
00:50:30.799 --> 00:50:37.760
the two 2006 championships and if you
00:50:34.200 --> 00:50:40.720
ask people like okay select this span um
00:50:37.760 --> 00:50:44.559
select the entity span the relations
00:50:40.720 --> 00:50:46.160
span and the um in the second entity the
00:50:44.559 --> 00:50:49.079
head entity the relation and the tail
00:50:46.160 --> 00:50:51.839
entity select it on this interface and
00:50:49.079 --> 00:50:54.200
then uh tell me is it this relation or
00:50:51.839 --> 00:50:55.640
this relation or this relation that's
00:50:54.200 --> 00:50:58.160
actually pretty hard and getting like
00:50:55.640 --> 00:51:01.280
crowd workers to start learning how to
00:50:58.160 --> 00:51:03.280
do that task is a bit tricky and it
00:51:01.280 --> 00:51:06.400
takes some you know it takes some time
00:51:03.280 --> 00:51:07.799
to get them onboarded basically um but
00:51:06.400 --> 00:51:09.760
basically what they said is instead
00:51:07.799 --> 00:51:11.359
we'll just ask them questions where the
00:51:09.760 --> 00:51:14.240
answer to the question basically gives
00:51:11.359 --> 00:51:17.160
us the answer to what the relation is so
00:51:14.240 --> 00:51:20.319
they ask like who finished something and
00:51:17.160 --> 00:51:23.680
the answer is like UCD and um what did
00:51:20.319 --> 00:51:25.359
someone finish the 2006 Championship
00:51:23.680 --> 00:51:28.920
what did someone fish some finish
00:51:25.359 --> 00:51:31.760
something as and basically um in doing
00:51:28.920 --> 00:51:33.319
this they created uh something called
00:51:31.760 --> 00:51:34.359
semantic roles which we're actually
00:51:33.319 --> 00:51:35.960
probably going to talk about a little
00:51:34.359 --> 00:51:37.559
bit later but you can take the semantic
00:51:35.960 --> 00:51:41.200
roles and then you can use them to
00:51:37.559 --> 00:51:43.920
annotate uh relation extraction data and
00:51:41.200 --> 00:51:46.720
then they trained a supervised neural
00:51:43.920 --> 00:51:46.720
tager for
00:51:48.799 --> 00:51:53.480
this
00:51:50.480 --> 00:51:56.040
cool um so another thing I'd like to
00:51:53.480 --> 00:51:57.880
talk about is I talked about learning um
00:51:56.040 --> 00:51:59.920
information about entities from entity
00:51:57.880 --> 00:52:02.079
embeddings but you can actually learn
00:51:59.920 --> 00:52:04.520
information about relations from
00:52:02.079 --> 00:52:07.680
relation information about other
00:52:04.520 --> 00:52:12.359
relations and this can help solve the
00:52:07.680 --> 00:52:16.119
problem um of like essentially the fact
00:52:12.359 --> 00:52:18.760
that open IE is not able to abstract and
00:52:16.119 --> 00:52:20.680
generalize so word embeddings or entity
00:52:18.760 --> 00:52:23.079
embeddings give information of the word
00:52:20.680 --> 00:52:26.920
in context um which can be indicative
00:52:23.079 --> 00:52:29.640
for knowledge uh knowledge bases
00:52:26.920 --> 00:52:32.640
but other relations or combinations
00:52:29.640 --> 00:52:34.960
thereof are also indicative of them and
00:52:32.640 --> 00:52:36.960
um if anybody is familiar with graphs or
00:52:34.960 --> 00:52:39.520
graph processing there's the whole idea
00:52:36.960 --> 00:52:41.400
of um link prediction where you're given
00:52:39.520 --> 00:52:42.680
like a a small number of links in a
00:52:41.400 --> 00:52:45.760
graph and you want to predict what other
00:52:42.680 --> 00:52:50.559
links are likely to uh
00:52:45.760 --> 00:52:52.920
exist and like as I said um a lot of uh
00:52:50.559 --> 00:52:54.839
you know very prominent AI researchers
00:52:52.920 --> 00:52:57.440
got their start in uh relation
00:52:54.839 --> 00:53:01.480
extraction and uh it sker is another one
00:52:57.440 --> 00:53:04.319
of them actually um and uh basically
00:53:01.480 --> 00:53:07.880
this 2009 paper proposed to use tensor
00:53:04.319 --> 00:53:09.400
de composition to do uh induction of
00:53:07.880 --> 00:53:13.520
relations
00:53:09.400 --> 00:53:15.319
and the way it worked is um you model
00:53:13.520 --> 00:53:18.400
relations by decomposing a tensor
00:53:15.319 --> 00:53:21.599
containing entity relation entity tles
00:53:18.400 --> 00:53:24.000
so you have the left entity the right
00:53:21.599 --> 00:53:27.160
entity and whether the relation exists
00:53:24.000 --> 00:53:31.319
is this big um uh big tensor in the
00:53:27.160 --> 00:53:33.160
Middle where these are embeddings of the
00:53:31.319 --> 00:53:35.760
left entity these are embeddings of the
00:53:33.160 --> 00:53:38.839
right entity and then the the depth of
00:53:35.760 --> 00:53:40.680
the tensor is like which relations exist
00:53:38.839 --> 00:53:43.760
and so we know that some exist so we
00:53:40.680 --> 00:53:46.640
give them a one we know others exist um
00:53:43.760 --> 00:53:48.680
don't exist so we give them a zero um
00:53:46.640 --> 00:53:51.040
and then we do a low rank approximation
00:53:48.680 --> 00:53:52.559
of this tensor and if we do a low rank
00:53:51.040 --> 00:53:55.720
approximation of the tensor we have
00:53:52.559 --> 00:53:57.280
reconstruction ER basically so when we
00:53:55.720 --> 00:53:59.960
reconstruct the are some things that
00:53:57.280 --> 00:54:01.960
were previously zero become one and so
00:53:59.960 --> 00:54:04.760
the things that were previously zero and
00:54:01.960 --> 00:54:07.880
then become close to one are the ones
00:54:04.760 --> 00:54:10.559
that we think like actually might exist
00:54:07.880 --> 00:54:12.000
they might be real um they might be real
00:54:10.559 --> 00:54:13.640
relations that we were just missing
00:54:12.000 --> 00:54:16.599
because our previous knowledge base was
00:54:13.640 --> 00:54:16.599
complete uh
00:54:18.640 --> 00:54:26.880
incomplete and um one thing that takes
00:54:21.799 --> 00:54:28.559
us a step further is uh what if if we
00:54:26.880 --> 00:54:30.079
actually do have a knowledge basee or
00:54:28.559 --> 00:54:31.839
what if we even have multiple knowledge
00:54:30.079 --> 00:54:35.520
bases like what if we have Wiki data and
00:54:31.839 --> 00:54:36.640
we have wordnet and we have um uh other
00:54:35.520 --> 00:54:38.920
things like
00:54:36.640 --> 00:54:40.680
this and in addition to that we also
00:54:38.920 --> 00:54:43.400
have open IE
00:54:40.680 --> 00:54:45.960
extractions so there's an idea of
00:54:43.400 --> 00:54:47.880
something called Universal schema and
00:54:45.960 --> 00:54:50.200
what Universal schema do is they embed
00:54:47.880 --> 00:54:55.119
relations from multiple schema or
00:54:50.200 --> 00:54:56.960
schemata in the same space and based on
00:54:55.119 --> 00:54:59.559
this they then
00:54:56.960 --> 00:55:01.359
predict which ones exist are likely to
00:54:59.559 --> 00:55:04.400
exist or which ones are not likely to
00:55:01.359 --> 00:55:06.680
exist so here we might have a free base
00:55:04.400 --> 00:55:08.640
or Wiki data we might have another uh
00:55:06.680 --> 00:55:11.559
kind of relation extraction data set
00:55:08.640 --> 00:55:15.480
called Tac and then on the training data
00:55:11.559 --> 00:55:17.040
set we have um like all of these uh
00:55:15.480 --> 00:55:20.240
things that are like positive or
00:55:17.040 --> 00:55:23.960
negative or something like this and then
00:55:20.240 --> 00:55:26.960
on the heldout data set we have only
00:55:23.960 --> 00:55:29.480
information about like open
00:55:26.960 --> 00:55:30.920
for example so um for all of the
00:55:29.480 --> 00:55:33.079
entities that exist in the knowledge
00:55:30.920 --> 00:55:34.839
base we know you know whether the
00:55:33.079 --> 00:55:36.039
relations exist for but for all the
00:55:34.839 --> 00:55:39.640
entities that don't exist in the
00:55:36.039 --> 00:55:41.760
database we don't know and so uh then
00:55:39.640 --> 00:55:43.839
just from the existence of open IE
00:55:41.760 --> 00:55:45.480
relations or non-existence of open IE
00:55:43.839 --> 00:55:47.920
relations we can predict that other
00:55:45.480 --> 00:55:49.359
relations might exist for example so
00:55:47.920 --> 00:55:51.079
this is a great way to combine the two
00:55:49.359 --> 00:55:53.920
together like open IE you can run it
00:55:51.079 --> 00:55:55.880
over you know very large data sets um
00:55:53.920 --> 00:55:58.000
but it doesn't have a good schema free
00:55:55.880 --> 00:56:00.400
uh Wiki data has a good schema but you
00:55:58.000 --> 00:56:02.960
can't you know it's all manually created
00:56:00.400 --> 00:56:04.720
so you can suggest other ones and one
00:56:02.960 --> 00:56:07.960
other like interesting thing is you can
00:56:04.720 --> 00:56:09.640
suggest other um things that might exist
00:56:07.960 --> 00:56:13.039
in Wiki data but you could also track
00:56:09.640 --> 00:56:15.039
that back to the original text that
00:56:13.039 --> 00:56:17.000
indicated that it might exist in Wiki
00:56:15.039 --> 00:56:18.720
data so then you could have a human go
00:56:17.000 --> 00:56:20.520
back and check it to make sure that
00:56:18.720 --> 00:56:24.200
that's actually true and trustworthy and
00:56:20.520 --> 00:56:24.200
other things like that
00:56:26.400 --> 00:56:31.400
cool um so if you like uh you like
00:56:29.400 --> 00:56:33.160
tensors or you like linear algebra or
00:56:31.400 --> 00:56:34.720
things like this this is maybe something
00:56:33.160 --> 00:56:37.880
that you could take a look at and think
00:56:34.720 --> 00:56:40.240
a little bit more about um any any
00:56:37.880 --> 00:56:40.240
questions
00:56:42.799 --> 00:56:46.240
here okay
00:56:46.880 --> 00:56:53.680
cool um so another thing I'd like to
00:56:50.640 --> 00:56:56.920
talk about is uh modeling relation paths
00:56:53.680 --> 00:57:00.359
so this is a really nice uh idea
00:56:56.920 --> 00:57:00.359
which is you
00:57:00.440 --> 00:57:05.000
can make inferences across multiple hops
00:57:04.240 --> 00:57:08.400
of
00:57:05.000 --> 00:57:12.280
relations um based on uh particular
00:57:08.400 --> 00:57:14.200
relations existing and so um multi-step
00:57:12.280 --> 00:57:17.280
passs can be informative for indicating
00:57:14.200 --> 00:57:20.000
whether individual relations exist so um
00:57:17.280 --> 00:57:24.400
for example uh given a word given a
00:57:20.000 --> 00:57:27.960
particular word in a paper title
00:57:24.400 --> 00:57:29.880
recommend a venue in which to the paper
00:57:27.960 --> 00:57:32.559
and so this is the the problem that they
00:57:29.880 --> 00:57:36.079
were trying to solve and then basically
00:57:32.559 --> 00:57:38.440
you have a word um you
00:57:36.079 --> 00:57:41.119
find if you have that word in your paper
00:57:38.440 --> 00:57:42.920
title you then find other papers that
00:57:41.119 --> 00:57:45.280
have that title uh that have that word
00:57:42.920 --> 00:57:48.359
in their title and those papers are in a
00:57:45.280 --> 00:57:52.039
journal and that gets a high weight with
00:57:48.359 --> 00:57:54.119
respect to like that your paper being
00:57:52.039 --> 00:57:56.839
you know relevant to that particular
00:57:54.119 --> 00:57:59.880
Journal you can also say
00:57:56.839 --> 00:58:01.000
okay I have a a word find papers with
00:57:59.880 --> 00:58:03.240
that word in the
00:58:01.000 --> 00:58:07.240
title find the first author of that
00:58:03.240 --> 00:58:09.280
paper find another paper uh that had
00:58:07.240 --> 00:58:11.599
that author as a first author and then
00:58:09.280 --> 00:58:13.240
find the Journal of it and they
00:58:11.599 --> 00:58:15.839
demonstrate a way where you can like
00:58:13.240 --> 00:58:18.280
expand these paths and feed them into a
00:58:15.839 --> 00:58:22.400
prediction model and use that to predict
00:58:18.280 --> 00:58:25.480
um you know additional relations so
00:58:22.400 --> 00:58:26.680
unlike this method here this method was
00:58:25.480 --> 00:58:29.240
saying like
00:58:26.680 --> 00:58:30.920
other single relations are indicative of
00:58:29.240 --> 00:58:34.160
a particular relation
00:58:30.920 --> 00:58:36.880
existing this paper is saying not just
00:58:34.160 --> 00:58:38.720
individual relations are indicative of
00:58:36.880 --> 00:58:40.640
another relation existing but actually
00:58:38.720 --> 00:58:43.839
relation paths are indicative of a
00:58:40.640 --> 00:58:46.400
relation existing so this is more um
00:58:43.839 --> 00:58:46.400
expressive
00:58:47.520 --> 00:58:55.359
basically um and this followup paper
00:58:52.640 --> 00:58:57.480
uh using differentiable logic rules
00:58:55.359 --> 00:59:00.799
actually made this endtoend
00:58:57.480 --> 00:59:03.079
trainable so this allows you to consider
00:59:00.799 --> 00:59:07.599
whole paths in a differentiable
00:59:03.079 --> 00:59:09.960
framework and so the way they did this
00:59:07.599 --> 00:59:13.359
is like if you have you know City in
00:59:09.960 --> 00:59:16.440
country and has office in country um
00:59:13.359 --> 00:59:18.920
that or sorry City and Country and has
00:59:16.440 --> 00:59:22.200
office in city that indicates has office
00:59:18.920 --> 00:59:24.160
in country and I I'm sure you know many
00:59:22.200 --> 00:59:26.760
people here have thought like learned
00:59:24.160 --> 00:59:29.520
about logic and you know and induction
00:59:26.760 --> 00:59:32.720
from or deduction from uh logic rules
00:59:29.520 --> 00:59:34.359
and stuff like this but the problem is
00:59:32.720 --> 00:59:37.079
deduction from logic rules is very
00:59:34.359 --> 00:59:39.039
fragile like there are cases where there
00:59:37.079 --> 00:59:41.119
are counter examples so if you say that
00:59:39.039 --> 00:59:43.280
something is always true deductively
00:59:41.119 --> 00:59:45.839
then um that can cause problems so in
00:59:43.280 --> 00:59:47.839
reality it's like if you have two pieces
00:59:45.839 --> 00:59:52.400
of information something can become much
00:59:47.839 --> 00:59:56.920
much more likely um and so you know just
00:59:52.400 --> 00:59:59.880
to give an example um somebody studying
00:59:56.920 --> 01:00:01.280
studying at CMU makes it very likely
00:59:59.880 --> 01:00:03.799
much more likely that they're studying
01:00:01.280 --> 01:00:06.359
computer science and much less likely
01:00:03.799 --> 01:00:08.000
that they're studying medicine or
01:00:06.359 --> 01:00:09.520
something like that but that doesn't
01:00:08.000 --> 01:00:11.720
mean that it like
01:00:09.520 --> 01:00:13.559
entirely the first one is definitely not
01:00:11.720 --> 01:00:15.480
entirely implied and I'm sure there's
01:00:13.559 --> 01:00:16.760
like a few people at CMU who are somehow
01:00:15.480 --> 01:00:18.440
studying medicine through a joint
01:00:16.760 --> 01:00:21.480
program with pit or something like that
01:00:18.440 --> 01:00:24.400
so you know like very it's very rare
01:00:21.480 --> 01:00:26.799
that logic rules are hard and fast and
01:00:24.400 --> 01:00:28.480
so basically what they do is they treat
01:00:26.799 --> 01:00:30.559
each path as a sequence of Matrix
01:00:28.480 --> 01:00:34.839
multiplies it where they have a rule
01:00:30.559 --> 01:00:36.599
weight um like this and um in the end
01:00:34.839 --> 01:00:38.359
that allows you to make a a prediction
01:00:36.599 --> 01:00:40.839
about whether a predic logic rule is
01:00:38.359 --> 01:00:40.839
correct or
01:00:40.880 --> 01:00:49.319
not um so this is uh i' I've been
01:00:46.880 --> 01:00:51.119
working mostly in like structured
01:00:49.319 --> 01:00:54.480
knowledge space structured knowledge
01:00:51.119 --> 01:00:56.599
graphs other uh other things like this
01:00:54.480 --> 01:00:59.760
um I I don't
01:00:56.599 --> 01:01:02.720
think there's a whole lot of work that
01:00:59.760 --> 01:01:05.640
directly applies this to language models
01:01:02.720 --> 01:01:07.319
um like differentiable logic rules and
01:01:05.640 --> 01:01:10.079
language models or things like that just
01:01:07.319 --> 01:01:12.440
because it's less clean it's you know uh
01:01:10.079 --> 01:01:13.839
harder um there there's a little bit of
01:01:12.440 --> 01:01:16.079
work which I'm going to talk about now
01:01:13.839 --> 01:01:18.599
but I think like this kind of work is
01:01:16.079 --> 01:01:21.440
interesting because a lot of models are
01:01:18.599 --> 01:01:23.119
not super great at reasoning and how to
01:01:21.440 --> 01:01:25.119
like allow them to be better at
01:01:23.119 --> 01:01:26.559
reasoning is kind of an open problem so
01:01:25.119 --> 01:01:28.039
learning from these old older works that
01:01:26.559 --> 01:01:30.200
did it in a more structured space and
01:01:28.039 --> 01:01:32.160
trying to figure out how to apply them
01:01:30.200 --> 01:01:34.400
to less structured spaces is still
01:01:32.160 --> 01:01:36.240
interesting I think
01:01:34.400 --> 01:01:39.160
so
01:01:36.240 --> 01:01:40.720
cool um then the final talk topic I want
01:01:39.160 --> 01:01:42.920
to talk about is probing knowledge in
01:01:40.720 --> 01:01:44.920
LMS and so we have these knowledge bases
01:01:42.920 --> 01:01:47.319
that encode you know tons and tons of
01:01:44.920 --> 01:01:49.880
knowledge um which allows us to figure
01:01:47.319 --> 01:01:52.200
out you know oh well how well do uh
01:01:49.880 --> 01:01:56.200
language models know about these
01:01:52.200 --> 01:01:59.079
things and so
01:01:56.200 --> 01:02:02.760
traditional um kind of QA machine
01:01:59.079 --> 01:02:04.799
reading comprehension rag models um
01:02:02.760 --> 01:02:06.359
usually referred to external resources
01:02:04.799 --> 01:02:10.039
to answer questions like Wikipedia
01:02:06.359 --> 01:02:14.359
articles um or things like this but then
01:02:10.039 --> 01:02:16.119
the question is without doing rag can we
01:02:14.359 --> 01:02:18.160
you know answer questions like what
01:02:16.119 --> 01:02:20.920
knowledge is
01:02:18.160 --> 01:02:24.079
encoded and so the first paper that kind
01:02:20.920 --> 01:02:26.520
of handled this sort of problem uh is
01:02:24.079 --> 01:02:29.200
this paper which actually was also
01:02:26.520 --> 01:02:33.359
called uh
01:02:29.200 --> 01:02:35.960
wama surprisingly um or released a
01:02:33.359 --> 01:02:41.000
resource called llama except it was l m
01:02:35.960 --> 01:02:44.880
a um but what they did is they
01:02:41.000 --> 01:02:46.960
uh used they in contrast to using
01:02:44.880 --> 01:02:50.000
structural queries like SQL or or
01:02:46.960 --> 01:02:52.119
Sparkle two query KBS they tried to use
01:02:50.000 --> 01:02:54.240
natural language prompts to query LM so
01:02:52.119 --> 01:02:58.160
this was actually one of the the first
01:02:54.240 --> 01:03:02.359
uh kind of paper on prompts uh prompting
01:02:58.160 --> 01:03:05.079
for uh language models in a way and the
01:03:02.359 --> 01:03:08.359
way they did this is they had um they
01:03:05.079 --> 01:03:10.039
did like Dante was born in mask and then
01:03:08.359 --> 01:03:13.279
they tried to fill in the mask using a
01:03:10.039 --> 01:03:15.839
mask language model and uh and output
01:03:13.279 --> 01:03:18.559
Florence so
01:03:15.839 --> 01:03:19.960
um when they did this work now now we
01:03:18.559 --> 01:03:21.359
don't do this quite as much but when
01:03:19.960 --> 01:03:23.520
they did this work they basically used
01:03:21.359 --> 01:03:25.440
the knowledge base as the ground truth
01:03:23.520 --> 01:03:28.880
and tried to probe whether the knowledge
01:03:25.440 --> 01:03:31.520
in in um in the knowledge base was also
01:03:28.880 --> 01:03:34.880
uh recoverable from the neural
01:03:31.520 --> 01:03:37.720
map um and they proposed the Llama
01:03:34.880 --> 01:03:39.760
Benchmark um basically it was manual
01:03:37.720 --> 01:03:42.480
prompts for 41 relations they created
01:03:39.760 --> 01:03:44.839
the prompts manually uh so like X was
01:03:42.480 --> 01:03:46.480
founded in y The Prompt template and
01:03:44.839 --> 01:03:49.400
they filled in the subjects and had the
01:03:46.480 --> 01:03:52.160
LMS uh for such as Bert predict the
01:03:49.400 --> 01:03:55.839
objects uh like blueberg LP was founded
01:03:52.160 --> 01:03:59.000
in mask and they demonstrated that like
01:03:55.839 --> 01:04:02.440
basically Elmo uh Transformer XL and
01:03:59.000 --> 01:04:04.960
Bert base got uh you know up to 31%
01:04:02.440 --> 01:04:06.480
accuracy now I'm sure uh the modern
01:04:04.960 --> 01:04:09.200
language models would have much higher
01:04:06.480 --> 01:04:11.279
accuracy than
01:04:09.200 --> 01:04:13.920
that
01:04:11.279 --> 01:04:17.839
um this is a a follow-up paper that we
01:04:13.920 --> 01:04:21.160
did to this um where we tried to do this
01:04:17.839 --> 01:04:23.400
multilingually um I I think this is
01:04:21.160 --> 01:04:25.680
really let
01:04:23.400 --> 01:04:29.520
me I think one thing that's interesting
01:04:25.680 --> 01:04:31.960
interesting about this paper is um even
01:04:29.520 --> 01:04:37.240
if you're not interested in multilingual
01:04:31.960 --> 01:04:38.920
stuff per se there is an interesting
01:04:37.240 --> 01:04:40.760
dichotomy about like what knowledge is
01:04:38.920 --> 01:04:43.079
included in LMS and whether we can
01:04:40.760 --> 01:04:46.000
retrieve it and the reason why I'm
01:04:43.079 --> 01:04:48.359
saying this is because in this paper
01:04:46.000 --> 01:04:51.200
we created
01:04:48.359 --> 01:04:52.599
queries from a knowledge base and
01:04:51.200 --> 01:04:54.160
because we created queries from a
01:04:52.599 --> 01:04:55.760
knowledge base and knowledge bases are
01:04:54.160 --> 01:04:57.240
multilingual we can also create
01:04:55.760 --> 01:05:00.039
multilingual queries from knowledge
01:04:57.240 --> 01:05:01.720
bases right so we can use exactly the
01:05:00.039 --> 01:05:03.359
same entities but just ask the same
01:05:01.720 --> 01:05:05.920
question in different languages and so
01:05:03.359 --> 01:05:07.480
we had a bunch of people manually uh
01:05:05.920 --> 01:05:10.119
create prompts for all of these
01:05:07.480 --> 01:05:13.000
languages here and you can see that in
01:05:10.119 --> 01:05:15.960
English it's much better at responding
01:05:13.000 --> 01:05:19.000
uh to these queries than it is in any
01:05:15.960 --> 01:05:21.039
other language and in particular like
01:05:19.000 --> 01:05:22.880
lower resource languages or languages
01:05:21.039 --> 01:05:26.400
that are less similar to English it did
01:05:22.880 --> 01:05:29.079
much worse and notably we we counted the
01:05:26.400 --> 01:05:32.160
answer correct if it got it
01:05:29.079 --> 01:05:34.279
um we we had two settings one setting is
01:05:32.160 --> 01:05:35.799
we counted the answer correct if it only
01:05:34.279 --> 01:05:38.359
if it answered in the language we
01:05:35.799 --> 01:05:39.680
queried it in but we in other setting we
01:05:38.359 --> 01:05:42.640
also counted the answer correct if it
01:05:39.680 --> 01:05:44.200
answered in any language so we um it
01:05:42.640 --> 01:05:46.640
didn't necessarily have to even know the
01:05:44.200 --> 01:05:48.200
name of the entity in that uh language
01:05:46.640 --> 01:05:50.520
and we would still count it
01:05:48.200 --> 01:05:54.720
correct and so what I mean by there's a
01:05:50.520 --> 01:05:56.440
dichotomy between the information that
01:05:54.720 --> 01:05:59.240
language models have
01:05:56.440 --> 01:06:02.480
encoded and whether they're able to
01:05:59.240 --> 01:06:02.480
retrieve it
01:06:02.680 --> 01:06:07.640
is in English it's able to answer the
01:06:06.000 --> 01:06:10.799
models we tested were able to answer
01:06:07.640 --> 01:06:13.000
like 177% of queries
01:06:10.799 --> 01:06:14.359
but if the fact that they're able to
01:06:13.000 --> 01:06:16.160
answer in English means that the
01:06:14.359 --> 01:06:18.520
language model quote unquote knows the
01:06:16.160 --> 01:06:20.200
answer right like it knows the answer in
01:06:18.520 --> 01:06:22.680
English we're asking exactly the same
01:06:20.200 --> 01:06:24.400
question in all the other languages so
01:06:22.680 --> 01:06:26.079
you know it should know the answer in
01:06:24.400 --> 01:06:27.680
the other languages too
01:06:26.079 --> 01:06:30.000
but it's not able to retrieve the answer
01:06:27.680 --> 01:06:33.079
because we asked in another language
01:06:30.000 --> 01:06:35.920
so um that brings up some interesting
01:06:33.079 --> 01:06:38.079
questions about how we can make models
01:06:35.920 --> 01:06:39.680
better at retrieving the the knowledge
01:06:38.079 --> 01:06:43.559
that they already know in English when
01:06:39.680 --> 01:06:45.520
you query them in other languages or um
01:06:43.559 --> 01:06:48.119
and there was another paper recently I
01:06:45.520 --> 01:06:52.720
don't know if I'd be able to find it um
01:06:48.119 --> 01:06:56.119
exactly which is um they
01:06:52.720 --> 01:07:01.799
prompted models with personas and so
01:06:56.119 --> 01:07:04.599
they said I um you know I am a old man I
01:07:01.799 --> 01:07:07.160
am an old woman I am a young man I am
01:07:04.599 --> 01:07:10.039
young woman I am a child or something
01:07:07.160 --> 01:07:12.799
like that um or they also talked about
01:07:10.039 --> 01:07:15.640
things like uh physical disabilities and
01:07:12.799 --> 01:07:17.200
things and they said um please answer
01:07:15.640 --> 01:07:19.640
this question after they prompted with a
01:07:17.200 --> 01:07:22.680
Persona and just having that Persona
01:07:19.640 --> 01:07:24.839
greatly changed the ability of the model
01:07:22.680 --> 01:07:26.400
to answer questions so it's this very
01:07:24.839 --> 01:07:28.200
weird thing which which is like the
01:07:26.400 --> 01:07:29.799
models are actually capable of answering
01:07:28.200 --> 01:07:31.520
the questions but based on how you probe
01:07:29.799 --> 01:07:32.880
them whether it's in like different
01:07:31.520 --> 01:07:34.599
languages or if you give them a
01:07:32.880 --> 01:07:36.839
different Persona they manage to answer
01:07:34.599 --> 01:07:39.000
things differently and so on the plus
01:07:36.839 --> 01:07:42.920
side like you can create you can make
01:07:39.000 --> 01:07:44.799
ways to reduce the language models
01:07:42.920 --> 01:07:45.920
performance by giving it like a Persona
01:07:44.799 --> 01:07:49.839
that shouldn't be good at answering
01:07:45.920 --> 01:07:53.279
questions or something like that um
01:07:49.839 --> 01:07:54.839
but on the plus side um like when you're
01:07:53.279 --> 01:07:57.279
doing code generation there was this
01:07:54.839 --> 01:07:58.960
magic prompt which is like um I have
01:07:57.279 --> 01:08:01.319
checked this carefully in all the unit
01:07:58.960 --> 01:08:03.240
tests pass and that would improve your
01:08:01.319 --> 01:08:05.760
code generation accuracy by like five
01:08:03.240 --> 01:08:07.559
five points or something like that so um
01:08:05.760 --> 01:08:09.240
you just get the the model in the right
01:08:07.559 --> 01:08:11.359
mood to answer the question accurately
01:08:09.240 --> 01:08:13.319
and it does a better job at doing it so
01:08:11.359 --> 01:08:15.960
it's kind of uh it goes in both
01:08:13.319 --> 01:08:15.960
directions I
01:08:16.679 --> 01:08:27.080
guess cool um yeah uh any any questions
01:08:23.679 --> 01:08:30.120
here um another thing that you can do uh
01:08:27.080 --> 01:08:31.000
is fine-tune models specifically so
01:08:30.120 --> 01:08:34.080
they're good at answering
01:08:31.000 --> 01:08:35.560
knowledge-based questions so um uh this
01:08:34.080 --> 01:08:38.080
paper demonstrated that you could find
01:08:35.560 --> 01:08:39.480
tune models uh on synthetically created
01:08:38.080 --> 01:08:41.159
knowledge based questions and that would
01:08:39.480 --> 01:08:42.920
improve the ability of the model to
01:08:41.159 --> 01:08:47.679
answer questions about knowledge
01:08:42.920 --> 01:08:47.679
bases um it's
01:08:49.120 --> 01:08:57.440
uh yeah um it's pretty straightforward
01:08:53.199 --> 01:08:57.440
so uh there's that
01:08:57.799 --> 01:09:03.120
um yeah we already talked about this in
01:09:00.000 --> 01:09:07.560
the rag class so I think I might skip
01:09:03.120 --> 01:09:10.239
that um a final paper that I'd like to
01:09:07.560 --> 01:09:12.600
talk about this is also a paper uh done
01:09:10.239 --> 01:09:13.759
by my student Jung B Jong and this is
01:09:12.600 --> 01:09:16.080
interesting from the point of view of
01:09:13.759 --> 01:09:18.000
multihop reasoning and so I talked a
01:09:16.080 --> 01:09:19.679
little bit about like multihop reasoning
01:09:18.000 --> 01:09:23.239
along reasoning
01:09:19.679 --> 01:09:26.159
chains um in knowledge bases and this is
01:09:23.239 --> 01:09:28.520
one example of multihop reasoning
01:09:26.159 --> 01:09:30.080
among along reasoning chains within the
01:09:28.520 --> 01:09:33.400
parameters of the model so testing
01:09:30.080 --> 01:09:36.759
whether models can answer
01:09:33.400 --> 01:09:38.480
um Can it answer multihop questions and
01:09:36.759 --> 01:09:40.839
basically what we did here is we took a
01:09:38.480 --> 01:09:42.679
knowledge base and a knowledge base can
01:09:40.839 --> 01:09:44.279
have
01:09:42.679 --> 01:09:49.480
um
01:09:44.279 --> 01:09:49.480
like uh country country is
01:09:49.600 --> 01:09:52.600
US
01:09:53.480 --> 01:09:58.600
president um and then a
01:10:00.880 --> 01:10:06.560
birthday um and so we can create these
01:10:04.280 --> 01:10:08.640
multihop questions right uh and just
01:10:06.560 --> 01:10:10.280
follow the relation links and then we
01:10:08.640 --> 01:10:11.440
know the answer to the multihop question
01:10:10.280 --> 01:10:13.560
by following the link and we can
01:10:11.440 --> 01:10:18.159
generate you know the question given a
01:10:13.560 --> 01:10:19.800
template um so we did this and had like
01:10:18.159 --> 01:10:22.800
question one which is return the artist
01:10:19.800 --> 01:10:25.719
who recorded party a over um and then
01:10:22.800 --> 01:10:28.159
where in Georgia does uh Usher live and
01:10:25.719 --> 01:10:29.920
then we can turn this into a question
01:10:28.159 --> 01:10:31.679
which part of Georgia in which part of
01:10:29.920 --> 01:10:34.239
Georgia does the artist that recorded
01:10:31.679 --> 01:10:37.560
the party8 overlive and so we now have a
01:10:34.239 --> 01:10:45.000
multi multihop question and what we did
01:10:37.560 --> 01:10:47.440
is we measured whether um the model was
01:10:45.000 --> 01:10:49.760
able to answer the first question the
01:10:47.440 --> 01:10:53.320
second question and the comp like
01:10:49.760 --> 01:10:56.120
compound question and what we found is
01:10:53.320 --> 01:10:59.440
like what we would expect
01:10:56.120 --> 01:11:01.719
if models were like perfect knowledge
01:10:59.440 --> 01:11:04.360
processors right
01:11:01.719 --> 01:11:08.120
is we have
01:11:04.360 --> 01:11:10.800
like yes on the first question
01:11:08.120 --> 01:11:14.000
no
01:11:10.800 --> 01:11:16.560
yes um yes on the first question and no
01:11:14.000 --> 01:11:16.560
on the first
01:11:17.199 --> 01:11:24.760
question and we would expect that
01:11:21.920 --> 01:11:26.080
basically if it knew both of the answers
01:11:24.760 --> 01:11:27.239
to the first question and the second
01:11:26.080 --> 01:11:30.600
question it would get the compound
01:11:27.239 --> 01:11:31.800
question right and if it got uh like
01:11:30.600 --> 01:11:34.800
either of them wrong it would get it
01:11:31.800 --> 01:11:37.120
wrong right um you know in the in the
01:11:34.800 --> 01:11:39.400
ideal world where the knowledge of the
01:11:37.120 --> 01:11:41.280
two sub questions is necessary to answer
01:11:39.400 --> 01:11:43.880
the comp composite question and the
01:11:41.280 --> 01:11:45.840
model is a perfect knowledge processor
01:11:43.880 --> 01:11:47.120
and basically what we found we tried a
01:11:45.840 --> 01:11:49.280
whole bunch of different types of
01:11:47.120 --> 01:11:51.199
questions and what we found is this is
01:11:49.280 --> 01:11:55.960
totally not the case like it's not the
01:11:51.199 --> 01:11:58.520
case at all um and what we found in said
01:11:55.960 --> 01:12:01.560
is if it's able to answer the second
01:11:58.520 --> 01:12:04.120
question correctly it was much more
01:12:01.560 --> 01:12:07.480
likely to be able to answer the
01:12:04.120 --> 01:12:08.840
composite question um even if it can
01:12:07.480 --> 01:12:11.000
answer the first question that has
01:12:08.840 --> 01:12:13.120
almost no relation with whether it could
01:12:11.000 --> 01:12:15.520
answer the composite question at all so
01:12:13.120 --> 01:12:17.679
it's more like somehow from the answer
01:12:15.520 --> 01:12:19.320
to the second question it was able to to
01:12:17.679 --> 01:12:22.280
get the answer right and it kind of
01:12:19.320 --> 01:12:24.040
makes sense actually because like um
01:12:22.280 --> 01:12:26.320
let's say the answer to the second
01:12:24.040 --> 01:12:27.920
question is some like really long list
01:12:26.320 --> 01:12:30.719
like who are all the presidents of the
01:12:27.920 --> 01:12:33.320
United States um or something like that
01:12:30.719 --> 01:12:35.639
that's just hard to answer um so if I
01:12:33.320 --> 01:12:38.000
said who are all the presidents of the
01:12:35.639 --> 01:12:40.800
country where Washington DC is located
01:12:38.000 --> 01:12:42.679
in um you know like the second question
01:12:40.800 --> 01:12:44.040
is really hard so that's hard to get but
01:12:42.679 --> 01:12:46.120
if I say
01:12:44.040 --> 01:12:49.920
um
01:12:46.120 --> 01:12:53.520
uh what what is the
01:12:49.920 --> 01:12:57.120
capital what is the capital of the
01:12:53.520 --> 01:12:57.120
country uh
01:12:57.400 --> 01:13:02.440
what is what is the capital of the
01:12:58.840 --> 01:13:05.400
country where the most
01:13:02.440 --> 01:13:06.800
um people live or something like that
01:13:05.400 --> 01:13:08.679
even if you weren't sure about the
01:13:06.800 --> 01:13:10.880
country where the most people live you
01:13:08.679 --> 01:13:13.040
could pick a random capital and get it
01:13:10.880 --> 01:13:16.199
right some of the time or something like
01:13:13.040 --> 01:13:18.239
that so um that's what we found in this
01:13:16.199 --> 01:13:19.800
paper and I I think like another nice
01:13:18.239 --> 01:13:22.360
thing about knowledge bases is they
01:13:19.800 --> 01:13:24.880
allow you to ask like really interesting
01:13:22.360 --> 01:13:26.400
questions like this about what language
01:13:24.880 --> 01:13:29.120
model know or what language models don't
01:13:26.400 --> 01:13:31.040
know in a structured way so um I think
01:13:29.120 --> 01:13:32.280
if you're interested in probing language
01:13:31.040 --> 01:13:35.320
models and what they know and what they
01:13:32.280 --> 01:13:38.639
can infer what logic they can do that's
01:13:35.320 --> 01:13:42.320
good um cool yeah that's all I have for
01:13:38.639 --> 01:13:44.920
today um are there any questions or
01:13:42.320 --> 01:13:48.679
discussion or things like that or happy
01:13:44.920 --> 01:13:48.679
to talk up here too