|
1 |
|
00:00:01,280 --> 00:00:06,759 |
|
so the class today is uh introduction to |
|
|
|
2 |
|
00:00:04,680 --> 00:00:09,480 |
|
natural language processing and I'll be |
|
|
|
3 |
|
00:00:06,759 --> 00:00:11,200 |
|
talking a little bit about you know what |
|
|
|
4 |
|
00:00:09,480 --> 00:00:14,719 |
|
is natural language processing why we're |
|
|
|
5 |
|
00:00:11,200 --> 00:00:16,720 |
|
motivated to do it and also some of the |
|
|
|
6 |
|
00:00:14,719 --> 00:00:18,039 |
|
difficulties that we encounter and I'll |
|
|
|
7 |
|
00:00:16,720 --> 00:00:19,880 |
|
at the end I'll also be talking about |
|
|
|
8 |
|
00:00:18,039 --> 00:00:22,519 |
|
class Logistics so you can ask any |
|
|
|
9 |
|
00:00:19,880 --> 00:00:25,439 |
|
Logistics questions at that |
|
|
|
10 |
|
00:00:22,519 --> 00:00:27,720 |
|
time so if we talk about what is NLP |
|
|
|
11 |
|
00:00:25,439 --> 00:00:29,320 |
|
anyway uh does anyone have any opinions |
|
|
|
12 |
|
00:00:27,720 --> 00:00:31,439 |
|
about the definition of what natural |
|
|
|
13 |
|
00:00:29,320 --> 00:00:33,239 |
|
language process would be oh one other |
|
|
|
14 |
|
00:00:31,439 --> 00:00:35,680 |
|
thing I should mention is I am recording |
|
|
|
15 |
|
00:00:33,239 --> 00:00:38,600 |
|
the class uh I put the class on YouTube |
|
|
|
16 |
|
00:00:35,680 --> 00:00:40,520 |
|
uh afterwards I will not take pictures |
|
|
|
17 |
|
00:00:38,600 --> 00:00:41,920 |
|
or video of any of you uh but if you |
|
|
|
18 |
|
00:00:40,520 --> 00:00:44,719 |
|
talk your voice might come in the |
|
|
|
19 |
|
00:00:41,920 --> 00:00:47,440 |
|
background so just uh be aware of that |
|
|
|
20 |
|
00:00:44,719 --> 00:00:49,000 |
|
um usually not it's a directional mic so |
|
|
|
21 |
|
00:00:47,440 --> 00:00:51,559 |
|
I try to repeat the questions after |
|
|
|
22 |
|
00:00:49,000 --> 00:00:54,079 |
|
everybody um but uh for the people who |
|
|
|
23 |
|
00:00:51,559 --> 00:00:57,680 |
|
are recordings uh listening to the |
|
|
|
24 |
|
00:00:54,079 --> 00:00:59,320 |
|
recordings um so anyway what is NLP |
|
|
|
25 |
|
00:00:57,680 --> 00:01:03,120 |
|
anyway does anybody have any ideas about |
|
|
|
26 |
|
00:00:59,320 --> 00:01:03,120 |
|
the definition of what NLP might |
|
|
|
27 |
|
00:01:06,119 --> 00:01:09,119 |
|
be |
|
|
|
28 |
|
00:01:15,439 --> 00:01:21,759 |
|
yes okay um it so the answer was it |
|
|
|
29 |
|
00:01:19,240 --> 00:01:25,759 |
|
helps machines understand language |
|
|
|
30 |
|
00:01:21,759 --> 00:01:27,920 |
|
better uh so to facilitate human human |
|
|
|
31 |
|
00:01:25,759 --> 00:01:31,159 |
|
and human machine interactions I think |
|
|
|
32 |
|
00:01:27,920 --> 00:01:32,759 |
|
that's very good um it's |
|
|
|
33 |
|
00:01:31,159 --> 00:01:36,520 |
|
uh similar to what I have written on my |
|
|
|
34 |
|
00:01:32,759 --> 00:01:38,040 |
|
slide here uh but natur in addition to |
|
|
|
35 |
|
00:01:36,520 --> 00:01:41,280 |
|
natural language understanding there's |
|
|
|
36 |
|
00:01:38,040 --> 00:01:46,000 |
|
one major other segment of NLP uh does |
|
|
|
37 |
|
00:01:41,280 --> 00:01:46,000 |
|
anyone uh have an idea what that might |
|
|
|
38 |
|
00:01:48,719 --> 00:01:53,079 |
|
be we often have a dichotomy between two |
|
|
|
39 |
|
00:01:51,399 --> 00:01:55,240 |
|
major segments natural language |
|
|
|
40 |
|
00:01:53,079 --> 00:01:57,520 |
|
understanding and natural language |
|
|
|
41 |
|
00:01:55,240 --> 00:01:59,439 |
|
generation yeah exactly so I I would say |
|
|
|
42 |
|
00:01:57,520 --> 00:02:03,119 |
|
that's almost perfect if you had said |
|
|
|
43 |
|
00:01:59,439 --> 00:02:06,640 |
|
understand and generate so very good um |
|
|
|
44 |
|
00:02:03,119 --> 00:02:08,560 |
|
so I I say natural technology to handle |
|
|
|
45 |
|
00:02:06,640 --> 00:02:11,400 |
|
human language usually text using |
|
|
|
46 |
|
00:02:08,560 --> 00:02:13,200 |
|
computers uh to Aid human machine |
|
|
|
47 |
|
00:02:11,400 --> 00:02:15,480 |
|
communication and this can include |
|
|
|
48 |
|
00:02:13,200 --> 00:02:17,879 |
|
things like question answering dialogue |
|
|
|
49 |
|
00:02:15,480 --> 00:02:20,840 |
|
or generation of code that can be |
|
|
|
50 |
|
00:02:17,879 --> 00:02:23,239 |
|
executed with uh |
|
|
|
51 |
|
00:02:20,840 --> 00:02:25,080 |
|
computers it can also Aid human human |
|
|
|
52 |
|
00:02:23,239 --> 00:02:27,440 |
|
communication and this can include |
|
|
|
53 |
|
00:02:25,080 --> 00:02:30,440 |
|
things like machine translation or spell |
|
|
|
54 |
|
00:02:27,440 --> 00:02:32,640 |
|
checking or assisted writing |
|
|
|
55 |
|
00:02:30,440 --> 00:02:34,560 |
|
and then a final uh segment that people |
|
|
|
56 |
|
00:02:32,640 --> 00:02:37,400 |
|
might think about a little bit less is |
|
|
|
57 |
|
00:02:34,560 --> 00:02:39,400 |
|
analyzing and understanding a language |
|
|
|
58 |
|
00:02:37,400 --> 00:02:42,400 |
|
and this includes things like syntactic |
|
|
|
59 |
|
00:02:39,400 --> 00:02:44,959 |
|
analysis text classification entity |
|
|
|
60 |
|
00:02:42,400 --> 00:02:47,400 |
|
recognition and linking and these can be |
|
|
|
61 |
|
00:02:44,959 --> 00:02:49,159 |
|
used for uh various reasons not |
|
|
|
62 |
|
00:02:47,400 --> 00:02:51,000 |
|
necessarily for direct human machine |
|
|
|
63 |
|
00:02:49,159 --> 00:02:52,720 |
|
communication but also for like |
|
|
|
64 |
|
00:02:51,000 --> 00:02:54,400 |
|
aggregating information across large |
|
|
|
65 |
|
00:02:52,720 --> 00:02:55,760 |
|
things for scientific studies and other |
|
|
|
66 |
|
00:02:54,400 --> 00:02:57,519 |
|
things like that I'll give a few |
|
|
|
67 |
|
00:02:55,760 --> 00:03:00,920 |
|
examples of |
|
|
|
68 |
|
00:02:57,519 --> 00:03:04,040 |
|
this um we now use an many times a day |
|
|
|
69 |
|
00:03:00,920 --> 00:03:06,480 |
|
sometimes without even knowing it so uh |
|
|
|
70 |
|
00:03:04,040 --> 00:03:09,400 |
|
whenever you're typing a doc in Google |
|
|
|
71 |
|
00:03:06,480 --> 00:03:11,599 |
|
Docs there's you know spell checking and |
|
|
|
72 |
|
00:03:09,400 --> 00:03:13,959 |
|
grammar checking going on behind it's |
|
|
|
73 |
|
00:03:11,599 --> 00:03:15,920 |
|
gotten frighten frighteningly good |
|
|
|
74 |
|
00:03:13,959 --> 00:03:18,280 |
|
recently that where it checks like most |
|
|
|
75 |
|
00:03:15,920 --> 00:03:20,720 |
|
of my mistakes and rarely Flags things |
|
|
|
76 |
|
00:03:18,280 --> 00:03:22,799 |
|
that are not mistakes so obviously they |
|
|
|
77 |
|
00:03:20,720 --> 00:03:25,080 |
|
have powerful models running behind that |
|
|
|
78 |
|
00:03:22,799 --> 00:03:25,080 |
|
uh |
|
|
|
79 |
|
00:03:25,640 --> 00:03:33,080 |
|
so and it can do things like answer |
|
|
|
80 |
|
00:03:28,720 --> 00:03:34,599 |
|
questions uh so I asked chat GPT who is |
|
|
|
81 |
|
00:03:33,080 --> 00:03:37,000 |
|
the current president of Carnegie melan |
|
|
|
82 |
|
00:03:34,599 --> 00:03:38,920 |
|
University and chat GPT said I did a |
|
|
|
83 |
|
00:03:37,000 --> 00:03:40,920 |
|
quick search for more information here |
|
|
|
84 |
|
00:03:38,920 --> 00:03:43,439 |
|
is what I found uh the current president |
|
|
|
85 |
|
00:03:40,920 --> 00:03:47,120 |
|
of car Mel University is faram Janan he |
|
|
|
86 |
|
00:03:43,439 --> 00:03:50,040 |
|
has been serving since July 1 etc etc so |
|
|
|
87 |
|
00:03:47,120 --> 00:03:50,040 |
|
as far as I can tell that's |
|
|
|
88 |
|
00:03:50,400 --> 00:03:56,319 |
|
correct um at the same time I asked how |
|
|
|
89 |
|
00:03:53,799 --> 00:04:00,280 |
|
many layers are included in the GP 3.5 |
|
|
|
90 |
|
00:03:56,319 --> 00:04:02,360 |
|
turbo architecture and it said to me |
|
|
|
91 |
|
00:04:00,280 --> 00:04:05,400 |
|
GPT 3.5 turbo which is an optimized |
|
|
|
92 |
|
00:04:02,360 --> 00:04:07,239 |
|
version of GPT 3.5 for faster responses |
|
|
|
93 |
|
00:04:05,400 --> 00:04:08,959 |
|
doesn't have a specific layer art |
|
|
|
94 |
|
00:04:07,239 --> 00:04:11,720 |
|
structure like the traditional gpt3 |
|
|
|
95 |
|
00:04:08,959 --> 00:04:13,560 |
|
models um and I don't know if this is |
|
|
|
96 |
|
00:04:11,720 --> 00:04:16,600 |
|
true or not but I'm pretty sure it's not |
|
|
|
97 |
|
00:04:13,560 --> 00:04:18,840 |
|
true I'm pretty sure that you know GPT |
|
|
|
98 |
|
00:04:16,600 --> 00:04:20,560 |
|
is a model that's much like other models |
|
|
|
99 |
|
00:04:18,840 --> 00:04:21,560 |
|
uh so it basically just made up the spec |
|
|
|
100 |
|
00:04:20,560 --> 00:04:22,880 |
|
because it didn't have any information |
|
|
|
101 |
|
00:04:21,560 --> 00:04:26,000 |
|
on the Internet or couldn't talk about |
|
|
|
102 |
|
00:04:22,880 --> 00:04:26,000 |
|
it so |
|
|
|
103 |
|
00:04:26,120 --> 00:04:33,479 |
|
um another thing is uh NLP can translate |
|
|
|
104 |
|
00:04:29,639 --> 00:04:37,759 |
|
text pretty well so I ran um Google |
|
|
|
105 |
|
00:04:33,479 --> 00:04:39,560 |
|
translate uh on Japanese uh this example |
|
|
|
106 |
|
00:04:37,759 --> 00:04:41,639 |
|
is a little bit old it's from uh you |
|
|
|
107 |
|
00:04:39,560 --> 00:04:44,639 |
|
know a few years ago about Co but I I |
|
|
|
108 |
|
00:04:41,639 --> 00:04:46,240 |
|
retranslated it a few days ago and it |
|
|
|
109 |
|
00:04:44,639 --> 00:04:47,680 |
|
comes up pretty good uh you can |
|
|
|
110 |
|
00:04:46,240 --> 00:04:49,639 |
|
basically understand what's going on |
|
|
|
111 |
|
00:04:47,680 --> 00:04:53,520 |
|
here it's not perfect but you can |
|
|
|
112 |
|
00:04:49,639 --> 00:04:56,400 |
|
understand the uh the general uh |
|
|
|
113 |
|
00:04:53,520 --> 00:04:58,560 |
|
gist at the same time uh if I put in a |
|
|
|
114 |
|
00:04:56,400 --> 00:05:02,280 |
|
relatively low resource language this is |
|
|
|
115 |
|
00:04:58,560 --> 00:05:05,759 |
|
Kurdish um it has a number of problems |
|
|
|
116 |
|
00:05:02,280 --> 00:05:08,199 |
|
when you try to understand it and just |
|
|
|
117 |
|
00:05:05,759 --> 00:05:12,400 |
|
to give an example this is talking about |
|
|
|
118 |
|
00:05:08,199 --> 00:05:14,320 |
|
uh some uh paleontology Discovery it |
|
|
|
119 |
|
00:05:12,400 --> 00:05:15,800 |
|
called this person a fossil scientist |
|
|
|
120 |
|
00:05:14,320 --> 00:05:17,440 |
|
instead of the kind of obvious English |
|
|
|
121 |
|
00:05:15,800 --> 00:05:20,120 |
|
term |
|
|
|
122 |
|
00:05:17,440 --> 00:05:23,520 |
|
paleontologist um and it's talking about |
|
|
|
123 |
|
00:05:20,120 --> 00:05:25,240 |
|
three different uh T-Rex species uh how |
|
|
|
124 |
|
00:05:23,520 --> 00:05:27,039 |
|
T-Rex should actually be split into |
|
|
|
125 |
|
00:05:25,240 --> 00:05:29,639 |
|
three species where T-Rex says king of |
|
|
|
126 |
|
00:05:27,039 --> 00:05:31,560 |
|
ferocious lizards emperator says emperor |
|
|
|
127 |
|
00:05:29,639 --> 00:05:33,720 |
|
of Savaged lizards and then T Regina |
|
|
|
128 |
|
00:05:31,560 --> 00:05:35,120 |
|
means clean of ferocious snail I'm |
|
|
|
129 |
|
00:05:33,720 --> 00:05:37,240 |
|
pretty sure that's not snail I'm pretty |
|
|
|
130 |
|
00:05:35,120 --> 00:05:41,080 |
|
sure that's lizard so uh you can see |
|
|
|
131 |
|
00:05:37,240 --> 00:05:41,080 |
|
that this is not uh this is not perfect |
|
|
|
132 |
|
00:05:41,280 --> 00:05:46,680 |
|
either some people might be thinking why |
|
|
|
133 |
|
00:05:43,960 --> 00:05:48,400 |
|
Google translate and why not GPD well it |
|
|
|
134 |
|
00:05:46,680 --> 00:05:49,960 |
|
turns out um according to one of the |
|
|
|
135 |
|
00:05:48,400 --> 00:05:51,759 |
|
recent studies we've done GPD is even |
|
|
|
136 |
|
00:05:49,960 --> 00:05:55,479 |
|
worse at these slow resource languages |
|
|
|
137 |
|
00:05:51,759 --> 00:05:58,120 |
|
so I use the best thing that's out |
|
|
|
138 |
|
00:05:55,479 --> 00:06:00,440 |
|
there um another thing is language |
|
|
|
139 |
|
00:05:58,120 --> 00:06:02,039 |
|
analysis can Aid scientific ific inquiry |
|
|
|
140 |
|
00:06:00,440 --> 00:06:03,600 |
|
so this is an example that I've been |
|
|
|
141 |
|
00:06:02,039 --> 00:06:06,120 |
|
using for a long time it's actually from |
|
|
|
142 |
|
00:06:03,600 --> 00:06:09,160 |
|
Martin sap another faculty member here |
|
|
|
143 |
|
00:06:06,120 --> 00:06:12,440 |
|
uh but I have been using it since uh |
|
|
|
144 |
|
00:06:09,160 --> 00:06:14,160 |
|
like before he joined and it uh this is |
|
|
|
145 |
|
00:06:12,440 --> 00:06:16,039 |
|
an example from computational social |
|
|
|
146 |
|
00:06:14,160 --> 00:06:18,599 |
|
science uh answering questions about |
|
|
|
147 |
|
00:06:16,039 --> 00:06:20,240 |
|
Society given observational data and |
|
|
|
148 |
|
00:06:18,599 --> 00:06:22,280 |
|
their question was do movie scripts |
|
|
|
149 |
|
00:06:20,240 --> 00:06:24,599 |
|
portray female or male characters with |
|
|
|
150 |
|
00:06:22,280 --> 00:06:27,520 |
|
more power or agency in movie script |
|
|
|
151 |
|
00:06:24,599 --> 00:06:30,120 |
|
films so it's asking kind of a so |
|
|
|
152 |
|
00:06:27,520 --> 00:06:32,160 |
|
societal question by using NLP |
|
|
|
153 |
|
00:06:30,120 --> 00:06:35,360 |
|
technology and the way they did it is |
|
|
|
154 |
|
00:06:32,160 --> 00:06:36,880 |
|
they basically analyzed text trying to |
|
|
|
155 |
|
00:06:35,360 --> 00:06:43,080 |
|
find |
|
|
|
156 |
|
00:06:36,880 --> 00:06:45,280 |
|
uh the uh agents and patients in a a |
|
|
|
157 |
|
00:06:43,080 --> 00:06:46,479 |
|
particular text which are the the things |
|
|
|
158 |
|
00:06:45,280 --> 00:06:49,280 |
|
that are doing things and the things |
|
|
|
159 |
|
00:06:46,479 --> 00:06:52,639 |
|
that things are being done to and you |
|
|
|
160 |
|
00:06:49,280 --> 00:06:54,440 |
|
can see that essentially male characters |
|
|
|
161 |
|
00:06:52,639 --> 00:06:56,560 |
|
in these movie scripts were given more |
|
|
|
162 |
|
00:06:54,440 --> 00:06:58,080 |
|
power in agency and female characters |
|
|
|
163 |
|
00:06:56,560 --> 00:06:59,960 |
|
were given less power in agency and they |
|
|
|
164 |
|
00:06:58,080 --> 00:07:02,680 |
|
were able to do this because they had |
|
|
|
165 |
|
00:06:59,960 --> 00:07:04,840 |
|
NLP technology that analyzed and |
|
|
|
166 |
|
00:07:02,680 --> 00:07:08,960 |
|
extracted useful data and made turned it |
|
|
|
167 |
|
00:07:04,840 --> 00:07:11,520 |
|
into a very easy form to do kind of |
|
|
|
168 |
|
00:07:08,960 --> 00:07:15,840 |
|
analysis of the variety that they want |
|
|
|
169 |
|
00:07:11,520 --> 00:07:17,400 |
|
so um I think that's a major use case of |
|
|
|
170 |
|
00:07:15,840 --> 00:07:19,400 |
|
NLP technology that does language |
|
|
|
171 |
|
00:07:17,400 --> 00:07:20,919 |
|
analysis nowadays turn it into a form |
|
|
|
172 |
|
00:07:19,400 --> 00:07:23,960 |
|
that allows you to very quickly do |
|
|
|
173 |
|
00:07:20,919 --> 00:07:27,440 |
|
aggregate queries and other things like |
|
|
|
174 |
|
00:07:23,960 --> 00:07:30,479 |
|
this um but at the same time uh language |
|
|
|
175 |
|
00:07:27,440 --> 00:07:33,520 |
|
analysis tools fail at very basic tasks |
|
|
|
176 |
|
00:07:30,479 --> 00:07:36,000 |
|
so these are |
|
|
|
177 |
|
00:07:33,520 --> 00:07:38,199 |
|
some things that I ran through a named |
|
|
|
178 |
|
00:07:36,000 --> 00:07:41,080 |
|
entity recognizer and these were kind of |
|
|
|
179 |
|
00:07:38,199 --> 00:07:43,160 |
|
very nice named entity recognizers uh |
|
|
|
180 |
|
00:07:41,080 --> 00:07:46,240 |
|
that a lot of people were using for |
|
|
|
181 |
|
00:07:43,160 --> 00:07:48,039 |
|
example Stanford core NLP and Spacey and |
|
|
|
182 |
|
00:07:46,240 --> 00:07:50,319 |
|
both of them I just threw in the first |
|
|
|
183 |
|
00:07:48,039 --> 00:07:53,120 |
|
thing that I found on the New York Times |
|
|
|
184 |
|
00:07:50,319 --> 00:07:55,199 |
|
at the time and it basically made at |
|
|
|
185 |
|
00:07:53,120 --> 00:07:58,319 |
|
least one mistake in the first sentence |
|
|
|
186 |
|
00:07:55,199 --> 00:08:00,840 |
|
and here it recognizes Baton Rouge as an |
|
|
|
187 |
|
00:07:58,319 --> 00:08:04,720 |
|
organization and here it recognized |
|
|
|
188 |
|
00:08:00,840 --> 00:08:07,000 |
|
hurricane EA as an organization so um |
|
|
|
189 |
|
00:08:04,720 --> 00:08:08,879 |
|
like even uh these things that we expect |
|
|
|
190 |
|
00:08:07,000 --> 00:08:10,360 |
|
should work pretty well make pretty |
|
|
|
191 |
|
00:08:08,879 --> 00:08:13,360 |
|
Solly |
|
|
|
192 |
|
00:08:10,360 --> 00:08:16,199 |
|
mistakes so in the class uh basically |
|
|
|
193 |
|
00:08:13,360 --> 00:08:18,479 |
|
what I want to cover is uh what goes |
|
|
|
194 |
|
00:08:16,199 --> 00:08:20,360 |
|
into building uh state-of-the-art NLP |
|
|
|
195 |
|
00:08:18,479 --> 00:08:24,000 |
|
systems that work really well on a wide |
|
|
|
196 |
|
00:08:20,360 --> 00:08:26,240 |
|
variety of tasks um where do current |
|
|
|
197 |
|
00:08:24,000 --> 00:08:28,840 |
|
systems |
|
|
|
198 |
|
00:08:26,240 --> 00:08:30,479 |
|
fail and how can we make appropriate |
|
|
|
199 |
|
00:08:28,840 --> 00:08:35,000 |
|
improvements and Achieve whatever we |
|
|
|
200 |
|
00:08:30,479 --> 00:08:37,719 |
|
want to do with nalp and this set of |
|
|
|
201 |
|
00:08:35,000 --> 00:08:39,360 |
|
questions that I'm asking here is |
|
|
|
202 |
|
00:08:37,719 --> 00:08:40,919 |
|
exactly the same as the set of questions |
|
|
|
203 |
|
00:08:39,360 --> 00:08:43,519 |
|
that I was asking two years ago before |
|
|
|
204 |
|
00:08:40,919 --> 00:08:45,480 |
|
chat GPT uh I still think they're |
|
|
|
205 |
|
00:08:43,519 --> 00:08:46,920 |
|
important questions but I think the |
|
|
|
206 |
|
00:08:45,480 --> 00:08:48,399 |
|
answers to these questions is very |
|
|
|
207 |
|
00:08:46,920 --> 00:08:50,040 |
|
different and because of that we're |
|
|
|
208 |
|
00:08:48,399 --> 00:08:52,120 |
|
updating the class materials to try to |
|
|
|
209 |
|
00:08:50,040 --> 00:08:54,399 |
|
cover you know the answers to these |
|
|
|
210 |
|
00:08:52,120 --> 00:08:56,000 |
|
questions and uh in kind of the era of |
|
|
|
211 |
|
00:08:54,399 --> 00:08:58,200 |
|
large language models and other things |
|
|
|
212 |
|
00:08:56,000 --> 00:08:59,720 |
|
like |
|
|
|
213 |
|
00:08:58,200 --> 00:09:02,079 |
|
that |
|
|
|
214 |
|
00:08:59,720 --> 00:09:03,360 |
|
so that's all I have for the intro maybe |
|
|
|
215 |
|
00:09:02,079 --> 00:09:06,640 |
|
maybe pretty straightforward are there |
|
|
|
216 |
|
00:09:03,360 --> 00:09:08,480 |
|
any questions or comments so far if not |
|
|
|
217 |
|
00:09:06,640 --> 00:09:14,399 |
|
I'll I'll just go |
|
|
|
218 |
|
00:09:08,480 --> 00:09:17,160 |
|
on okay great so I want to uh first go |
|
|
|
219 |
|
00:09:14,399 --> 00:09:19,480 |
|
into a very high Lev overview of NLP |
|
|
|
220 |
|
00:09:17,160 --> 00:09:20,839 |
|
system building and most of the stuff |
|
|
|
221 |
|
00:09:19,480 --> 00:09:22,399 |
|
that I want to do today is to set the |
|
|
|
222 |
|
00:09:20,839 --> 00:09:24,320 |
|
stage for what I'm going to be talking |
|
|
|
223 |
|
00:09:22,399 --> 00:09:25,040 |
|
about in more detail uh over the rest of |
|
|
|
224 |
|
00:09:24,320 --> 00:09:29,200 |
|
the |
|
|
|
225 |
|
00:09:25,040 --> 00:09:31,720 |
|
class and we could think of NLP syst |
|
|
|
226 |
|
00:09:29,200 --> 00:09:34,040 |
|
systems through this kind of General |
|
|
|
227 |
|
00:09:31,720 --> 00:09:36,560 |
|
framework where we want to create a |
|
|
|
228 |
|
00:09:34,040 --> 00:09:40,600 |
|
function to map an input X into an |
|
|
|
229 |
|
00:09:36,560 --> 00:09:44,440 |
|
output y uh where X and or Y involve |
|
|
|
230 |
|
00:09:40,600 --> 00:09:47,000 |
|
language and uh do some people have |
|
|
|
231 |
|
00:09:44,440 --> 00:09:50,120 |
|
favorite NLP tasks or NLP tasks that you |
|
|
|
232 |
|
00:09:47,000 --> 00:09:52,399 |
|
want to uh want to be handling in some |
|
|
|
233 |
|
00:09:50,120 --> 00:09:57,000 |
|
way or maybe what what do you think are |
|
|
|
234 |
|
00:09:52,399 --> 00:09:57,000 |
|
the most popular and important NLP tasks |
|
|
|
235 |
|
00:09:58,120 --> 00:10:03,200 |
|
nowadays |
|
|
|
236 |
|
00:10:00,800 --> 00:10:06,120 |
|
okay so translation is maybe easy what's |
|
|
|
237 |
|
00:10:03,200 --> 00:10:06,120 |
|
the input and output of |
|
|
|
238 |
|
00:10:11,440 --> 00:10:15,720 |
|
translation okay yeah so uh in |
|
|
|
239 |
|
00:10:13,800 --> 00:10:17,959 |
|
Translation inputs text in one language |
|
|
|
240 |
|
00:10:15,720 --> 00:10:21,760 |
|
output is text in another language and |
|
|
|
241 |
|
00:10:17,959 --> 00:10:21,760 |
|
then what what is a good |
|
|
|
242 |
|
00:10:27,680 --> 00:10:32,160 |
|
translation yeah corre or or the same is |
|
|
|
243 |
|
00:10:30,320 --> 00:10:35,839 |
|
the input basically yes um it also |
|
|
|
244 |
|
00:10:32,160 --> 00:10:37,760 |
|
should be fluent but I agree any other |
|
|
|
245 |
|
00:10:35,839 --> 00:10:39,839 |
|
things generation the reason why I said |
|
|
|
246 |
|
00:10:37,760 --> 00:10:41,519 |
|
it's tough is it's pretty broad um and |
|
|
|
247 |
|
00:10:39,839 --> 00:10:43,360 |
|
it's not like we could be doing |
|
|
|
248 |
|
00:10:41,519 --> 00:10:46,360 |
|
generation with lots of different inputs |
|
|
|
249 |
|
00:10:43,360 --> 00:10:51,440 |
|
but um yeah any any other things maybe a |
|
|
|
250 |
|
00:10:46,360 --> 00:10:51,440 |
|
little bit different yeah like |
|
|
|
251 |
|
00:10:51,480 --> 00:10:55,959 |
|
scenario a scenario and a multiple |
|
|
|
252 |
|
00:10:54,000 --> 00:10:58,200 |
|
choice question about the scenario and |
|
|
|
253 |
|
00:10:55,959 --> 00:10:59,680 |
|
so what would the scenario in the |
|
|
|
254 |
|
00:10:58,200 --> 00:11:01,760 |
|
multiple choice question are probably |
|
|
|
255 |
|
00:10:59,680 --> 00:11:04,040 |
|
the input and then the output |
|
|
|
256 |
|
00:11:01,760 --> 00:11:06,480 |
|
is an answer to the multiple choice |
|
|
|
257 |
|
00:11:04,040 --> 00:11:07,920 |
|
question um and then there it's kind of |
|
|
|
258 |
|
00:11:06,480 --> 00:11:12,279 |
|
obvious like what is good it's the |
|
|
|
259 |
|
00:11:07,920 --> 00:11:14,880 |
|
correct answer sure um interestingly I |
|
|
|
260 |
|
00:11:12,279 --> 00:11:17,440 |
|
think a lot of llm evaluation is done on |
|
|
|
261 |
|
00:11:14,880 --> 00:11:21,160 |
|
these multiple choice questions but I'm |
|
|
|
262 |
|
00:11:17,440 --> 00:11:22,320 |
|
yet to encounter an actual application |
|
|
|
263 |
|
00:11:21,160 --> 00:11:24,880 |
|
that cares about multiple choice |
|
|
|
264 |
|
00:11:22,320 --> 00:11:26,880 |
|
question answering so uh there's kind of |
|
|
|
265 |
|
00:11:24,880 --> 00:11:30,959 |
|
a funny disconnect there but uh yeah I |
|
|
|
266 |
|
00:11:26,880 --> 00:11:33,519 |
|
saw hand that think about V search comp |
|
|
|
267 |
|
00:11:30,959 --> 00:11:36,360 |
|
yeah Vector search uh that's very good |
|
|
|
268 |
|
00:11:33,519 --> 00:11:36,360 |
|
so the input |
|
|
|
269 |
|
00:11:37,120 --> 00:11:45,000 |
|
is can con it into or understanding and |
|
|
|
270 |
|
00:11:42,560 --> 00:11:45,000 |
|
it to |
|
|
|
271 |
|
00:11:47,360 --> 00:11:53,760 |
|
another okay yeah so I'd say the input |
|
|
|
272 |
|
00:11:49,880 --> 00:11:56,160 |
|
there is a query and a document base um |
|
|
|
273 |
|
00:11:53,760 --> 00:11:57,959 |
|
and then the output is maybe an index |
|
|
|
274 |
|
00:11:56,160 --> 00:11:59,800 |
|
into the document or or something else |
|
|
|
275 |
|
00:11:57,959 --> 00:12:01,279 |
|
like that sure um and then something |
|
|
|
276 |
|
00:11:59,800 --> 00:12:05,040 |
|
that's good here here's a good question |
|
|
|
277 |
|
00:12:01,279 --> 00:12:05,040 |
|
what what's a good result from |
|
|
|
278 |
|
00:12:06,560 --> 00:12:10,200 |
|
that what's a good |
|
|
|
279 |
|
00:12:10,839 --> 00:12:19,279 |
|
output be sort of simar the major |
|
|
|
280 |
|
00:12:15,560 --> 00:12:21,680 |
|
problem there I see is how you def SAR |
|
|
|
281 |
|
00:12:19,279 --> 00:12:26,199 |
|
and how you |
|
|
|
282 |
|
00:12:21,680 --> 00:12:29,760 |
|
a always like you understand |
|
|
|
283 |
|
00:12:26,199 --> 00:12:33,000 |
|
whether is actually |
|
|
|
284 |
|
00:12:29,760 --> 00:12:35,079 |
|
yeah exactly so that um just to repeat |
|
|
|
285 |
|
00:12:33,000 --> 00:12:36,880 |
|
it's like uh we need to have a |
|
|
|
286 |
|
00:12:35,079 --> 00:12:38,399 |
|
similarity a good similarity metric we |
|
|
|
287 |
|
00:12:36,880 --> 00:12:40,120 |
|
need to have a good threshold where we |
|
|
|
288 |
|
00:12:38,399 --> 00:12:41,760 |
|
get like the ones we want and we don't |
|
|
|
289 |
|
00:12:40,120 --> 00:12:43,240 |
|
get the ones we don't want we're going |
|
|
|
290 |
|
00:12:41,760 --> 00:12:44,959 |
|
to talk more about that in the retrieval |
|
|
|
291 |
|
00:12:43,240 --> 00:12:48,440 |
|
lecture exactly how we evaluate and |
|
|
|
292 |
|
00:12:44,959 --> 00:12:49,920 |
|
stuff but um yeah good so this is a good |
|
|
|
293 |
|
00:12:48,440 --> 00:12:53,279 |
|
uh here are some good examples I have |
|
|
|
294 |
|
00:12:49,920 --> 00:12:55,519 |
|
some examples of my own um the first one |
|
|
|
295 |
|
00:12:53,279 --> 00:12:58,360 |
|
is uh kind of the very generic one maybe |
|
|
|
296 |
|
00:12:55,519 --> 00:13:00,800 |
|
kind of like generation here but text in |
|
|
|
297 |
|
00:12:58,360 --> 00:13:02,959 |
|
continuing text uh so this is language |
|
|
|
298 |
|
00:13:00,800 --> 00:13:04,160 |
|
modeling so you have a text and then you |
|
|
|
299 |
|
00:13:02,959 --> 00:13:05,440 |
|
have the continuation you want to |
|
|
|
300 |
|
00:13:04,160 --> 00:13:07,680 |
|
predict the |
|
|
|
301 |
|
00:13:05,440 --> 00:13:10,480 |
|
continuation um text and text in another |
|
|
|
302 |
|
00:13:07,680 --> 00:13:13,040 |
|
language is translation uh text in a |
|
|
|
303 |
|
00:13:10,480 --> 00:13:15,800 |
|
label could be text classification uh |
|
|
|
304 |
|
00:13:13,040 --> 00:13:17,760 |
|
text in linguistic structure or uh some |
|
|
|
305 |
|
00:13:15,800 --> 00:13:21,360 |
|
s kind of entities or something like |
|
|
|
306 |
|
00:13:17,760 --> 00:13:22,680 |
|
that could be uh language analysis or um |
|
|
|
307 |
|
00:13:21,360 --> 00:13:24,839 |
|
information |
|
|
|
308 |
|
00:13:22,680 --> 00:13:29,440 |
|
extraction uh we could also have image |
|
|
|
309 |
|
00:13:24,839 --> 00:13:31,320 |
|
and text uh which is image captioning um |
|
|
|
310 |
|
00:13:29,440 --> 00:13:33,560 |
|
or speech and text which is speech |
|
|
|
311 |
|
00:13:31,320 --> 00:13:35,240 |
|
recognition and I take the very broad |
|
|
|
312 |
|
00:13:33,560 --> 00:13:38,000 |
|
view of natural language processing |
|
|
|
313 |
|
00:13:35,240 --> 00:13:39,519 |
|
which is if it's any variety of language |
|
|
|
314 |
|
00:13:38,000 --> 00:13:41,519 |
|
uh if you're handling language in some |
|
|
|
315 |
|
00:13:39,519 --> 00:13:42,800 |
|
way it's natural language processing it |
|
|
|
316 |
|
00:13:41,519 --> 00:13:45,880 |
|
doesn't necessarily have to be text |
|
|
|
317 |
|
00:13:42,800 --> 00:13:47,480 |
|
input text output um so that's relevant |
|
|
|
318 |
|
00:13:45,880 --> 00:13:50,199 |
|
for the projects that you're thinking |
|
|
|
319 |
|
00:13:47,480 --> 00:13:52,160 |
|
about too at the end of this course so |
|
|
|
320 |
|
00:13:50,199 --> 00:13:55,519 |
|
the the most common FAQ for this course |
|
|
|
321 |
|
00:13:52,160 --> 00:13:57,839 |
|
is does my project count and if you're |
|
|
|
322 |
|
00:13:55,519 --> 00:13:59,360 |
|
uncertain you should ask but usually |
|
|
|
323 |
|
00:13:57,839 --> 00:14:01,040 |
|
like if it has some sort of language |
|
|
|
324 |
|
00:13:59,360 --> 00:14:05,079 |
|
involved then I'll usually say yes it |
|
|
|
325 |
|
00:14:01,040 --> 00:14:07,920 |
|
does kind so um if it's like uh code to |
|
|
|
326 |
|
00:14:05,079 --> 00:14:09,680 |
|
code there that's not code is not |
|
|
|
327 |
|
00:14:07,920 --> 00:14:11,480 |
|
natural language it is language but it's |
|
|
|
328 |
|
00:14:09,680 --> 00:14:13,000 |
|
not natural language so that might be |
|
|
|
329 |
|
00:14:11,480 --> 00:14:15,320 |
|
borderline we might have to discuss |
|
|
|
330 |
|
00:14:13,000 --> 00:14:15,320 |
|
about |
|
|
|
331 |
|
00:14:15,759 --> 00:14:21,800 |
|
that cool um so next I'd like to talk |
|
|
|
332 |
|
00:14:18,880 --> 00:14:25,240 |
|
about methods for creating NLP systems |
|
|
|
333 |
|
00:14:21,800 --> 00:14:27,839 |
|
um and there's a lot of different ways |
|
|
|
334 |
|
00:14:25,240 --> 00:14:29,720 |
|
to create MLP systems all of these are |
|
|
|
335 |
|
00:14:27,839 --> 00:14:32,880 |
|
alive and well in |
|
|
|
336 |
|
00:14:29,720 --> 00:14:35,759 |
|
2024 uh the first one is Rule uh |
|
|
|
337 |
|
00:14:32,880 --> 00:14:37,959 |
|
rule-based system creation and so the |
|
|
|
338 |
|
00:14:35,759 --> 00:14:40,399 |
|
way this works is like let's say you |
|
|
|
339 |
|
00:14:37,959 --> 00:14:42,480 |
|
want to build a text classifier you just |
|
|
|
340 |
|
00:14:40,399 --> 00:14:46,560 |
|
write the simple python function that |
|
|
|
341 |
|
00:14:42,480 --> 00:14:48,639 |
|
classifies things into uh sports or |
|
|
|
342 |
|
00:14:46,560 --> 00:14:50,240 |
|
other and the way it classifies it into |
|
|
|
343 |
|
00:14:48,639 --> 00:14:52,959 |
|
sports or other is it checks whether |
|
|
|
344 |
|
00:14:50,240 --> 00:14:55,160 |
|
baseball soccer football and Tennis are |
|
|
|
345 |
|
00:14:52,959 --> 00:14:59,399 |
|
included in the document and classifies |
|
|
|
346 |
|
00:14:55,160 --> 00:15:01,959 |
|
it into uh Sports if so uh other if not |
|
|
|
347 |
|
00:14:59,399 --> 00:15:05,279 |
|
so has anyone written something like |
|
|
|
348 |
|
00:15:01,959 --> 00:15:09,720 |
|
this maybe not a text classifier but um |
|
|
|
349 |
|
00:15:05,279 --> 00:15:11,880 |
|
you know to identify entities or uh |
|
|
|
350 |
|
00:15:09,720 --> 00:15:14,279 |
|
split words |
|
|
|
351 |
|
00:15:11,880 --> 00:15:16,680 |
|
or something like |
|
|
|
352 |
|
00:15:14,279 --> 00:15:18,399 |
|
that has anybody not ever written |
|
|
|
353 |
|
00:15:16,680 --> 00:15:22,800 |
|
anything like |
|
|
|
354 |
|
00:15:18,399 --> 00:15:24,639 |
|
this yeah that's what I thought so um |
|
|
|
355 |
|
00:15:22,800 --> 00:15:26,079 |
|
rule-based systems are very convenient |
|
|
|
356 |
|
00:15:24,639 --> 00:15:28,920 |
|
when you don't really care about how |
|
|
|
357 |
|
00:15:26,079 --> 00:15:30,759 |
|
good your system is um or you're doing |
|
|
|
358 |
|
00:15:28,920 --> 00:15:32,360 |
|
that's really really simple and like |
|
|
|
359 |
|
00:15:30,759 --> 00:15:35,600 |
|
it'll be perfect even if you do the very |
|
|
|
360 |
|
00:15:32,360 --> 00:15:37,079 |
|
simple thing and so I I think it's worth |
|
|
|
361 |
|
00:15:35,600 --> 00:15:39,959 |
|
talking a little bit about them and I'll |
|
|
|
362 |
|
00:15:37,079 --> 00:15:43,319 |
|
talk a little bit about that uh this |
|
|
|
363 |
|
00:15:39,959 --> 00:15:45,680 |
|
time the second thing which like very |
|
|
|
364 |
|
00:15:43,319 --> 00:15:47,680 |
|
rapidly over the course of maybe three |
|
|
|
365 |
|
00:15:45,680 --> 00:15:50,279 |
|
years or so has become actually maybe |
|
|
|
366 |
|
00:15:47,680 --> 00:15:52,720 |
|
the dominant Paradigm in NLP is |
|
|
|
367 |
|
00:15:50,279 --> 00:15:56,360 |
|
prompting uh in prompting a language |
|
|
|
368 |
|
00:15:52,720 --> 00:15:58,560 |
|
model and the way this works is uh you |
|
|
|
369 |
|
00:15:56,360 --> 00:16:00,720 |
|
ask a language model if the following |
|
|
|
370 |
|
00:15:58,560 --> 00:16:03,079 |
|
sent is about sports reply Sports |
|
|
|
371 |
|
00:16:00,720 --> 00:16:06,120 |
|
otherwise reply other and you feed it to |
|
|
|
372 |
|
00:16:03,079 --> 00:16:08,480 |
|
your favorite LM uh usually that's GPT |
|
|
|
373 |
|
00:16:06,120 --> 00:16:11,399 |
|
something or other uh sometimes it's an |
|
|
|
374 |
|
00:16:08,480 --> 00:16:14,440 |
|
open source model of some variety and |
|
|
|
375 |
|
00:16:11,399 --> 00:16:17,759 |
|
then uh it will give you the |
|
|
|
376 |
|
00:16:14,440 --> 00:16:20,639 |
|
answer and then finally uh fine-tuning |
|
|
|
377 |
|
00:16:17,759 --> 00:16:22,240 |
|
uh so you take some paired data and you |
|
|
|
378 |
|
00:16:20,639 --> 00:16:23,600 |
|
do machine learning from paired data |
|
|
|
379 |
|
00:16:22,240 --> 00:16:25,680 |
|
where you have something like I love to |
|
|
|
380 |
|
00:16:23,600 --> 00:16:27,440 |
|
play baseball uh the stock price is |
|
|
|
381 |
|
00:16:25,680 --> 00:16:29,519 |
|
going up he got a hatrick yesterday he |
|
|
|
382 |
|
00:16:27,440 --> 00:16:32,759 |
|
is wearing tennis shoes and you assign |
|
|
|
383 |
|
00:16:29,519 --> 00:16:35,319 |
|
all these uh labels to them training a |
|
|
|
384 |
|
00:16:32,759 --> 00:16:38,160 |
|
model and you can even start out with a |
|
|
|
385 |
|
00:16:35,319 --> 00:16:41,480 |
|
prompting based model and fine-tune a a |
|
|
|
386 |
|
00:16:38,160 --> 00:16:41,480 |
|
language model |
|
|
|
387 |
|
00:16:42,920 --> 00:16:49,399 |
|
also so one major consideration when |
|
|
|
388 |
|
00:16:47,519 --> 00:16:52,000 |
|
you're Building Systems like this is the |
|
|
|
389 |
|
00:16:49,399 --> 00:16:56,440 |
|
data requirements for building such a |
|
|
|
390 |
|
00:16:52,000 --> 00:16:59,319 |
|
system and for rules or prompting where |
|
|
|
391 |
|
00:16:56,440 --> 00:17:02,240 |
|
it's just based on intuition really no |
|
|
|
392 |
|
00:16:59,319 --> 00:17:04,640 |
|
data is needed whatsoever it you don't |
|
|
|
393 |
|
00:17:02,240 --> 00:17:08,240 |
|
need a single example and you can start |
|
|
|
394 |
|
00:17:04,640 --> 00:17:11,000 |
|
writing rules or like just just to give |
|
|
|
395 |
|
00:17:08,240 --> 00:17:12,640 |
|
an example the rules and prompts I wrote |
|
|
|
396 |
|
00:17:11,000 --> 00:17:14,679 |
|
here I didn't look at any examples and I |
|
|
|
397 |
|
00:17:12,640 --> 00:17:17,240 |
|
just wrote them uh so this is something |
|
|
|
398 |
|
00:17:14,679 --> 00:17:20,000 |
|
that you could start out |
|
|
|
399 |
|
00:17:17,240 --> 00:17:21,559 |
|
with uh the problem is you also have no |
|
|
|
400 |
|
00:17:20,000 --> 00:17:24,720 |
|
idea how well it works if you don't have |
|
|
|
401 |
|
00:17:21,559 --> 00:17:26,760 |
|
any data whatsoever right so um you'll |
|
|
|
402 |
|
00:17:24,720 --> 00:17:30,400 |
|
you might be in trouble if you think |
|
|
|
403 |
|
00:17:26,760 --> 00:17:30,400 |
|
something should be working |
|
|
|
404 |
|
00:17:30,919 --> 00:17:34,440 |
|
so normally the next thing that people |
|
|
|
405 |
|
00:17:32,919 --> 00:17:36,880 |
|
move to nowadays when they're building |
|
|
|
406 |
|
00:17:34,440 --> 00:17:39,559 |
|
practical systems is rules are prompting |
|
|
|
407 |
|
00:17:36,880 --> 00:17:41,240 |
|
based on spot checks so that basically |
|
|
|
408 |
|
00:17:39,559 --> 00:17:42,919 |
|
means that you start out with a |
|
|
|
409 |
|
00:17:41,240 --> 00:17:45,840 |
|
rule-based system or a prompting based |
|
|
|
410 |
|
00:17:42,919 --> 00:17:47,240 |
|
system and then you go in and you run it |
|
|
|
411 |
|
00:17:45,840 --> 00:17:48,720 |
|
on some data that you're interested in |
|
|
|
412 |
|
00:17:47,240 --> 00:17:50,799 |
|
you just kind of qualitatively look at |
|
|
|
413 |
|
00:17:48,720 --> 00:17:52,160 |
|
the data and say oh it's messing up here |
|
|
|
414 |
|
00:17:50,799 --> 00:17:53,440 |
|
then you go in and fix your prompt a |
|
|
|
415 |
|
00:17:52,160 --> 00:17:54,919 |
|
little bit or you go in and fix your |
|
|
|
416 |
|
00:17:53,440 --> 00:17:57,320 |
|
rules a little bit or something like |
|
|
|
417 |
|
00:17:54,919 --> 00:18:00,400 |
|
that so uh this is kind of the second |
|
|
|
418 |
|
00:17:57,320 --> 00:18:00,400 |
|
level of difficulty |
|
|
|
419 |
|
00:18:01,400 --> 00:18:04,640 |
|
so the third level of difficulty would |
|
|
|
420 |
|
00:18:03,159 --> 00:18:07,400 |
|
be something like rules are prompting |
|
|
|
421 |
|
00:18:04,640 --> 00:18:09,039 |
|
with rigorous evaluation and so here you |
|
|
|
422 |
|
00:18:07,400 --> 00:18:12,840 |
|
would create a development set with |
|
|
|
423 |
|
00:18:09,039 --> 00:18:14,840 |
|
inputs and outputs uh so you uh create |
|
|
|
424 |
|
00:18:12,840 --> 00:18:17,039 |
|
maybe 200 to 2,000 |
|
|
|
425 |
|
00:18:14,840 --> 00:18:20,080 |
|
examples um |
|
|
|
426 |
|
00:18:17,039 --> 00:18:21,720 |
|
and then evaluate your actual accuracy |
|
|
|
427 |
|
00:18:20,080 --> 00:18:23,880 |
|
so you need an evaluation metric you |
|
|
|
428 |
|
00:18:21,720 --> 00:18:26,120 |
|
need other things like this this is the |
|
|
|
429 |
|
00:18:23,880 --> 00:18:28,400 |
|
next level of difficulty but if you're |
|
|
|
430 |
|
00:18:26,120 --> 00:18:30,240 |
|
going to be a serious you know NLP |
|
|
|
431 |
|
00:18:28,400 --> 00:18:33,000 |
|
engineer or something like this you |
|
|
|
432 |
|
00:18:30,240 --> 00:18:34,720 |
|
definitely will be doing this a lot I |
|
|
|
433 |
|
00:18:33,000 --> 00:18:37,760 |
|
feel and |
|
|
|
434 |
|
00:18:34,720 --> 00:18:40,360 |
|
then so that here now you start needing |
|
|
|
435 |
|
00:18:37,760 --> 00:18:41,960 |
|
a depth set and a test set and then |
|
|
|
436 |
|
00:18:40,360 --> 00:18:46,280 |
|
finally fine-tuning you need an |
|
|
|
437 |
|
00:18:41,960 --> 00:18:48,480 |
|
additional training set um and uh this |
|
|
|
438 |
|
00:18:46,280 --> 00:18:52,240 |
|
will generally be a lot bigger than 200 |
|
|
|
439 |
|
00:18:48,480 --> 00:18:56,080 |
|
to 2,000 examples and generally the rule |
|
|
|
440 |
|
00:18:52,240 --> 00:18:56,080 |
|
is that every time you |
|
|
|
441 |
|
00:18:57,320 --> 00:19:01,080 |
|
double |
|
|
|
442 |
|
00:18:59,520 --> 00:19:02,400 |
|
every time you double your training set |
|
|
|
443 |
|
00:19:01,080 --> 00:19:07,480 |
|
size you get about a constant |
|
|
|
444 |
|
00:19:02,400 --> 00:19:07,480 |
|
Improvement so if you start |
|
|
|
445 |
|
00:19:07,799 --> 00:19:15,080 |
|
out if you start out down here with |
|
|
|
446 |
|
00:19:12,240 --> 00:19:17,039 |
|
um zero shot accuracy with a language |
|
|
|
447 |
|
00:19:15,080 --> 00:19:21,559 |
|
model you you create a small printing |
|
|
|
448 |
|
00:19:17,039 --> 00:19:21,559 |
|
set and you get you know a pretty big |
|
|
|
449 |
|
00:19:22,000 --> 00:19:29,120 |
|
increase and then every time you double |
|
|
|
450 |
|
00:19:26,320 --> 00:19:30,799 |
|
it it increases by constant fact it's |
|
|
|
451 |
|
00:19:29,120 --> 00:19:32,480 |
|
kind of like just in general in machine |
|
|
|
452 |
|
00:19:30,799 --> 00:19:37,360 |
|
learning this is a trend that we tend to |
|
|
|
453 |
|
00:19:32,480 --> 00:19:40,679 |
|
see so um So based on this |
|
|
|
454 |
|
00:19:37,360 --> 00:19:41,880 |
|
uh there's kind of like you get a big |
|
|
|
455 |
|
00:19:40,679 --> 00:19:44,200 |
|
gain from having a little bit of |
|
|
|
456 |
|
00:19:41,880 --> 00:19:45,760 |
|
training data but the gains very quickly |
|
|
|
457 |
|
00:19:44,200 --> 00:19:48,919 |
|
drop off and you start spending a lot of |
|
|
|
458 |
|
00:19:45,760 --> 00:19:48,919 |
|
time annotating |
|
|
|
459 |
|
00:19:51,000 --> 00:19:55,880 |
|
an so um yeah this is the the general |
|
|
|
460 |
|
00:19:54,760 --> 00:19:58,280 |
|
overview of the different types of |
|
|
|
461 |
|
00:19:55,880 --> 00:20:00,000 |
|
system building uh any any question |
|
|
|
462 |
|
00:19:58,280 --> 00:20:01,559 |
|
questions about this or comments or |
|
|
|
463 |
|
00:20:00,000 --> 00:20:04,000 |
|
things like |
|
|
|
464 |
|
00:20:01,559 --> 00:20:05,840 |
|
this I think one thing that's changed |
|
|
|
465 |
|
00:20:04,000 --> 00:20:08,159 |
|
really drastically from the last time I |
|
|
|
466 |
|
00:20:05,840 --> 00:20:09,600 |
|
taught this class is the fact that |
|
|
|
467 |
|
00:20:08,159 --> 00:20:11,000 |
|
number one and number two are the things |
|
|
|
468 |
|
00:20:09,600 --> 00:20:13,799 |
|
that people are actually doing in |
|
|
|
469 |
|
00:20:11,000 --> 00:20:15,360 |
|
practice uh which was you know people |
|
|
|
470 |
|
00:20:13,799 --> 00:20:16,679 |
|
who actually care about systems are |
|
|
|
471 |
|
00:20:15,360 --> 00:20:18,880 |
|
doing number one and number two is the |
|
|
|
472 |
|
00:20:16,679 --> 00:20:20,440 |
|
main thing it used to be that if you |
|
|
|
473 |
|
00:20:18,880 --> 00:20:22,679 |
|
were actually serious about building a |
|
|
|
474 |
|
00:20:20,440 --> 00:20:24,320 |
|
system uh you really needed to do the |
|
|
|
475 |
|
00:20:22,679 --> 00:20:27,080 |
|
funing and now it's kind of like more |
|
|
|
476 |
|
00:20:24,320 --> 00:20:27,080 |
|
optional |
|
|
|
477 |
|
00:20:27,159 --> 00:20:30,159 |
|
so |
|
|
|
478 |
|
00:20:44,039 --> 00:20:50,960 |
|
yeah |
|
|
|
479 |
|
00:20:46,320 --> 00:20:53,960 |
|
so it's it's definitely an empirical |
|
|
|
480 |
|
00:20:50,960 --> 00:20:53,960 |
|
observation |
|
|
|
481 |
|
00:20:54,720 --> 00:21:01,080 |
|
um in terms of the theoretical |
|
|
|
482 |
|
00:20:57,640 --> 00:21:03,120 |
|
background I am not I can't immediately |
|
|
|
483 |
|
00:21:01,080 --> 00:21:05,840 |
|
point to a |
|
|
|
484 |
|
00:21:03,120 --> 00:21:10,039 |
|
particular paper that does that but I |
|
|
|
485 |
|
00:21:05,840 --> 00:21:12,720 |
|
think if you think about |
|
|
|
486 |
|
00:21:10,039 --> 00:21:14,720 |
|
the I I think I have seen that they do |
|
|
|
487 |
|
00:21:12,720 --> 00:21:17,039 |
|
exist in the past but I I can't think of |
|
|
|
488 |
|
00:21:14,720 --> 00:21:19,000 |
|
it right now I can try to uh try to come |
|
|
|
489 |
|
00:21:17,039 --> 00:21:23,720 |
|
up with an example of |
|
|
|
490 |
|
00:21:19,000 --> 00:21:23,720 |
|
that so yeah I I should take |
|
|
|
491 |
|
00:21:26,799 --> 00:21:31,960 |
|
notes or someone wants to share one on |
|
|
|
492 |
|
00:21:29,360 --> 00:21:33,360 |
|
Piaza uh if you have any ideas and want |
|
|
|
493 |
|
00:21:31,960 --> 00:21:34,520 |
|
to share on Patza I'm sure that would be |
|
|
|
494 |
|
00:21:33,360 --> 00:21:35,640 |
|
great it'd be great to have a discussion |
|
|
|
495 |
|
00:21:34,520 --> 00:21:39,320 |
|
on |
|
|
|
496 |
|
00:21:35,640 --> 00:21:44,960 |
|
Patza um Pi |
|
|
|
497 |
|
00:21:39,320 --> 00:21:46,880 |
|
one cool okay so next I want to try to |
|
|
|
498 |
|
00:21:44,960 --> 00:21:48,200 |
|
make a rule-based system and I'm going |
|
|
|
499 |
|
00:21:46,880 --> 00:21:49,360 |
|
to make a rule-based system for |
|
|
|
500 |
|
00:21:48,200 --> 00:21:51,799 |
|
sentiment |
|
|
|
501 |
|
00:21:49,360 --> 00:21:53,480 |
|
analysis uh and this is a bad idea I |
|
|
|
502 |
|
00:21:51,799 --> 00:21:55,400 |
|
would not encourage you to ever do this |
|
|
|
503 |
|
00:21:53,480 --> 00:21:57,440 |
|
in real life but I want to do it here to |
|
|
|
504 |
|
00:21:55,400 --> 00:21:59,640 |
|
show you why it's a bad idea and like |
|
|
|
505 |
|
00:21:57,440 --> 00:22:01,200 |
|
what are some of the hard problems that |
|
|
|
506 |
|
00:21:59,640 --> 00:22:03,960 |
|
you encounter when trying to create a |
|
|
|
507 |
|
00:22:01,200 --> 00:22:06,600 |
|
system based on rules |
|
|
|
508 |
|
00:22:03,960 --> 00:22:08,080 |
|
and then we'll move into building a |
|
|
|
509 |
|
00:22:06,600 --> 00:22:12,360 |
|
machine learning base system after we |
|
|
|
510 |
|
00:22:08,080 --> 00:22:15,400 |
|
finish this so if we look at the example |
|
|
|
511 |
|
00:22:12,360 --> 00:22:18,559 |
|
test this is review sentiment analysis |
|
|
|
512 |
|
00:22:15,400 --> 00:22:21,799 |
|
it's one of the most valuable uh tasks |
|
|
|
513 |
|
00:22:18,559 --> 00:22:24,039 |
|
uh that people do in NLP nowadays |
|
|
|
514 |
|
00:22:21,799 --> 00:22:26,400 |
|
because it allows people to know how |
|
|
|
515 |
|
00:22:24,039 --> 00:22:29,200 |
|
customers are thinking about products uh |
|
|
|
516 |
|
00:22:26,400 --> 00:22:30,799 |
|
improve their you know their product |
|
|
|
517 |
|
00:22:29,200 --> 00:22:32,919 |
|
development and other things like that |
|
|
|
518 |
|
00:22:30,799 --> 00:22:34,799 |
|
may monitor people's you know |
|
|
|
519 |
|
00:22:32,919 --> 00:22:36,760 |
|
satisfaction with their social media |
|
|
|
520 |
|
00:22:34,799 --> 00:22:39,200 |
|
service other things like this so |
|
|
|
521 |
|
00:22:36,760 --> 00:22:42,720 |
|
basically the way it works is um you |
|
|
|
522 |
|
00:22:39,200 --> 00:22:44,400 |
|
have uh outputs or you have sentences |
|
|
|
523 |
|
00:22:42,720 --> 00:22:46,720 |
|
inputs like I hate this movie I love |
|
|
|
524 |
|
00:22:44,400 --> 00:22:48,520 |
|
this movie I saw this movie and this |
|
|
|
525 |
|
00:22:46,720 --> 00:22:50,600 |
|
gets mapped into positive neutral or |
|
|
|
526 |
|
00:22:48,520 --> 00:22:53,120 |
|
negative so I hate this movie would be |
|
|
|
527 |
|
00:22:50,600 --> 00:22:55,480 |
|
negative I love this movie positive and |
|
|
|
528 |
|
00:22:53,120 --> 00:22:59,039 |
|
I saw this movie is |
|
|
|
529 |
|
00:22:55,480 --> 00:23:01,200 |
|
neutral so um |
|
|
|
530 |
|
00:22:59,039 --> 00:23:05,200 |
|
that that's the task input tax output |
|
|
|
531 |
|
00:23:01,200 --> 00:23:08,880 |
|
labels uh Kary uh sentence |
|
|
|
532 |
|
00:23:05,200 --> 00:23:11,679 |
|
label and in order to do this uh we |
|
|
|
533 |
|
00:23:08,880 --> 00:23:13,120 |
|
would like to build a model um and we're |
|
|
|
534 |
|
00:23:11,679 --> 00:23:16,159 |
|
going to build the model in a rule based |
|
|
|
535 |
|
00:23:13,120 --> 00:23:19,000 |
|
way but it we'll still call it a model |
|
|
|
536 |
|
00:23:16,159 --> 00:23:21,600 |
|
and the way it works is we do feature |
|
|
|
537 |
|
00:23:19,000 --> 00:23:23,159 |
|
extraction um so we extract the Salient |
|
|
|
538 |
|
00:23:21,600 --> 00:23:25,279 |
|
features for making the decision about |
|
|
|
539 |
|
00:23:23,159 --> 00:23:27,320 |
|
what to Output next we do score |
|
|
|
540 |
|
00:23:25,279 --> 00:23:29,880 |
|
calculation calculate a score for one or |
|
|
|
541 |
|
00:23:27,320 --> 00:23:32,320 |
|
more possib ities and we have a decision |
|
|
|
542 |
|
00:23:29,880 --> 00:23:33,520 |
|
function so we choose one of those |
|
|
|
543 |
|
00:23:32,320 --> 00:23:37,679 |
|
several |
|
|
|
544 |
|
00:23:33,520 --> 00:23:40,120 |
|
possibilities and so for feature |
|
|
|
545 |
|
00:23:37,679 --> 00:23:42,200 |
|
extraction uh formally what this looks |
|
|
|
546 |
|
00:23:40,120 --> 00:23:44,240 |
|
like is we have some function and it |
|
|
|
547 |
|
00:23:42,200 --> 00:23:48,039 |
|
extracts a feature |
|
|
|
548 |
|
00:23:44,240 --> 00:23:51,159 |
|
Vector for score calculation um we |
|
|
|
549 |
|
00:23:48,039 --> 00:23:54,240 |
|
calculate the scores based on either a |
|
|
|
550 |
|
00:23:51,159 --> 00:23:56,279 |
|
binary classification uh where we have a |
|
|
|
551 |
|
00:23:54,240 --> 00:23:58,279 |
|
a weight vector and we take the dot |
|
|
|
552 |
|
00:23:56,279 --> 00:24:00,120 |
|
product with our feature vector or we |
|
|
|
553 |
|
00:23:58,279 --> 00:24:02,480 |
|
have multi class classification where we |
|
|
|
554 |
|
00:24:00,120 --> 00:24:04,520 |
|
have a weight Matrix and we take the |
|
|
|
555 |
|
00:24:02,480 --> 00:24:08,640 |
|
product with uh the vector and that |
|
|
|
556 |
|
00:24:04,520 --> 00:24:08,640 |
|
gives us you know squares over multiple |
|
|
|
557 |
|
00:24:08,919 --> 00:24:14,840 |
|
classes and then we have a decision uh |
|
|
|
558 |
|
00:24:11,600 --> 00:24:17,520 |
|
rule so this decision rule tells us what |
|
|
|
559 |
|
00:24:14,840 --> 00:24:20,080 |
|
the output is going to be um does anyone |
|
|
|
560 |
|
00:24:17,520 --> 00:24:22,200 |
|
know what a typical decision rule is |
|
|
|
561 |
|
00:24:20,080 --> 00:24:24,520 |
|
maybe maybe so obvious that you don't |
|
|
|
562 |
|
00:24:22,200 --> 00:24:28,760 |
|
think about it often |
|
|
|
563 |
|
00:24:24,520 --> 00:24:31,000 |
|
but uh a threshold um so like for would |
|
|
|
564 |
|
00:24:28,760 --> 00:24:34,440 |
|
that be for binary a single binary |
|
|
|
565 |
|
00:24:31,000 --> 00:24:37,000 |
|
scaler score or a multiple |
|
|
|
566 |
|
00:24:34,440 --> 00:24:38,520 |
|
class binary yeah so and then you would |
|
|
|
567 |
|
00:24:37,000 --> 00:24:39,960 |
|
pick a threshold and if it's over the |
|
|
|
568 |
|
00:24:38,520 --> 00:24:42,919 |
|
threshold |
|
|
|
569 |
|
00:24:39,960 --> 00:24:45,760 |
|
you say yes and if it's under the |
|
|
|
570 |
|
00:24:42,919 --> 00:24:50,279 |
|
threshold you say no um another option |
|
|
|
571 |
|
00:24:45,760 --> 00:24:51,679 |
|
would be um you have a threshold and you |
|
|
|
572 |
|
00:24:50,279 --> 00:24:56,080 |
|
say |
|
|
|
573 |
|
00:24:51,679 --> 00:24:56,080 |
|
yes no |
|
|
|
574 |
|
00:24:56,200 --> 00:25:00,559 |
|
obain so you know you don't give an |
|
|
|
575 |
|
00:24:58,360 --> 00:25:02,520 |
|
answer and depending on how you're |
|
|
|
576 |
|
00:25:00,559 --> 00:25:03,720 |
|
evaluated what what is a good classifier |
|
|
|
577 |
|
00:25:02,520 --> 00:25:07,799 |
|
you might want to abstain some of the |
|
|
|
578 |
|
00:25:03,720 --> 00:25:10,960 |
|
time also um for multiclass what what's |
|
|
|
579 |
|
00:25:07,799 --> 00:25:10,960 |
|
a standard decision role for |
|
|
|
580 |
|
00:25:11,120 --> 00:25:16,720 |
|
multiclass argmax yeah exactly so um |
|
|
|
581 |
|
00:25:14,279 --> 00:25:19,520 |
|
basically you you find the index that |
|
|
|
582 |
|
00:25:16,720 --> 00:25:22,000 |
|
has the highest score in you output |
|
|
|
583 |
|
00:25:19,520 --> 00:25:24,480 |
|
it we're going to be talking about other |
|
|
|
584 |
|
00:25:22,000 --> 00:25:26,559 |
|
decision rules also um like |
|
|
|
585 |
|
00:25:24,480 --> 00:25:29,480 |
|
self-consistency and minimum based risk |
|
|
|
586 |
|
00:25:26,559 --> 00:25:30,760 |
|
later uh for text generation so you can |
|
|
|
587 |
|
00:25:29,480 --> 00:25:33,000 |
|
just keep that in mind and then we'll |
|
|
|
588 |
|
00:25:30,760 --> 00:25:36,279 |
|
forget about it for like several |
|
|
|
589 |
|
00:25:33,000 --> 00:25:39,559 |
|
classes um so for sentiment |
|
|
|
590 |
|
00:25:36,279 --> 00:25:42,159 |
|
class um I have a Cod |
|
|
|
591 |
|
00:25:39,559 --> 00:25:45,159 |
|
walk |
|
|
|
592 |
|
00:25:42,159 --> 00:25:45,159 |
|
here |
|
|
|
593 |
|
00:25:46,240 --> 00:25:54,320 |
|
and this is pretty simple um but if |
|
|
|
594 |
|
00:25:50,320 --> 00:25:58,559 |
|
you're bored uh of the class and would |
|
|
|
595 |
|
00:25:54,320 --> 00:26:01,000 |
|
like to um try out yourself you can |
|
|
|
596 |
|
00:25:58,559 --> 00:26:04,480 |
|
Challenge and try to get a better score |
|
|
|
597 |
|
00:26:01,000 --> 00:26:06,120 |
|
than I do um over the next few minutes |
|
|
|
598 |
|
00:26:04,480 --> 00:26:06,880 |
|
but we have this rule based classifier |
|
|
|
599 |
|
00:26:06,120 --> 00:26:10,240 |
|
in |
|
|
|
600 |
|
00:26:06,880 --> 00:26:12,640 |
|
here and I will open it up in my vs |
|
|
|
601 |
|
00:26:10,240 --> 00:26:15,360 |
|
code |
|
|
|
602 |
|
00:26:12,640 --> 00:26:18,360 |
|
to try to create a rule-based classifier |
|
|
|
603 |
|
00:26:15,360 --> 00:26:18,360 |
|
and basically the way this |
|
|
|
604 |
|
00:26:22,799 --> 00:26:29,960 |
|
works is |
|
|
|
605 |
|
00:26:25,159 --> 00:26:29,960 |
|
that we have a feature |
|
|
|
606 |
|
00:26:31,720 --> 00:26:37,720 |
|
extraction we have feature extraction we |
|
|
|
607 |
|
00:26:34,120 --> 00:26:40,679 |
|
have scoring and we have um a decision |
|
|
|
608 |
|
00:26:37,720 --> 00:26:43,480 |
|
rle so here for our feature extraction I |
|
|
|
609 |
|
00:26:40,679 --> 00:26:44,720 |
|
have created a list of good words and a |
|
|
|
610 |
|
00:26:43,480 --> 00:26:46,720 |
|
list of bad |
|
|
|
611 |
|
00:26:44,720 --> 00:26:48,960 |
|
words |
|
|
|
612 |
|
00:26:46,720 --> 00:26:51,320 |
|
and what we do is we just count the |
|
|
|
613 |
|
00:26:48,960 --> 00:26:53,000 |
|
number of good words that appeared and |
|
|
|
614 |
|
00:26:51,320 --> 00:26:55,320 |
|
count the number of bad words that |
|
|
|
615 |
|
00:26:53,000 --> 00:26:57,880 |
|
appeared then we also have a bias |
|
|
|
616 |
|
00:26:55,320 --> 00:27:01,159 |
|
feature so the bias feature is a feature |
|
|
|
617 |
|
00:26:57,880 --> 00:27:03,679 |
|
that's always one and so what that |
|
|
|
618 |
|
00:27:01,159 --> 00:27:06,799 |
|
results in is we have a dimension three |
|
|
|
619 |
|
00:27:03,679 --> 00:27:08,880 |
|
feature Vector um where this is like the |
|
|
|
620 |
|
00:27:06,799 --> 00:27:11,320 |
|
number of good words this is the number |
|
|
|
621 |
|
00:27:08,880 --> 00:27:15,320 |
|
of bad words and then you have the |
|
|
|
622 |
|
00:27:11,320 --> 00:27:17,760 |
|
bias and then I also Define the feature |
|
|
|
623 |
|
00:27:15,320 --> 00:27:20,039 |
|
weights that so for every good word we |
|
|
|
624 |
|
00:27:17,760 --> 00:27:22,200 |
|
add one to our score for every bad word |
|
|
|
625 |
|
00:27:20,039 --> 00:27:25,559 |
|
we add uh we subtract one from our score |
|
|
|
626 |
|
00:27:22,200 --> 00:27:29,399 |
|
and for the BIOS we absor and so we then |
|
|
|
627 |
|
00:27:25,559 --> 00:27:30,480 |
|
take the dot product between |
|
|
|
628 |
|
00:27:29,399 --> 00:27:34,360 |
|
these |
|
|
|
629 |
|
00:27:30,480 --> 00:27:36,919 |
|
two and we get minus |
|
|
|
630 |
|
00:27:34,360 --> 00:27:37,640 |
|
0.5 and that gives us uh that gives us |
|
|
|
631 |
|
00:27:36,919 --> 00:27:41,000 |
|
the |
|
|
|
632 |
|
00:27:37,640 --> 00:27:46,000 |
|
squore so let's run |
|
|
|
633 |
|
00:27:41,000 --> 00:27:50,320 |
|
that um and I read in some |
|
|
|
634 |
|
00:27:46,000 --> 00:27:52,600 |
|
data and what this data looks like is |
|
|
|
635 |
|
00:27:50,320 --> 00:27:55,000 |
|
basically we have a |
|
|
|
636 |
|
00:27:52,600 --> 00:27:57,559 |
|
review um which says the rock is |
|
|
|
637 |
|
00:27:55,000 --> 00:27:59,480 |
|
destined to be the 21st Century's new |
|
|
|
638 |
|
00:27:57,559 --> 00:28:01,240 |
|
Conan and that he's going to make a |
|
|
|
639 |
|
00:27:59,480 --> 00:28:03,600 |
|
splash even greater than Arnold |
|
|
|
640 |
|
00:28:01,240 --> 00:28:07,000 |
|
Schwarzenegger jeanclaude vanam or |
|
|
|
641 |
|
00:28:03,600 --> 00:28:09,519 |
|
Steven Seagal um so this seems pretty |
|
|
|
642 |
|
00:28:07,000 --> 00:28:10,840 |
|
positive right I like that's a pretty |
|
|
|
643 |
|
00:28:09,519 --> 00:28:13,200 |
|
high order to be better than Arnold |
|
|
|
644 |
|
00:28:10,840 --> 00:28:16,080 |
|
Schwarzenegger or John Claude vanam uh |
|
|
|
645 |
|
00:28:13,200 --> 00:28:19,519 |
|
if you're familiar with action movies um |
|
|
|
646 |
|
00:28:16,080 --> 00:28:22,840 |
|
and so of course this gets a positive |
|
|
|
647 |
|
00:28:19,519 --> 00:28:24,120 |
|
label and so uh we have run classifier |
|
|
|
648 |
|
00:28:22,840 --> 00:28:25,240 |
|
actually maybe I should call this |
|
|
|
649 |
|
00:28:24,120 --> 00:28:27,600 |
|
decision rule because this is |
|
|
|
650 |
|
00:28:25,240 --> 00:28:29,120 |
|
essentially our decision Rule and here |
|
|
|
651 |
|
00:28:27,600 --> 00:28:32,600 |
|
basically do the thing that I mentioned |
|
|
|
652 |
|
00:28:29,120 --> 00:28:35,440 |
|
here the yes no obstain or in this case |
|
|
|
653 |
|
00:28:32,600 --> 00:28:38,360 |
|
positive negative neutral so if the |
|
|
|
654 |
|
00:28:35,440 --> 00:28:40,159 |
|
score is greater than zero we uh return |
|
|
|
655 |
|
00:28:38,360 --> 00:28:42,480 |
|
one if the score is less than zero we |
|
|
|
656 |
|
00:28:40,159 --> 00:28:44,679 |
|
return negative one which is negative |
|
|
|
657 |
|
00:28:42,480 --> 00:28:47,240 |
|
and otherwise we returns |
|
|
|
658 |
|
00:28:44,679 --> 00:28:48,760 |
|
zero um we have an accuracy calculation |
|
|
|
659 |
|
00:28:47,240 --> 00:28:51,519 |
|
function just calculating the outputs |
|
|
|
660 |
|
00:28:48,760 --> 00:28:55,840 |
|
are good and |
|
|
|
661 |
|
00:28:51,519 --> 00:28:57,440 |
|
um this is uh the overall label count in |
|
|
|
662 |
|
00:28:55,840 --> 00:28:59,919 |
|
the in the output so we can see there |
|
|
|
663 |
|
00:28:57,440 --> 00:29:03,120 |
|
slightly more positives than there are |
|
|
|
664 |
|
00:28:59,919 --> 00:29:06,080 |
|
negatives and then we can run this and |
|
|
|
665 |
|
00:29:03,120 --> 00:29:10,200 |
|
we get a a score of |
|
|
|
666 |
|
00:29:06,080 --> 00:29:14,760 |
|
43 and so one one thing that I have |
|
|
|
667 |
|
00:29:10,200 --> 00:29:19,279 |
|
found um is I I do a lot of kind |
|
|
|
668 |
|
00:29:14,760 --> 00:29:21,240 |
|
of research on how to make NLP systems |
|
|
|
669 |
|
00:29:19,279 --> 00:29:23,600 |
|
better and one of the things I found |
|
|
|
670 |
|
00:29:21,240 --> 00:29:26,679 |
|
really invaluable |
|
|
|
671 |
|
00:29:23,600 --> 00:29:27,840 |
|
is if you're in a situation where you |
|
|
|
672 |
|
00:29:26,679 --> 00:29:29,720 |
|
have a |
|
|
|
673 |
|
00:29:27,840 --> 00:29:31,760 |
|
set task and you just want to make the |
|
|
|
674 |
|
00:29:29,720 --> 00:29:33,760 |
|
system better on the set task doing |
|
|
|
675 |
|
00:29:31,760 --> 00:29:35,159 |
|
comprehensive error analysis and |
|
|
|
676 |
|
00:29:33,760 --> 00:29:37,320 |
|
understanding where your system is |
|
|
|
677 |
|
00:29:35,159 --> 00:29:39,880 |
|
failing is one of the best ways to do |
|
|
|
678 |
|
00:29:37,320 --> 00:29:42,200 |
|
that and I would like to do a very |
|
|
|
679 |
|
00:29:39,880 --> 00:29:43,640 |
|
rudimentary version of this here and |
|
|
|
680 |
|
00:29:42,200 --> 00:29:46,519 |
|
what I'm doing essentially is I'm just |
|
|
|
681 |
|
00:29:43,640 --> 00:29:47,480 |
|
randomly picking uh several examples |
|
|
|
682 |
|
00:29:46,519 --> 00:29:49,320 |
|
that were |
|
|
|
683 |
|
00:29:47,480 --> 00:29:52,000 |
|
correct |
|
|
|
684 |
|
00:29:49,320 --> 00:29:54,840 |
|
um and so like let let's look at the |
|
|
|
685 |
|
00:29:52,000 --> 00:29:58,200 |
|
examples here um here the true label is |
|
|
|
686 |
|
00:29:54,840 --> 00:30:00,760 |
|
zero um in this predicted one um it may |
|
|
|
687 |
|
00:29:58,200 --> 00:30:03,440 |
|
not be as cutting as Woody or as true as |
|
|
|
688 |
|
00:30:00,760 --> 00:30:05,039 |
|
back in the Glory Days of uh weekend and |
|
|
|
689 |
|
00:30:03,440 --> 00:30:07,440 |
|
two or three things that I know about |
|
|
|
690 |
|
00:30:05,039 --> 00:30:09,640 |
|
her but who else engaged in film Mak |
|
|
|
691 |
|
00:30:07,440 --> 00:30:12,679 |
|
today is so cognizant of the cultural |
|
|
|
692 |
|
00:30:09,640 --> 00:30:14,480 |
|
and moral issues involved in the process |
|
|
|
693 |
|
00:30:12,679 --> 00:30:17,600 |
|
so what words in here are a good |
|
|
|
694 |
|
00:30:14,480 --> 00:30:20,840 |
|
indication that this is a neutral |
|
|
|
695 |
|
00:30:17,600 --> 00:30:20,840 |
|
sentence any |
|
|
|
696 |
|
00:30:23,760 --> 00:30:28,399 |
|
ideas little bit tough |
|
|
|
697 |
|
00:30:26,240 --> 00:30:30,919 |
|
huh starting to think maybe we should be |
|
|
|
698 |
|
00:30:28,399 --> 00:30:30,919 |
|
using machine |
|
|
|
699 |
|
00:30:31,480 --> 00:30:37,440 |
|
learning |
|
|
|
700 |
|
00:30:34,080 --> 00:30:40,320 |
|
um even by the intentionally low |
|
|
|
701 |
|
00:30:37,440 --> 00:30:41,559 |
|
standards of fratboy humor sority boys |
|
|
|
702 |
|
00:30:40,320 --> 00:30:43,840 |
|
is a |
|
|
|
703 |
|
00:30:41,559 --> 00:30:46,080 |
|
Bowser I think frat boy is maybe |
|
|
|
704 |
|
00:30:43,840 --> 00:30:47,360 |
|
negative sentiment if you're familiar |
|
|
|
705 |
|
00:30:46,080 --> 00:30:50,360 |
|
with |
|
|
|
706 |
|
00:30:47,360 --> 00:30:51,960 |
|
us us I don't have any negative |
|
|
|
707 |
|
00:30:50,360 --> 00:30:54,519 |
|
sentiment but the people who say it that |
|
|
|
708 |
|
00:30:51,960 --> 00:30:55,960 |
|
way have negative senent maybe so if we |
|
|
|
709 |
|
00:30:54,519 --> 00:31:01,080 |
|
wanted to go in and do that we could |
|
|
|
710 |
|
00:30:55,960 --> 00:31:01,080 |
|
maybe I won't save this but |
|
|
|
711 |
|
00:31:01,519 --> 00:31:08,919 |
|
uh |
|
|
|
712 |
|
00:31:04,240 --> 00:31:11,840 |
|
um oh whoops I'll go back and fix it uh |
|
|
|
713 |
|
00:31:08,919 --> 00:31:14,840 |
|
crass crass is pretty obviously negative |
|
|
|
714 |
|
00:31:11,840 --> 00:31:14,840 |
|
right so I can add |
|
|
|
715 |
|
00:31:17,039 --> 00:31:21,080 |
|
crass actually let me just add |
|
|
|
716 |
|
00:31:21,760 --> 00:31:29,159 |
|
CR and then um I'll go back and have our |
|
|
|
717 |
|
00:31:26,559 --> 00:31:29,159 |
|
train accurate |
|
|
|
718 |
|
00:31:32,159 --> 00:31:36,240 |
|
wa maybe maybe I need to run the whole |
|
|
|
719 |
|
00:31:33,960 --> 00:31:36,240 |
|
thing |
|
|
|
720 |
|
00:31:36,960 --> 00:31:39,960 |
|
again |
|
|
|
721 |
|
00:31:40,960 --> 00:31:45,880 |
|
and that budg the training accuracy a |
|
|
|
722 |
|
00:31:43,679 --> 00:31:50,360 |
|
little um the dev test accuracy not very |
|
|
|
723 |
|
00:31:45,880 --> 00:31:53,919 |
|
much so I could go through and do this |
|
|
|
724 |
|
00:31:50,360 --> 00:31:53,919 |
|
um let me add |
|
|
|
725 |
|
00:31:54,000 --> 00:31:58,320 |
|
unengaging so I could go through and do |
|
|
|
726 |
|
00:31:56,000 --> 00:32:01,720 |
|
this all day and you probably be very |
|
|
|
727 |
|
00:31:58,320 --> 00:32:01,720 |
|
bored on |
|
|
|
728 |
|
00:32:04,240 --> 00:32:08,360 |
|
engage but I won't do that uh because we |
|
|
|
729 |
|
00:32:06,919 --> 00:32:10,679 |
|
have much more important things to be |
|
|
|
730 |
|
00:32:08,360 --> 00:32:14,679 |
|
doing |
|
|
|
731 |
|
00:32:10,679 --> 00:32:16,440 |
|
um and uh so anyway we um we could go |
|
|
|
732 |
|
00:32:14,679 --> 00:32:18,919 |
|
through and design all the features here |
|
|
|
733 |
|
00:32:16,440 --> 00:32:21,279 |
|
but like why is this complicated like |
|
|
|
734 |
|
00:32:18,919 --> 00:32:22,600 |
|
the the reason why it was complicated |
|
|
|
735 |
|
00:32:21,279 --> 00:32:25,840 |
|
became pretty |
|
|
|
736 |
|
00:32:22,600 --> 00:32:27,840 |
|
clear from the uh from the very |
|
|
|
737 |
|
00:32:25,840 --> 00:32:29,639 |
|
beginning uh the very first example I |
|
|
|
738 |
|
00:32:27,840 --> 00:32:32,200 |
|
showed you which was that was a really |
|
|
|
739 |
|
00:32:29,639 --> 00:32:34,720 |
|
complicated sentence like all of us |
|
|
|
740 |
|
00:32:32,200 --> 00:32:36,240 |
|
could see that it wasn't like really |
|
|
|
741 |
|
00:32:34,720 --> 00:32:38,679 |
|
strongly positive it wasn't really |
|
|
|
742 |
|
00:32:36,240 --> 00:32:40,519 |
|
strongly negative it was kind of like in |
|
|
|
743 |
|
00:32:38,679 --> 00:32:42,919 |
|
the middle but it was in the middle and |
|
|
|
744 |
|
00:32:40,519 --> 00:32:44,600 |
|
it said it in a very long way uh you |
|
|
|
745 |
|
00:32:42,919 --> 00:32:46,120 |
|
know not using any clearly positive |
|
|
|
746 |
|
00:32:44,600 --> 00:32:47,639 |
|
sentiment words not using any clearly |
|
|
|
747 |
|
00:32:46,120 --> 00:32:49,760 |
|
negative sentiment |
|
|
|
748 |
|
00:32:47,639 --> 00:32:53,760 |
|
words |
|
|
|
749 |
|
00:32:49,760 --> 00:32:56,519 |
|
um so yeah basically I I |
|
|
|
750 |
|
00:32:53,760 --> 00:33:00,559 |
|
improved um but what are the difficult |
|
|
|
751 |
|
00:32:56,519 --> 00:33:03,720 |
|
cases uh that we saw here so the first |
|
|
|
752 |
|
00:33:00,559 --> 00:33:07,639 |
|
one is low frequency |
|
|
|
753 |
|
00:33:03,720 --> 00:33:09,760 |
|
words so um here's an example the action |
|
|
|
754 |
|
00:33:07,639 --> 00:33:11,519 |
|
switches between past and present but |
|
|
|
755 |
|
00:33:09,760 --> 00:33:13,120 |
|
the material link is too tenuous to |
|
|
|
756 |
|
00:33:11,519 --> 00:33:16,840 |
|
Anchor the emotional connections at |
|
|
|
757 |
|
00:33:13,120 --> 00:33:19,519 |
|
purport to span a 125 year divide so |
|
|
|
758 |
|
00:33:16,840 --> 00:33:21,080 |
|
this is negative um tenuous is kind of a |
|
|
|
759 |
|
00:33:19,519 --> 00:33:22,799 |
|
negative word purport is kind of a |
|
|
|
760 |
|
00:33:21,080 --> 00:33:24,760 |
|
negative word but it doesn't appear very |
|
|
|
761 |
|
00:33:22,799 --> 00:33:26,159 |
|
frequently so I would need to spend all |
|
|
|
762 |
|
00:33:24,760 --> 00:33:29,720 |
|
my time looking for these words and |
|
|
|
763 |
|
00:33:26,159 --> 00:33:32,480 |
|
trying to them in um here's yet another |
|
|
|
764 |
|
00:33:29,720 --> 00:33:34,240 |
|
horse franchise mucking up its storyline |
|
|
|
765 |
|
00:33:32,480 --> 00:33:36,639 |
|
with glitches casual fans could correct |
|
|
|
766 |
|
00:33:34,240 --> 00:33:40,159 |
|
in their sleep negative |
|
|
|
767 |
|
00:33:36,639 --> 00:33:42,600 |
|
again um so the solutions here are keep |
|
|
|
768 |
|
00:33:40,159 --> 00:33:46,880 |
|
working until we get all of them which |
|
|
|
769 |
|
00:33:42,600 --> 00:33:49,159 |
|
is maybe not super fun um or incorporate |
|
|
|
770 |
|
00:33:46,880 --> 00:33:51,639 |
|
external resources such as sentiment |
|
|
|
771 |
|
00:33:49,159 --> 00:33:52,880 |
|
dictionaries that people created uh we |
|
|
|
772 |
|
00:33:51,639 --> 00:33:55,960 |
|
could do that but that's a lot of |
|
|
|
773 |
|
00:33:52,880 --> 00:33:57,480 |
|
engineering effort to make something |
|
|
|
774 |
|
00:33:55,960 --> 00:34:00,639 |
|
work |
|
|
|
775 |
|
00:33:57,480 --> 00:34:03,720 |
|
um another one is conjugation so we saw |
|
|
|
776 |
|
00:34:00,639 --> 00:34:06,600 |
|
unengaging I guess that's an example of |
|
|
|
777 |
|
00:34:03,720 --> 00:34:08,359 |
|
conjugation uh some other ones are |
|
|
|
778 |
|
00:34:06,600 --> 00:34:10,520 |
|
operatic sprawling picture that's |
|
|
|
779 |
|
00:34:08,359 --> 00:34:12,040 |
|
entertainingly acted magnificently shot |
|
|
|
780 |
|
00:34:10,520 --> 00:34:15,480 |
|
and gripping enough to sustain most of |
|
|
|
781 |
|
00:34:12,040 --> 00:34:17,399 |
|
its 170 minute length so here we have |
|
|
|
782 |
|
00:34:15,480 --> 00:34:19,079 |
|
magnificently so even if I added |
|
|
|
783 |
|
00:34:17,399 --> 00:34:20,480 |
|
magnificent this wouldn't have been |
|
|
|
784 |
|
00:34:19,079 --> 00:34:23,800 |
|
clocked |
|
|
|
785 |
|
00:34:20,480 --> 00:34:26,599 |
|
right um it's basically an overlong |
|
|
|
786 |
|
00:34:23,800 --> 00:34:28,839 |
|
episode of tales from the cryp so that's |
|
|
|
787 |
|
00:34:26,599 --> 00:34:31,480 |
|
maybe another |
|
|
|
788 |
|
00:34:28,839 --> 00:34:33,040 |
|
example um so some things that we could |
|
|
|
789 |
|
00:34:31,480 --> 00:34:35,320 |
|
do or what we would have done before the |
|
|
|
790 |
|
00:34:33,040 --> 00:34:37,720 |
|
modern Paradigm of machine learning is |
|
|
|
791 |
|
00:34:35,320 --> 00:34:40,079 |
|
we would run some sort of normalizer |
|
|
|
792 |
|
00:34:37,720 --> 00:34:42,800 |
|
like a stemmer or other things like this |
|
|
|
793 |
|
00:34:40,079 --> 00:34:45,240 |
|
in order to convert this into uh the |
|
|
|
794 |
|
00:34:42,800 --> 00:34:48,599 |
|
root wordss that we already have seen |
|
|
|
795 |
|
00:34:45,240 --> 00:34:52,040 |
|
somewhere in our data or have already |
|
|
|
796 |
|
00:34:48,599 --> 00:34:54,040 |
|
handed so that requires um conjugation |
|
|
|
797 |
|
00:34:52,040 --> 00:34:55,879 |
|
analysis or morphological analysis as we |
|
|
|
798 |
|
00:34:54,040 --> 00:34:57,400 |
|
say it in |
|
|
|
799 |
|
00:34:55,879 --> 00:35:00,680 |
|
technicals |
|
|
|
800 |
|
00:34:57,400 --> 00:35:03,960 |
|
negation this is a tricky one so this |
|
|
|
801 |
|
00:35:00,680 --> 00:35:06,760 |
|
one's not nearly as Dreadful as expected |
|
|
|
802 |
|
00:35:03,960 --> 00:35:08,800 |
|
so Dreadful is a pretty bad word right |
|
|
|
803 |
|
00:35:06,760 --> 00:35:13,000 |
|
but not nearly as Dreadful as expected |
|
|
|
804 |
|
00:35:08,800 --> 00:35:14,440 |
|
is like a solidly neutral um you know or |
|
|
|
805 |
|
00:35:13,000 --> 00:35:16,359 |
|
maybe even |
|
|
|
806 |
|
00:35:14,440 --> 00:35:18,920 |
|
positive I would I would say that's |
|
|
|
807 |
|
00:35:16,359 --> 00:35:20,640 |
|
neutral but you know uh neutral or |
|
|
|
808 |
|
00:35:18,920 --> 00:35:23,800 |
|
positive it's definitely not |
|
|
|
809 |
|
00:35:20,640 --> 00:35:26,359 |
|
negative um serving s doesn't serve up a |
|
|
|
810 |
|
00:35:23,800 --> 00:35:29,480 |
|
whole lot of laughs so laughs is |
|
|
|
811 |
|
00:35:26,359 --> 00:35:31,880 |
|
obviously positive but not serving UPS |
|
|
|
812 |
|
00:35:29,480 --> 00:35:34,440 |
|
is obviously |
|
|
|
813 |
|
00:35:31,880 --> 00:35:36,839 |
|
negative so if negation modifies the |
|
|
|
814 |
|
00:35:34,440 --> 00:35:38,240 |
|
word disregard it now we would probably |
|
|
|
815 |
|
00:35:36,839 --> 00:35:41,440 |
|
need to do some sort of syntactic |
|
|
|
816 |
|
00:35:38,240 --> 00:35:45,599 |
|
analysis or semantic analysis of |
|
|
|
817 |
|
00:35:41,440 --> 00:35:47,520 |
|
some metaphor an analogy so puts a human |
|
|
|
818 |
|
00:35:45,599 --> 00:35:50,640 |
|
face on a land most westerners are |
|
|
|
819 |
|
00:35:47,520 --> 00:35:52,880 |
|
unfamiliar though uh this is |
|
|
|
820 |
|
00:35:50,640 --> 00:35:54,960 |
|
positive green might want to hang on to |
|
|
|
821 |
|
00:35:52,880 --> 00:35:58,800 |
|
that ski mask as robbery may be the only |
|
|
|
822 |
|
00:35:54,960 --> 00:35:58,800 |
|
way to pay for this next project |
|
|
|
823 |
|
00:35:58,839 --> 00:36:03,640 |
|
so this this is saying that the movie |
|
|
|
824 |
|
00:36:01,960 --> 00:36:05,560 |
|
was so bad that the director will have |
|
|
|
825 |
|
00:36:03,640 --> 00:36:08,359 |
|
to rob people in order to get money for |
|
|
|
826 |
|
00:36:05,560 --> 00:36:11,000 |
|
the next project so that's kind of bad I |
|
|
|
827 |
|
00:36:08,359 --> 00:36:12,880 |
|
guess um has all the depth of a waiting |
|
|
|
828 |
|
00:36:11,000 --> 00:36:14,520 |
|
pool this is kind of my favorite one |
|
|
|
829 |
|
00:36:12,880 --> 00:36:15,880 |
|
because it's really short and sweet but |
|
|
|
830 |
|
00:36:14,520 --> 00:36:18,800 |
|
you know you need to know how deep a |
|
|
|
831 |
|
00:36:15,880 --> 00:36:21,440 |
|
waiting pool is um so that's |
|
|
|
832 |
|
00:36:18,800 --> 00:36:22,960 |
|
negative so the solution here I don't |
|
|
|
833 |
|
00:36:21,440 --> 00:36:24,680 |
|
really even know how to handle this with |
|
|
|
834 |
|
00:36:22,960 --> 00:36:26,880 |
|
a rule based system I have no idea how |
|
|
|
835 |
|
00:36:24,680 --> 00:36:30,040 |
|
we would possibly do this yeah machine |
|
|
|
836 |
|
00:36:26,880 --> 00:36:32,400 |
|
learning based models seem to be pretty |
|
|
|
837 |
|
00:36:30,040 --> 00:36:37,000 |
|
adaptive okay and then I start doing |
|
|
|
838 |
|
00:36:32,400 --> 00:36:37,000 |
|
these ones um anyone have a good |
|
|
|
839 |
|
00:36:38,160 --> 00:36:46,800 |
|
idea any any other friends who know |
|
|
|
840 |
|
00:36:42,520 --> 00:36:50,040 |
|
Japanese no okay um so yeah that's |
|
|
|
841 |
|
00:36:46,800 --> 00:36:52,839 |
|
positive um that one's negative uh and |
|
|
|
842 |
|
00:36:50,040 --> 00:36:54,920 |
|
the solution here is learn Japanese I |
|
|
|
843 |
|
00:36:52,839 --> 00:36:56,800 |
|
guess or whatever other language you |
|
|
|
844 |
|
00:36:54,920 --> 00:37:00,040 |
|
want to process so like obviously |
|
|
|
845 |
|
00:36:56,800 --> 00:37:03,720 |
|
rule-based systems don't scale very |
|
|
|
846 |
|
00:37:00,040 --> 00:37:05,119 |
|
well so um we've moved but like rule |
|
|
|
847 |
|
00:37:03,720 --> 00:37:06,319 |
|
based systems don't scale very well |
|
|
|
848 |
|
00:37:05,119 --> 00:37:08,160 |
|
we're not going to be using them for |
|
|
|
849 |
|
00:37:06,319 --> 00:37:11,400 |
|
most of the things we do in this class |
|
|
|
850 |
|
00:37:08,160 --> 00:37:14,240 |
|
but I do think it's sometimes useful to |
|
|
|
851 |
|
00:37:11,400 --> 00:37:15,640 |
|
try to create one for your task maybe |
|
|
|
852 |
|
00:37:14,240 --> 00:37:16,680 |
|
right at the very beginning of a project |
|
|
|
853 |
|
00:37:15,640 --> 00:37:18,560 |
|
because it gives you an idea about |
|
|
|
854 |
|
00:37:16,680 --> 00:37:21,160 |
|
what's really hard about the task in |
|
|
|
855 |
|
00:37:18,560 --> 00:37:22,480 |
|
some cases so um yeah I wouldn't |
|
|
|
856 |
|
00:37:21,160 --> 00:37:25,599 |
|
entirely discount them I'm not |
|
|
|
857 |
|
00:37:22,480 --> 00:37:27,400 |
|
introducing them for no reason |
|
|
|
858 |
|
00:37:25,599 --> 00:37:29,880 |
|
whatsoever |
|
|
|
859 |
|
00:37:27,400 --> 00:37:34,160 |
|
so next is machine learning based anal |
|
|
|
860 |
|
00:37:29,880 --> 00:37:35,400 |
|
and machine learning uh in general uh I |
|
|
|
861 |
|
00:37:34,160 --> 00:37:36,640 |
|
here actually when I say machine |
|
|
|
862 |
|
00:37:35,400 --> 00:37:38,160 |
|
learning I'm going to be talking about |
|
|
|
863 |
|
00:37:36,640 --> 00:37:39,560 |
|
the traditional fine-tuning approach |
|
|
|
864 |
|
00:37:38,160 --> 00:37:43,520 |
|
where we have a training set Dev set |
|
|
|
865 |
|
00:37:39,560 --> 00:37:46,359 |
|
test set and so we take our training set |
|
|
|
866 |
|
00:37:43,520 --> 00:37:49,680 |
|
we run some learning algorithm over it |
|
|
|
867 |
|
00:37:46,359 --> 00:37:52,319 |
|
we have a learned feature extractor F A |
|
|
|
868 |
|
00:37:49,680 --> 00:37:55,839 |
|
possibly learned feature extractor F |
|
|
|
869 |
|
00:37:52,319 --> 00:37:57,880 |
|
possibly learned scoring function W and |
|
|
|
870 |
|
00:37:55,839 --> 00:38:00,800 |
|
uh then we apply our inference algorithm |
|
|
|
871 |
|
00:37:57,880 --> 00:38:02,839 |
|
our decision Rule and make decisions |
|
|
|
872 |
|
00:38:00,800 --> 00:38:04,200 |
|
when I say possibly learned actually the |
|
|
|
873 |
|
00:38:02,839 --> 00:38:06,119 |
|
first example I'm going to give of a |
|
|
|
874 |
|
00:38:04,200 --> 00:38:07,760 |
|
machine learning based technique is uh |
|
|
|
875 |
|
00:38:06,119 --> 00:38:10,079 |
|
doesn't have a learned feature extractor |
|
|
|
876 |
|
00:38:07,760 --> 00:38:12,800 |
|
but most things that we use nowadays do |
|
|
|
877 |
|
00:38:10,079 --> 00:38:12,800 |
|
have learned feature |
|
|
|
878 |
|
00:38:13,200 --> 00:38:18,040 |
|
extractors so our first attempt is going |
|
|
|
879 |
|
00:38:15,640 --> 00:38:21,760 |
|
to be a bag of words model uh and the |
|
|
|
880 |
|
00:38:18,040 --> 00:38:27,119 |
|
way a bag of wordss model works is uh |
|
|
|
881 |
|
00:38:21,760 --> 00:38:30,160 |
|
essentially we start out by looking up a |
|
|
|
882 |
|
00:38:27,119 --> 00:38:33,240 |
|
Vector where one element in the vector |
|
|
|
883 |
|
00:38:30,160 --> 00:38:36,240 |
|
is uh is one and all the other elements |
|
|
|
884 |
|
00:38:33,240 --> 00:38:38,040 |
|
in the vector are zero and so if the |
|
|
|
885 |
|
00:38:36,240 --> 00:38:40,319 |
|
word is different the position in the |
|
|
|
886 |
|
00:38:38,040 --> 00:38:42,839 |
|
vector that's one will be different we |
|
|
|
887 |
|
00:38:40,319 --> 00:38:46,280 |
|
add all of these together and this gives |
|
|
|
888 |
|
00:38:42,839 --> 00:38:48,200 |
|
us a vector where each element is the |
|
|
|
889 |
|
00:38:46,280 --> 00:38:50,359 |
|
frequency of that word in the vector and |
|
|
|
890 |
|
00:38:48,200 --> 00:38:52,520 |
|
then we multiply that by weights and we |
|
|
|
891 |
|
00:38:50,359 --> 00:38:55,520 |
|
get a |
|
|
|
892 |
|
00:38:52,520 --> 00:38:57,160 |
|
score and um here as I said this is not |
|
|
|
893 |
|
00:38:55,520 --> 00:39:00,359 |
|
a learned feature |
|
|
|
894 |
|
00:38:57,160 --> 00:39:02,079 |
|
uh Vector this is basically uh sorry not |
|
|
|
895 |
|
00:39:00,359 --> 00:39:04,359 |
|
a learn feature extractor this is |
|
|
|
896 |
|
00:39:02,079 --> 00:39:06,200 |
|
basically a fixed feature extractor but |
|
|
|
897 |
|
00:39:04,359 --> 00:39:09,839 |
|
the weights themselves are |
|
|
|
898 |
|
00:39:06,200 --> 00:39:11,640 |
|
learned um so my my question is I |
|
|
|
899 |
|
00:39:09,839 --> 00:39:14,599 |
|
mentioned a whole lot of problems before |
|
|
|
900 |
|
00:39:11,640 --> 00:39:17,480 |
|
I mentioned infrequent words I mentioned |
|
|
|
901 |
|
00:39:14,599 --> 00:39:20,760 |
|
conjugation I mentioned uh different |
|
|
|
902 |
|
00:39:17,480 --> 00:39:22,880 |
|
languages I mentioned syntax and |
|
|
|
903 |
|
00:39:20,760 --> 00:39:24,599 |
|
metaphor so which of these do we think |
|
|
|
904 |
|
00:39:22,880 --> 00:39:25,440 |
|
would be fixed by this sort of learning |
|
|
|
905 |
|
00:39:24,599 --> 00:39:27,400 |
|
based |
|
|
|
906 |
|
00:39:25,440 --> 00:39:29,640 |
|
approach |
|
|
|
907 |
|
00:39:27,400 --> 00:39:29,640 |
|
any |
|
|
|
908 |
|
00:39:29,920 --> 00:39:35,200 |
|
ideas maybe not fixed maybe made |
|
|
|
909 |
|
00:39:32,520 --> 00:39:35,200 |
|
significantly |
|
|
|
910 |
|
00:39:36,880 --> 00:39:41,560 |
|
better any Brave uh brave |
|
|
|
911 |
|
00:39:44,880 --> 00:39:48,440 |
|
people maybe maybe |
|
|
|
912 |
|
00:39:53,720 --> 00:39:58,400 |
|
negation okay so maybe doesn't when it |
|
|
|
913 |
|
00:39:55,760 --> 00:39:58,400 |
|
have a negative qu |
|
|
|
914 |
|
00:40:02,960 --> 00:40:07,560 |
|
yeah yeah so for the conjugation if we |
|
|
|
915 |
|
00:40:05,520 --> 00:40:09,200 |
|
had the conjugations of the stems mapped |
|
|
|
916 |
|
00:40:07,560 --> 00:40:11,119 |
|
in the same position that might fix a |
|
|
|
917 |
|
00:40:09,200 --> 00:40:12,920 |
|
conjugation problem but I would say if |
|
|
|
918 |
|
00:40:11,119 --> 00:40:15,200 |
|
you don't do that then this kind of |
|
|
|
919 |
|
00:40:12,920 --> 00:40:18,160 |
|
fixes conjugation a little bit but maybe |
|
|
|
920 |
|
00:40:15,200 --> 00:40:21,319 |
|
not not really yeah kind of fix |
|
|
|
921 |
|
00:40:18,160 --> 00:40:24,079 |
|
conjugation because like they're using |
|
|
|
922 |
|
00:40:21,319 --> 00:40:26,760 |
|
the same there |
|
|
|
923 |
|
00:40:24,079 --> 00:40:28,400 |
|
probably different variations so we |
|
|
|
924 |
|
00:40:26,760 --> 00:40:31,359 |
|
learn how to |
|
|
|
925 |
|
00:40:28,400 --> 00:40:33,400 |
|
classify surrounding |
|
|
|
926 |
|
00:40:31,359 --> 00:40:35,000 |
|
structure yeah if it's a big enough |
|
|
|
927 |
|
00:40:33,400 --> 00:40:36,760 |
|
training set you might have covered the |
|
|
|
928 |
|
00:40:35,000 --> 00:40:37,880 |
|
various conjugations but if you haven't |
|
|
|
929 |
|
00:40:36,760 --> 00:40:43,000 |
|
and you don't have any rule-based |
|
|
|
930 |
|
00:40:37,880 --> 00:40:43,000 |
|
processing it it might still be problems |
|
|
|
931 |
|
00:40:45,400 --> 00:40:50,359 |
|
yeah yeah so in frequent words if you |
|
|
|
932 |
|
00:40:48,280 --> 00:40:52,560 |
|
have a large enough training set yeah |
|
|
|
933 |
|
00:40:50,359 --> 00:40:54,599 |
|
you'll be able to fix it to some extent |
|
|
|
934 |
|
00:40:52,560 --> 00:40:56,480 |
|
so none of the problems are entirely |
|
|
|
935 |
|
00:40:54,599 --> 00:40:57,880 |
|
fixed but a lot of them are made better |
|
|
|
936 |
|
00:40:56,480 --> 00:40:58,960 |
|
different languages is also made better |
|
|
|
937 |
|
00:40:57,880 --> 00:41:00,119 |
|
if you have training data in that |
|
|
|
938 |
|
00:40:58,960 --> 00:41:04,599 |
|
language but if you don't then you're |
|
|
|
939 |
|
00:41:00,119 --> 00:41:06,240 |
|
out of BL so um so now what I'd like to |
|
|
|
940 |
|
00:41:04,599 --> 00:41:10,800 |
|
do is I'd look to like to look at what |
|
|
|
941 |
|
00:41:06,240 --> 00:41:15,079 |
|
our vectors represent so basically um in |
|
|
|
942 |
|
00:41:10,800 --> 00:41:16,880 |
|
uh in binary classification each word um |
|
|
|
943 |
|
00:41:15,079 --> 00:41:19,119 |
|
sorry so the vectors themselves |
|
|
|
944 |
|
00:41:16,880 --> 00:41:21,880 |
|
represent the counts of the words here |
|
|
|
945 |
|
00:41:19,119 --> 00:41:25,319 |
|
I'm talking about what the weight uh |
|
|
|
946 |
|
00:41:21,880 --> 00:41:28,520 |
|
vectors or matrices correspond to and |
|
|
|
947 |
|
00:41:25,319 --> 00:41:31,640 |
|
the weight uh Vector here will be |
|
|
|
948 |
|
00:41:28,520 --> 00:41:33,680 |
|
positive if the word it tends to be |
|
|
|
949 |
|
00:41:31,640 --> 00:41:36,680 |
|
positive if in a binary classification |
|
|
|
950 |
|
00:41:33,680 --> 00:41:38,400 |
|
case in a multiclass classification case |
|
|
|
951 |
|
00:41:36,680 --> 00:41:42,480 |
|
we'll actually have a matrix that looks |
|
|
|
952 |
|
00:41:38,400 --> 00:41:45,480 |
|
like this where um each column or row uh |
|
|
|
953 |
|
00:41:42,480 --> 00:41:47,079 |
|
corresponds to the word and each row or |
|
|
|
954 |
|
00:41:45,480 --> 00:41:49,319 |
|
column corresponds to a label and it |
|
|
|
955 |
|
00:41:47,079 --> 00:41:51,960 |
|
will be higher if that row tends to uh |
|
|
|
956 |
|
00:41:49,319 --> 00:41:54,800 |
|
correlate with that uh that word tends |
|
|
|
957 |
|
00:41:51,960 --> 00:41:56,920 |
|
to correlate that little |
|
|
|
958 |
|
00:41:54,800 --> 00:41:59,240 |
|
bit so |
|
|
|
959 |
|
00:41:56,920 --> 00:42:04,079 |
|
this um training of the bag of words |
|
|
|
960 |
|
00:41:59,240 --> 00:42:07,720 |
|
model is can be done uh so simply that |
|
|
|
961 |
|
00:42:04,079 --> 00:42:10,200 |
|
we uh can put it in a single slide so |
|
|
|
962 |
|
00:42:07,720 --> 00:42:11,599 |
|
basically here uh what we do is we start |
|
|
|
963 |
|
00:42:10,200 --> 00:42:14,760 |
|
out with the feature |
|
|
|
964 |
|
00:42:11,599 --> 00:42:18,880 |
|
weights and for each example in our data |
|
|
|
965 |
|
00:42:14,760 --> 00:42:20,800 |
|
set we extract features um the exact way |
|
|
|
966 |
|
00:42:18,880 --> 00:42:23,920 |
|
I'm extracting features is basically |
|
|
|
967 |
|
00:42:20,800 --> 00:42:25,720 |
|
splitting uh splitting the words using |
|
|
|
968 |
|
00:42:23,920 --> 00:42:28,000 |
|
the python split function and then uh |
|
|
|
969 |
|
00:42:25,720 --> 00:42:31,319 |
|
Counting number of times each word |
|
|
|
970 |
|
00:42:28,000 --> 00:42:33,160 |
|
exists uh we then run the classifier so |
|
|
|
971 |
|
00:42:31,319 --> 00:42:36,280 |
|
actually running the classifier is |
|
|
|
972 |
|
00:42:33,160 --> 00:42:38,200 |
|
exactly the same as what we did for the |
|
|
|
973 |
|
00:42:36,280 --> 00:42:42,640 |
|
uh the rule based system it's just that |
|
|
|
974 |
|
00:42:38,200 --> 00:42:47,359 |
|
we have feature vectors instead and |
|
|
|
975 |
|
00:42:42,640 --> 00:42:51,559 |
|
then if the predicted value is |
|
|
|
976 |
|
00:42:47,359 --> 00:42:55,160 |
|
not value then for each of the |
|
|
|
977 |
|
00:42:51,559 --> 00:42:56,680 |
|
features uh in the feature space we |
|
|
|
978 |
|
00:42:55,160 --> 00:43:02,200 |
|
upweight |
|
|
|
979 |
|
00:42:56,680 --> 00:43:03,599 |
|
the um we upweight The Weight by the |
|
|
|
980 |
|
00:43:02,200 --> 00:43:06,000 |
|
vector |
|
|
|
981 |
|
00:43:03,599 --> 00:43:09,920 |
|
size by or by the amount of the vector |
|
|
|
982 |
|
00:43:06,000 --> 00:43:13,240 |
|
if Y is positive and we downweight the |
|
|
|
983 |
|
00:43:09,920 --> 00:43:16,240 |
|
vector uh by the size of the vector if Y |
|
|
|
984 |
|
00:43:13,240 --> 00:43:18,520 |
|
is negative so this is really really |
|
|
|
985 |
|
00:43:16,240 --> 00:43:20,559 |
|
simple it's uh probably the simplest |
|
|
|
986 |
|
00:43:18,520 --> 00:43:25,079 |
|
possible algorithm for training one of |
|
|
|
987 |
|
00:43:20,559 --> 00:43:27,559 |
|
these models um but I have an |
|
|
|
988 |
|
00:43:25,079 --> 00:43:30,040 |
|
example in this that you can also take a |
|
|
|
989 |
|
00:43:27,559 --> 00:43:31,960 |
|
look at here's a trained bag of words |
|
|
|
990 |
|
00:43:30,040 --> 00:43:33,680 |
|
classifier and we could step through |
|
|
|
991 |
|
00:43:31,960 --> 00:43:34,960 |
|
this is on exactly the same data set as |
|
|
|
992 |
|
00:43:33,680 --> 00:43:37,240 |
|
I did before we're training on the |
|
|
|
993 |
|
00:43:34,960 --> 00:43:42,359 |
|
training set |
|
|
|
994 |
|
00:43:37,240 --> 00:43:43,640 |
|
um and uh evaluating on the dev set um I |
|
|
|
995 |
|
00:43:42,359 --> 00:43:45,880 |
|
also have some extra stuff like I'm |
|
|
|
996 |
|
00:43:43,640 --> 00:43:47,079 |
|
Shuffling the order of the data IDs |
|
|
|
997 |
|
00:43:45,880 --> 00:43:49,440 |
|
which is really important if you're |
|
|
|
998 |
|
00:43:47,079 --> 00:43:53,160 |
|
doing this sort of incremental algorithm |
|
|
|
999 |
|
00:43:49,440 --> 00:43:54,960 |
|
uh because uh what if what if your |
|
|
|
1000 |
|
00:43:53,160 --> 00:43:57,400 |
|
creating data set was ordered in this |
|
|
|
1001 |
|
00:43:54,960 --> 00:44:00,040 |
|
way where you have all of the positive |
|
|
|
1002 |
|
00:43:57,400 --> 00:44:00,040 |
|
labels on |
|
|
|
1003 |
|
00:44:00,359 --> 00:44:04,520 |
|
top and then you have all of the |
|
|
|
1004 |
|
00:44:02,280 --> 00:44:06,680 |
|
negative labels on the |
|
|
|
1005 |
|
00:44:04,520 --> 00:44:08,200 |
|
bottom if you do something like this it |
|
|
|
1006 |
|
00:44:06,680 --> 00:44:10,200 |
|
would see only negative labels at the |
|
|
|
1007 |
|
00:44:08,200 --> 00:44:11,800 |
|
end of training and you might have |
|
|
|
1008 |
|
00:44:10,200 --> 00:44:14,400 |
|
problems because your model would only |
|
|
|
1009 |
|
00:44:11,800 --> 00:44:17,440 |
|
predict negatives so we also Shuffle |
|
|
|
1010 |
|
00:44:14,400 --> 00:44:20,319 |
|
data um and then step through we run the |
|
|
|
1011 |
|
00:44:17,440 --> 00:44:22,559 |
|
classifier and I'm going to run uh five |
|
|
|
1012 |
|
00:44:20,319 --> 00:44:23,640 |
|
epochs of training through the data set |
|
|
|
1013 |
|
00:44:22,559 --> 00:44:27,160 |
|
uh very |
|
|
|
1014 |
|
00:44:23,640 --> 00:44:29,599 |
|
fast and calculate our accuracy |
|
|
|
1015 |
|
00:44:27,160 --> 00:44:33,280 |
|
and this got 75% accuracy on the |
|
|
|
1016 |
|
00:44:29,599 --> 00:44:36,160 |
|
training data set and uh 56% accuracy on |
|
|
|
1017 |
|
00:44:33,280 --> 00:44:40,000 |
|
the Deb data set so uh if you remember |
|
|
|
1018 |
|
00:44:36,160 --> 00:44:41,520 |
|
our rule-based classifier had 42 uh 42 |
|
|
|
1019 |
|
00:44:40,000 --> 00:44:43,880 |
|
accuracy and now our training based |
|
|
|
1020 |
|
00:44:41,520 --> 00:44:45,760 |
|
classifier has 56 accuracy but it's |
|
|
|
1021 |
|
00:44:43,880 --> 00:44:49,359 |
|
overfitting heavily to the training side |
|
|
|
1022 |
|
00:44:45,760 --> 00:44:50,880 |
|
so um basically this is a pretty strong |
|
|
|
1023 |
|
00:44:49,359 --> 00:44:53,480 |
|
advertisement for why we should be using |
|
|
|
1024 |
|
00:44:50,880 --> 00:44:54,960 |
|
machine learning you know I the amount |
|
|
|
1025 |
|
00:44:53,480 --> 00:44:57,800 |
|
of code that we had for this machine |
|
|
|
1026 |
|
00:44:54,960 --> 00:44:59,720 |
|
learning model is basically very similar |
|
|
|
1027 |
|
00:44:57,800 --> 00:45:02,680 |
|
um it's not using any external libraries |
|
|
|
1028 |
|
00:44:59,720 --> 00:45:02,680 |
|
but we're getting better at |
|
|
|
1029 |
|
00:45:03,599 --> 00:45:08,800 |
|
this |
|
|
|
1030 |
|
00:45:05,800 --> 00:45:08,800 |
|
cool |
|
|
|
1031 |
|
00:45:09,559 --> 00:45:16,000 |
|
so cool any any questions |
|
|
|
1032 |
|
00:45:13,520 --> 00:45:18,240 |
|
here and so I'm going to talk about the |
|
|
|
1033 |
|
00:45:16,000 --> 00:45:20,760 |
|
connection to between this algorithm and |
|
|
|
1034 |
|
00:45:18,240 --> 00:45:22,839 |
|
neural networks in the next class um |
|
|
|
1035 |
|
00:45:20,760 --> 00:45:24,200 |
|
because this actually is using a very |
|
|
|
1036 |
|
00:45:22,839 --> 00:45:26,319 |
|
similar training algorithm to what we |
|
|
|
1037 |
|
00:45:24,200 --> 00:45:27,480 |
|
use in neural networks with some uh |
|
|
|
1038 |
|
00:45:26,319 --> 00:45:30,079 |
|
particular |
|
|
|
1039 |
|
00:45:27,480 --> 00:45:32,839 |
|
assumptions cool um so what's missing in |
|
|
|
1040 |
|
00:45:30,079 --> 00:45:34,800 |
|
bag of words um still handling of |
|
|
|
1041 |
|
00:45:32,839 --> 00:45:36,880 |
|
conjugation or compound words is not |
|
|
|
1042 |
|
00:45:34,800 --> 00:45:39,160 |
|
perfect it we can do it to some extent |
|
|
|
1043 |
|
00:45:36,880 --> 00:45:41,079 |
|
to the point where we can uh memorize |
|
|
|
1044 |
|
00:45:39,160 --> 00:45:44,079 |
|
things so I love this movie I love this |
|
|
|
1045 |
|
00:45:41,079 --> 00:45:46,920 |
|
movie another thing is handling word Ser |
|
|
|
1046 |
|
00:45:44,079 --> 00:45:49,240 |
|
uh similarities so I love this movie and |
|
|
|
1047 |
|
00:45:46,920 --> 00:45:50,720 |
|
I adore this movie uh these basically |
|
|
|
1048 |
|
00:45:49,240 --> 00:45:52,119 |
|
mean the same thing as humans we know |
|
|
|
1049 |
|
00:45:50,720 --> 00:45:54,200 |
|
they mean the same thing so we should be |
|
|
|
1050 |
|
00:45:52,119 --> 00:45:56,079 |
|
able to take advantage of that fact to |
|
|
|
1051 |
|
00:45:54,200 --> 00:45:57,839 |
|
learn better models but we're not doing |
|
|
|
1052 |
|
00:45:56,079 --> 00:46:02,760 |
|
that in this model at the moment because |
|
|
|
1053 |
|
00:45:57,839 --> 00:46:05,440 |
|
each unit is uh treated as a atomic unit |
|
|
|
1054 |
|
00:46:02,760 --> 00:46:08,040 |
|
and there's no idea of |
|
|
|
1055 |
|
00:46:05,440 --> 00:46:11,040 |
|
similarity also handling of combination |
|
|
|
1056 |
|
00:46:08,040 --> 00:46:12,760 |
|
features so um I love this movie and I |
|
|
|
1057 |
|
00:46:11,040 --> 00:46:14,920 |
|
don't love this movie I hate this movie |
|
|
|
1058 |
|
00:46:12,760 --> 00:46:17,079 |
|
and I don't hate this movie actually |
|
|
|
1059 |
|
00:46:14,920 --> 00:46:20,400 |
|
this is a little bit tricky because |
|
|
|
1060 |
|
00:46:17,079 --> 00:46:23,240 |
|
negative words are slightly indicative |
|
|
|
1061 |
|
00:46:20,400 --> 00:46:25,280 |
|
of it being negative but actually what |
|
|
|
1062 |
|
00:46:23,240 --> 00:46:28,119 |
|
they do is they negate the other things |
|
|
|
1063 |
|
00:46:25,280 --> 00:46:28,119 |
|
that you're saying in the |
|
|
|
1064 |
|
00:46:28,240 --> 00:46:36,559 |
|
sentence |
|
|
|
1065 |
|
00:46:30,720 --> 00:46:40,480 |
|
so um like love is positive hate is |
|
|
|
1066 |
|
00:46:36,559 --> 00:46:40,480 |
|
negative but like don't |
|
|
|
1067 |
|
00:46:50,359 --> 00:46:56,079 |
|
love it's actually kind of like this |
|
|
|
1068 |
|
00:46:52,839 --> 00:46:59,359 |
|
right like um Love is very positive POS |
|
|
|
1069 |
|
00:46:56,079 --> 00:47:01,760 |
|
hate is very negative but don't love is |
|
|
|
1070 |
|
00:46:59,359 --> 00:47:04,680 |
|
like slightly less positive than don't |
|
|
|
1071 |
|
00:47:01,760 --> 00:47:06,160 |
|
hate right so um It's actually kind of |
|
|
|
1072 |
|
00:47:04,680 --> 00:47:07,559 |
|
tricky because you need to combine them |
|
|
|
1073 |
|
00:47:06,160 --> 00:47:10,720 |
|
together and figure out what's going on |
|
|
|
1074 |
|
00:47:07,559 --> 00:47:12,280 |
|
based on that another example that a lot |
|
|
|
1075 |
|
00:47:10,720 --> 00:47:14,160 |
|
of people might not think of immediately |
|
|
|
1076 |
|
00:47:12,280 --> 00:47:17,880 |
|
but is super super common in sentiment |
|
|
|
1077 |
|
00:47:14,160 --> 00:47:20,160 |
|
analysis or any other thing is butt so |
|
|
|
1078 |
|
00:47:17,880 --> 00:47:22,599 |
|
basically what but does is it throws |
|
|
|
1079 |
|
00:47:20,160 --> 00:47:24,160 |
|
away all the stuff that you said before |
|
|
|
1080 |
|
00:47:22,599 --> 00:47:26,119 |
|
um and you can just pay attention to the |
|
|
|
1081 |
|
00:47:24,160 --> 00:47:29,000 |
|
stuff that you saw beforehand so like we |
|
|
|
1082 |
|
00:47:26,119 --> 00:47:30,440 |
|
could even add this to our um like if |
|
|
|
1083 |
|
00:47:29,000 --> 00:47:31,760 |
|
you want to add this to your rule based |
|
|
|
1084 |
|
00:47:30,440 --> 00:47:33,240 |
|
classifier you can do that you just |
|
|
|
1085 |
|
00:47:31,760 --> 00:47:34,640 |
|
search for butt and delete everything |
|
|
|
1086 |
|
00:47:33,240 --> 00:47:37,240 |
|
before it and see if that inputs your |
|
|
|
1087 |
|
00:47:34,640 --> 00:47:39,240 |
|
accuracy might be might be a fun very |
|
|
|
1088 |
|
00:47:37,240 --> 00:47:43,480 |
|
quick thing |
|
|
|
1089 |
|
00:47:39,240 --> 00:47:44,880 |
|
to cool so the better solution which is |
|
|
|
1090 |
|
00:47:43,480 --> 00:47:46,800 |
|
what we're going to talk about for every |
|
|
|
1091 |
|
00:47:44,880 --> 00:47:49,480 |
|
other class other than uh other than |
|
|
|
1092 |
|
00:47:46,800 --> 00:47:52,160 |
|
this one is neural network models and |
|
|
|
1093 |
|
00:47:49,480 --> 00:47:55,800 |
|
basically uh what they do is they do a |
|
|
|
1094 |
|
00:47:52,160 --> 00:47:59,400 |
|
lookup of uh dense word embeddings so |
|
|
|
1095 |
|
00:47:55,800 --> 00:48:02,520 |
|
instead of looking up uh individual uh |
|
|
|
1096 |
|
00:47:59,400 --> 00:48:04,640 |
|
sparse uh vectors individual one hot |
|
|
|
1097 |
|
00:48:02,520 --> 00:48:06,920 |
|
vectors they look up dense word |
|
|
|
1098 |
|
00:48:04,640 --> 00:48:09,680 |
|
embeddings and then throw them into some |
|
|
|
1099 |
|
00:48:06,920 --> 00:48:11,880 |
|
complicated function to extract features |
|
|
|
1100 |
|
00:48:09,680 --> 00:48:16,359 |
|
and based on the features uh multiply by |
|
|
|
1101 |
|
00:48:11,880 --> 00:48:18,280 |
|
weights and get a score um and if you're |
|
|
|
1102 |
|
00:48:16,359 --> 00:48:20,359 |
|
doing text classification in the |
|
|
|
1103 |
|
00:48:18,280 --> 00:48:22,520 |
|
traditional way this is normally what |
|
|
|
1104 |
|
00:48:20,359 --> 00:48:23,760 |
|
you do um if you're doing text |
|
|
|
1105 |
|
00:48:22,520 --> 00:48:25,960 |
|
classification with something like |
|
|
|
1106 |
|
00:48:23,760 --> 00:48:27,280 |
|
prompting you're still actually doing |
|
|
|
1107 |
|
00:48:25,960 --> 00:48:29,960 |
|
this because you're calculating the |
|
|
|
1108 |
|
00:48:27,280 --> 00:48:32,960 |
|
score of the next word to predict and |
|
|
|
1109 |
|
00:48:29,960 --> 00:48:34,720 |
|
that's done in exactly the same way so |
|
|
|
1110 |
|
00:48:32,960 --> 00:48:37,760 |
|
uh even if you're using a large language |
|
|
|
1111 |
|
00:48:34,720 --> 00:48:39,359 |
|
model like GPT this is still probably |
|
|
|
1112 |
|
00:48:37,760 --> 00:48:41,800 |
|
happening under the hood unless open the |
|
|
|
1113 |
|
00:48:39,359 --> 00:48:43,400 |
|
eye invented something that very |
|
|
|
1114 |
|
00:48:41,800 --> 00:48:45,559 |
|
different in Alien than anything else |
|
|
|
1115 |
|
00:48:43,400 --> 00:48:48,440 |
|
that we know of but I I'm guessing that |
|
|
|
1116 |
|
00:48:45,559 --> 00:48:48,440 |
|
that propably hasn't |
|
|
|
1117 |
|
00:48:48,480 --> 00:48:52,880 |
|
happen um one nice thing about neural |
|
|
|
1118 |
|
00:48:50,880 --> 00:48:54,480 |
|
networks is neural networks |
|
|
|
1119 |
|
00:48:52,880 --> 00:48:57,559 |
|
theoretically are powerful enough to |
|
|
|
1120 |
|
00:48:54,480 --> 00:49:00,000 |
|
solve any task if you make them uh deep |
|
|
|
1121 |
|
00:48:57,559 --> 00:49:01,160 |
|
enough or wide enough uh like if you |
|
|
|
1122 |
|
00:49:00,000 --> 00:49:04,520 |
|
make them wide enough and then if you |
|
|
|
1123 |
|
00:49:01,160 --> 00:49:06,799 |
|
make them deep it also helps further so |
|
|
|
1124 |
|
00:49:04,520 --> 00:49:08,079 |
|
anytime somebody says well you can't |
|
|
|
1125 |
|
00:49:06,799 --> 00:49:11,119 |
|
just solve that problem with neural |
|
|
|
1126 |
|
00:49:08,079 --> 00:49:13,240 |
|
networks you know that they're lying |
|
|
|
1127 |
|
00:49:11,119 --> 00:49:15,720 |
|
basically because they theoretically can |
|
|
|
1128 |
|
00:49:13,240 --> 00:49:17,359 |
|
solve every problem uh but you have you |
|
|
|
1129 |
|
00:49:15,720 --> 00:49:19,799 |
|
have issues of data you have issues of |
|
|
|
1130 |
|
00:49:17,359 --> 00:49:23,079 |
|
other things like that so you know they |
|
|
|
1131 |
|
00:49:19,799 --> 00:49:23,079 |
|
don't just necessarily work |
|
|
|
1132 |
|
00:49:23,119 --> 00:49:28,040 |
|
outs cool um so the final thing I'd like |
|
|
|
1133 |
|
00:49:26,400 --> 00:49:29,319 |
|
to talk about is the road map going |
|
|
|
1134 |
|
00:49:28,040 --> 00:49:31,319 |
|
forward some of the things I'm going to |
|
|
|
1135 |
|
00:49:29,319 --> 00:49:32,799 |
|
cover in the class and some of the |
|
|
|
1136 |
|
00:49:31,319 --> 00:49:35,200 |
|
logistics |
|
|
|
1137 |
|
00:49:32,799 --> 00:49:36,799 |
|
issues so um the first thing I'm going |
|
|
|
1138 |
|
00:49:35,200 --> 00:49:38,240 |
|
to talk about in the class is language |
|
|
|
1139 |
|
00:49:36,799 --> 00:49:40,559 |
|
modeling fun |
|
|
|
1140 |
|
00:49:38,240 --> 00:49:42,720 |
|
fundamentals and uh so this could |
|
|
|
1141 |
|
00:49:40,559 --> 00:49:44,240 |
|
include language models uh that just |
|
|
|
1142 |
|
00:49:42,720 --> 00:49:46,559 |
|
predict the next words it could include |
|
|
|
1143 |
|
00:49:44,240 --> 00:49:50,559 |
|
language models that predict the output |
|
|
|
1144 |
|
00:49:46,559 --> 00:49:51,599 |
|
given the uh the input or the prompt um |
|
|
|
1145 |
|
00:49:50,559 --> 00:49:54,559 |
|
I'm going to be talking about |
|
|
|
1146 |
|
00:49:51,599 --> 00:49:56,520 |
|
representing words uh how how we get |
|
|
|
1147 |
|
00:49:54,559 --> 00:49:59,319 |
|
word representation subword models other |
|
|
|
1148 |
|
00:49:56,520 --> 00:50:01,440 |
|
things like that uh then go kind of |
|
|
|
1149 |
|
00:49:59,319 --> 00:50:04,200 |
|
deeper into language modeling uh how do |
|
|
|
1150 |
|
00:50:01,440 --> 00:50:07,799 |
|
we do it how do we evaluate it other |
|
|
|
1151 |
|
00:50:04,200 --> 00:50:10,920 |
|
things um sequence encoding uh and this |
|
|
|
1152 |
|
00:50:07,799 --> 00:50:13,240 |
|
is going to cover things like uh |
|
|
|
1153 |
|
00:50:10,920 --> 00:50:16,280 |
|
Transformers uh self attention modals |
|
|
|
1154 |
|
00:50:13,240 --> 00:50:18,559 |
|
but also very quickly cnns and rnns |
|
|
|
1155 |
|
00:50:16,280 --> 00:50:20,880 |
|
which are useful in some |
|
|
|
1156 |
|
00:50:18,559 --> 00:50:22,200 |
|
cases um and then we're going to |
|
|
|
1157 |
|
00:50:20,880 --> 00:50:24,040 |
|
specifically go very deep into the |
|
|
|
1158 |
|
00:50:22,200 --> 00:50:25,960 |
|
Transformer architecture and also talk a |
|
|
|
1159 |
|
00:50:24,040 --> 00:50:27,280 |
|
little bit about some of the modern uh |
|
|
|
1160 |
|
00:50:25,960 --> 00:50:30,240 |
|
improvements to the Transformer |
|
|
|
1161 |
|
00:50:27,280 --> 00:50:31,839 |
|
architecture so the Transformer we're |
|
|
|
1162 |
|
00:50:30,240 --> 00:50:33,839 |
|
using nowadays is very different than |
|
|
|
1163 |
|
00:50:31,839 --> 00:50:36,200 |
|
the Transformer that was invented in |
|
|
|
1164 |
|
00:50:33,839 --> 00:50:37,240 |
|
2017 uh so we're going to talk well I |
|
|
|
1165 |
|
00:50:36,200 --> 00:50:38,760 |
|
wouldn't say very different but |
|
|
|
1166 |
|
00:50:37,240 --> 00:50:41,359 |
|
different enough that it's important so |
|
|
|
1167 |
|
00:50:38,760 --> 00:50:43,280 |
|
we're going to talk about some of those |
|
|
|
1168 |
|
00:50:41,359 --> 00:50:45,079 |
|
things second thing I'd like to talk |
|
|
|
1169 |
|
00:50:43,280 --> 00:50:47,000 |
|
about is training and inference methods |
|
|
|
1170 |
|
00:50:45,079 --> 00:50:48,839 |
|
so this includes uh generation |
|
|
|
1171 |
|
00:50:47,000 --> 00:50:52,119 |
|
algorithms uh so we're going to have a |
|
|
|
1172 |
|
00:50:48,839 --> 00:50:55,520 |
|
whole class on how we generate text uh |
|
|
|
1173 |
|
00:50:52,119 --> 00:50:58,319 |
|
in different ways uh prompting how uh we |
|
|
|
1174 |
|
00:50:55,520 --> 00:50:59,720 |
|
can prompt things I hear uh world class |
|
|
|
1175 |
|
00:50:58,319 --> 00:51:01,799 |
|
prompt engineers make a lot of money |
|
|
|
1176 |
|
00:50:59,720 --> 00:51:05,480 |
|
nowadays so uh you'll want to pay |
|
|
|
1177 |
|
00:51:01,799 --> 00:51:08,760 |
|
attention to that one um and instruction |
|
|
|
1178 |
|
00:51:05,480 --> 00:51:11,520 |
|
tuning uh so how do we train models to |
|
|
|
1179 |
|
00:51:08,760 --> 00:51:13,720 |
|
handle a lot of different tasks and |
|
|
|
1180 |
|
00:51:11,520 --> 00:51:15,839 |
|
reinforcement learning so how do we uh |
|
|
|
1181 |
|
00:51:13,720 --> 00:51:18,520 |
|
you know like actually generate outputs |
|
|
|
1182 |
|
00:51:15,839 --> 00:51:19,839 |
|
uh kind of Judge them and then learn |
|
|
|
1183 |
|
00:51:18,520 --> 00:51:22,599 |
|
from |
|
|
|
1184 |
|
00:51:19,839 --> 00:51:25,880 |
|
there also experimental design and |
|
|
|
1185 |
|
00:51:22,599 --> 00:51:28,079 |
|
evaluation so experimental design uh so |
|
|
|
1186 |
|
00:51:25,880 --> 00:51:30,480 |
|
how do we design an experiment well uh |
|
|
|
1187 |
|
00:51:28,079 --> 00:51:32,000 |
|
so that it backs up what we want to be |
|
|
|
1188 |
|
00:51:30,480 --> 00:51:34,559 |
|
uh our conclusions that we want to be |
|
|
|
1189 |
|
00:51:32,000 --> 00:51:37,000 |
|
backing up how do we do human annotation |
|
|
|
1190 |
|
00:51:34,559 --> 00:51:38,880 |
|
of data in a reliable way this is |
|
|
|
1191 |
|
00:51:37,000 --> 00:51:41,160 |
|
getting harder and harder as models get |
|
|
|
1192 |
|
00:51:38,880 --> 00:51:43,359 |
|
better and better because uh getting |
|
|
|
1193 |
|
00:51:41,160 --> 00:51:45,000 |
|
humans who don't care very much about |
|
|
|
1194 |
|
00:51:43,359 --> 00:51:48,559 |
|
The annotation task they might do worse |
|
|
|
1195 |
|
00:51:45,000 --> 00:51:51,119 |
|
than gp4 so um you need to be careful of |
|
|
|
1196 |
|
00:51:48,559 --> 00:51:52,240 |
|
that also debugging and interpretation |
|
|
|
1197 |
|
00:51:51,119 --> 00:51:53,960 |
|
technique so what are some of the |
|
|
|
1198 |
|
00:51:52,240 --> 00:51:55,160 |
|
automatic techniques that you can do to |
|
|
|
1199 |
|
00:51:53,960 --> 00:51:57,720 |
|
quickly figure out what's going wrong |
|
|
|
1200 |
|
00:51:55,160 --> 00:52:00,040 |
|
with your models and improve |
|
|
|
1201 |
|
00:51:57,720 --> 00:52:01,599 |
|
them and uh bias and fairness |
|
|
|
1202 |
|
00:52:00,040 --> 00:52:04,200 |
|
considerations so it's really really |
|
|
|
1203 |
|
00:52:01,599 --> 00:52:05,799 |
|
important nowadays uh that models are |
|
|
|
1204 |
|
00:52:04,200 --> 00:52:07,880 |
|
being deployed to real people in the |
|
|
|
1205 |
|
00:52:05,799 --> 00:52:09,880 |
|
real world and like actually causing |
|
|
|
1206 |
|
00:52:07,880 --> 00:52:11,760 |
|
harm to people in some cases that we |
|
|
|
1207 |
|
00:52:09,880 --> 00:52:15,160 |
|
need to be worried about |
|
|
|
1208 |
|
00:52:11,760 --> 00:52:17,000 |
|
that Advanced Training in architectures |
|
|
|
1209 |
|
00:52:15,160 --> 00:52:19,280 |
|
so we're going to talk about distill |
|
|
|
1210 |
|
00:52:17,000 --> 00:52:21,400 |
|
distillation and quantization how can we |
|
|
|
1211 |
|
00:52:19,280 --> 00:52:23,520 |
|
make small language models uh that |
|
|
|
1212 |
|
00:52:21,400 --> 00:52:24,880 |
|
actually still work well like not large |
|
|
|
1213 |
|
00:52:23,520 --> 00:52:27,559 |
|
you can run them on your phone you can |
|
|
|
1214 |
|
00:52:24,880 --> 00:52:29,920 |
|
run them on your local |
|
|
|
1215 |
|
00:52:27,559 --> 00:52:31,640 |
|
laptop um ensembling and mixtures of |
|
|
|
1216 |
|
00:52:29,920 --> 00:52:33,480 |
|
experts how can we combine together |
|
|
|
1217 |
|
00:52:31,640 --> 00:52:34,760 |
|
multiple models in order to create |
|
|
|
1218 |
|
00:52:33,480 --> 00:52:35,880 |
|
models that are better than the sum of |
|
|
|
1219 |
|
00:52:34,760 --> 00:52:38,799 |
|
their |
|
|
|
1220 |
|
00:52:35,880 --> 00:52:40,720 |
|
parts and um retrieval and retrieval |
|
|
|
1221 |
|
00:52:38,799 --> 00:52:43,920 |
|
augmented |
|
|
|
1222 |
|
00:52:40,720 --> 00:52:45,480 |
|
generation long sequence models uh so |
|
|
|
1223 |
|
00:52:43,920 --> 00:52:49,920 |
|
how do we handle long |
|
|
|
1224 |
|
00:52:45,480 --> 00:52:52,240 |
|
outputs um and uh we're going to talk |
|
|
|
1225 |
|
00:52:49,920 --> 00:52:55,760 |
|
about applications to complex reasoning |
|
|
|
1226 |
|
00:52:52,240 --> 00:52:57,760 |
|
tasks code generation language agents |
|
|
|
1227 |
|
00:52:55,760 --> 00:52:59,920 |
|
and knowledge-based QA and information |
|
|
|
1228 |
|
00:52:57,760 --> 00:53:04,160 |
|
extraction I picked |
|
|
|
1229 |
|
00:52:59,920 --> 00:53:06,760 |
|
these because they seem to be maybe the |
|
|
|
1230 |
|
00:53:04,160 --> 00:53:09,880 |
|
most important at least in research |
|
|
|
1231 |
|
00:53:06,760 --> 00:53:11,440 |
|
nowadays and also they cover uh the |
|
|
|
1232 |
|
00:53:09,880 --> 00:53:13,640 |
|
things that when I talk to people in |
|
|
|
1233 |
|
00:53:11,440 --> 00:53:15,280 |
|
Industry are kind of most interested in |
|
|
|
1234 |
|
00:53:13,640 --> 00:53:17,559 |
|
so hopefully it'll be useful regardless |
|
|
|
1235 |
|
00:53:15,280 --> 00:53:19,799 |
|
of uh whether you plan on doing research |
|
|
|
1236 |
|
00:53:17,559 --> 00:53:22,839 |
|
or or plan on doing industry related |
|
|
|
1237 |
|
00:53:19,799 --> 00:53:24,160 |
|
things uh by by the way the two things |
|
|
|
1238 |
|
00:53:22,839 --> 00:53:25,920 |
|
that when I talk to people in Industry |
|
|
|
1239 |
|
00:53:24,160 --> 00:53:29,599 |
|
they're most interested in are Rag and |
|
|
|
1240 |
|
00:53:25,920 --> 00:53:31,079 |
|
code generation at the moment for now um |
|
|
|
1241 |
|
00:53:29,599 --> 00:53:32,319 |
|
so those are ones that you'll want to |
|
|
|
1242 |
|
00:53:31,079 --> 00:53:34,680 |
|
pay attention |
|
|
|
1243 |
|
00:53:32,319 --> 00:53:36,599 |
|
to and then finally we have a few |
|
|
|
1244 |
|
00:53:34,680 --> 00:53:40,079 |
|
lectures on Linguistics and |
|
|
|
1245 |
|
00:53:36,599 --> 00:53:42,720 |
|
multilinguality um I love Linguistics |
|
|
|
1246 |
|
00:53:40,079 --> 00:53:44,839 |
|
but uh to be honest at the moment most |
|
|
|
1247 |
|
00:53:42,720 --> 00:53:47,760 |
|
of our Cutting Edge models don't |
|
|
|
1248 |
|
00:53:44,839 --> 00:53:49,240 |
|
explicitly use linguistic structure um |
|
|
|
1249 |
|
00:53:47,760 --> 00:53:50,799 |
|
but I still think it's useful to know |
|
|
|
1250 |
|
00:53:49,240 --> 00:53:52,760 |
|
about it especially if you're working on |
|
|
|
1251 |
|
00:53:50,799 --> 00:53:54,880 |
|
multilingual things especially if you're |
|
|
|
1252 |
|
00:53:52,760 --> 00:53:57,040 |
|
interested in very robust generalization |
|
|
|
1253 |
|
00:53:54,880 --> 00:53:58,920 |
|
to new models so we're going to talk a |
|
|
|
1254 |
|
00:53:57,040 --> 00:54:02,599 |
|
little bit about that and also |
|
|
|
1255 |
|
00:53:58,920 --> 00:54:06,079 |
|
multilingual LP I'm going to have |
|
|
|
1256 |
|
00:54:02,599 --> 00:54:09,119 |
|
fure so also if you have any suggestions |
|
|
|
1257 |
|
00:54:06,079 --> 00:54:11,400 |
|
um we have two guest lecture slots still |
|
|
|
1258 |
|
00:54:09,119 --> 00:54:12,799 |
|
open uh that I'm trying to fill so if |
|
|
|
1259 |
|
00:54:11,400 --> 00:54:15,440 |
|
you have any things that you really want |
|
|
|
1260 |
|
00:54:12,799 --> 00:54:16,440 |
|
to hear about um I could either add them |
|
|
|
1261 |
|
00:54:15,440 --> 00:54:19,319 |
|
to the |
|
|
|
1262 |
|
00:54:16,440 --> 00:54:21,079 |
|
existing you know content or I could |
|
|
|
1263 |
|
00:54:19,319 --> 00:54:23,240 |
|
invite a guest lecturer who's working on |
|
|
|
1264 |
|
00:54:21,079 --> 00:54:24,079 |
|
that topic so you know please feel free |
|
|
|
1265 |
|
00:54:23,240 --> 00:54:26,760 |
|
to tell |
|
|
|
1266 |
|
00:54:24,079 --> 00:54:29,160 |
|
me um then the class format and |
|
|
|
1267 |
|
00:54:26,760 --> 00:54:32,280 |
|
structure uh the class |
|
|
|
1268 |
|
00:54:29,160 --> 00:54:34,000 |
|
content my goal is to learn in detail |
|
|
|
1269 |
|
00:54:32,280 --> 00:54:36,640 |
|
about building NLP systems from a |
|
|
|
1270 |
|
00:54:34,000 --> 00:54:40,520 |
|
research perspective so this is a 700 |
|
|
|
1271 |
|
00:54:36,640 --> 00:54:43,599 |
|
level course so it's aiming to be for |
|
|
|
1272 |
|
00:54:40,520 --> 00:54:46,960 |
|
people who really want to try new and |
|
|
|
1273 |
|
00:54:43,599 --> 00:54:49,280 |
|
Innovative things in uh kind of natural |
|
|
|
1274 |
|
00:54:46,960 --> 00:54:51,359 |
|
language processing it's not going to |
|
|
|
1275 |
|
00:54:49,280 --> 00:54:52,760 |
|
focus solely on reimplementing things |
|
|
|
1276 |
|
00:54:51,359 --> 00:54:54,319 |
|
that have been done before including in |
|
|
|
1277 |
|
00:54:52,760 --> 00:54:55,280 |
|
the project I'm going to be expecting |
|
|
|
1278 |
|
00:54:54,319 --> 00:54:58,480 |
|
everybody to do something something |
|
|
|
1279 |
|
00:54:55,280 --> 00:54:59,920 |
|
that's kind of new whether it's coming |
|
|
|
1280 |
|
00:54:58,480 --> 00:55:01,359 |
|
up with a new method or applying |
|
|
|
1281 |
|
00:54:59,920 --> 00:55:03,559 |
|
existing methods to a place where they |
|
|
|
1282 |
|
00:55:01,359 --> 00:55:05,079 |
|
haven't been used before or building out |
|
|
|
1283 |
|
00:55:03,559 --> 00:55:06,640 |
|
things for a new language or something |
|
|
|
1284 |
|
00:55:05,079 --> 00:55:08,359 |
|
like that so that's kind of one of the |
|
|
|
1285 |
|
00:55:06,640 --> 00:55:11,480 |
|
major goals of this |
|
|
|
1286 |
|
00:55:08,359 --> 00:55:13,000 |
|
class um learn basic and advanced topics |
|
|
|
1287 |
|
00:55:11,480 --> 00:55:15,559 |
|
in machine learning approaches to NLP |
|
|
|
1288 |
|
00:55:13,000 --> 00:55:18,359 |
|
and language models learn some basic |
|
|
|
1289 |
|
00:55:15,559 --> 00:55:21,480 |
|
linguistic knowledge useful in NLP uh |
|
|
|
1290 |
|
00:55:18,359 --> 00:55:23,200 |
|
see case studies of NLP applications and |
|
|
|
1291 |
|
00:55:21,480 --> 00:55:25,680 |
|
learn how to identify unique problems |
|
|
|
1292 |
|
00:55:23,200 --> 00:55:29,039 |
|
for each um one thing i' like to point |
|
|
|
1293 |
|
00:55:25,680 --> 00:55:31,160 |
|
out is I'm not going to cover every NLP |
|
|
|
1294 |
|
00:55:29,039 --> 00:55:32,920 |
|
application ever because that would be |
|
|
|
1295 |
|
00:55:31,160 --> 00:55:35,520 |
|
absolutely impossible NLP is being used |
|
|
|
1296 |
|
00:55:32,920 --> 00:55:37,079 |
|
in so many different areas nowadays but |
|
|
|
1297 |
|
00:55:35,520 --> 00:55:38,960 |
|
what I want people to pay attention to |
|
|
|
1298 |
|
00:55:37,079 --> 00:55:41,280 |
|
like even if you're not super interested |
|
|
|
1299 |
|
00:55:38,960 --> 00:55:42,400 |
|
in code generation for example what you |
|
|
|
1300 |
|
00:55:41,280 --> 00:55:44,200 |
|
can do is you can look at code |
|
|
|
1301 |
|
00:55:42,400 --> 00:55:46,160 |
|
generation look at how people identify |
|
|
|
1302 |
|
00:55:44,200 --> 00:55:47,680 |
|
problems look at the methods that people |
|
|
|
1303 |
|
00:55:46,160 --> 00:55:50,880 |
|
have proposed to solve those unique |
|
|
|
1304 |
|
00:55:47,680 --> 00:55:53,039 |
|
problems and then kind of map that try |
|
|
|
1305 |
|
00:55:50,880 --> 00:55:54,799 |
|
to do some generalization onto your own |
|
|
|
1306 |
|
00:55:53,039 --> 00:55:57,799 |
|
problems of Interest so uh that's kind |
|
|
|
1307 |
|
00:55:54,799 --> 00:56:00,280 |
|
of the goal of the NLP |
|
|
|
1308 |
|
00:55:57,799 --> 00:56:02,440 |
|
applications finally uh learning how to |
|
|
|
1309 |
|
00:56:00,280 --> 00:56:05,160 |
|
debug when and where NLP systems fail |
|
|
|
1310 |
|
00:56:02,440 --> 00:56:08,200 |
|
and build improvements based on this so |
|
|
|
1311 |
|
00:56:05,160 --> 00:56:10,200 |
|
um ever since I was a graduate student |
|
|
|
1312 |
|
00:56:08,200 --> 00:56:12,720 |
|
this has been like one of the really |
|
|
|
1313 |
|
00:56:10,200 --> 00:56:15,920 |
|
important things that I feel like I've |
|
|
|
1314 |
|
00:56:12,720 --> 00:56:17,440 |
|
done well or done better than some other |
|
|
|
1315 |
|
00:56:15,920 --> 00:56:19,280 |
|
people and I I feel like it's a really |
|
|
|
1316 |
|
00:56:17,440 --> 00:56:21,119 |
|
good way to like even if you're only |
|
|
|
1317 |
|
00:56:19,280 --> 00:56:22,680 |
|
interested in improving accuracy knowing |
|
|
|
1318 |
|
00:56:21,119 --> 00:56:25,039 |
|
why your system's failing still is the |
|
|
|
1319 |
|
00:56:22,680 --> 00:56:27,599 |
|
best way to do that I so I'm going to |
|
|
|
1320 |
|
00:56:25,039 --> 00:56:30,559 |
|
put a lot of emphasis on |
|
|
|
1321 |
|
00:56:27,599 --> 00:56:32,559 |
|
that in terms of the class format um |
|
|
|
1322 |
|
00:56:30,559 --> 00:56:36,280 |
|
before class for some classes there are |
|
|
|
1323 |
|
00:56:32,559 --> 00:56:37,880 |
|
recommended reading uh this can be |
|
|
|
1324 |
|
00:56:36,280 --> 00:56:39,559 |
|
helpful to read I'm never going to |
|
|
|
1325 |
|
00:56:37,880 --> 00:56:41,119 |
|
expect you to definitely have read it |
|
|
|
1326 |
|
00:56:39,559 --> 00:56:42,480 |
|
before the class but I would suggest |
|
|
|
1327 |
|
00:56:41,119 --> 00:56:45,160 |
|
that maybe you'll get more out of the |
|
|
|
1328 |
|
00:56:42,480 --> 00:56:47,319 |
|
class if you do that um during class |
|
|
|
1329 |
|
00:56:45,160 --> 00:56:48,079 |
|
we'll have the lecture um in discussion |
|
|
|
1330 |
|
00:56:47,319 --> 00:56:50,559 |
|
with |
|
|
|
1331 |
|
00:56:48,079 --> 00:56:52,359 |
|
everybody um sometimes we'll have a code |
|
|
|
1332 |
|
00:56:50,559 --> 00:56:55,839 |
|
or data walk |
|
|
|
1333 |
|
00:56:52,359 --> 00:56:58,760 |
|
um actually this is a a little bit old I |
|
|
|
1334 |
|
00:56:55,839 --> 00:57:01,880 |
|
I have this slide we're this year we're |
|
|
|
1335 |
|
00:56:58,760 --> 00:57:04,160 |
|
going to be adding more uh code and data |
|
|
|
1336 |
|
00:57:01,880 --> 00:57:07,400 |
|
walks during office hours and the way it |
|
|
|
1337 |
|
00:57:04,160 --> 00:57:09,400 |
|
will work is one of the Tas we have |
|
|
|
1338 |
|
00:57:07,400 --> 00:57:11,160 |
|
seven Tas who I'm going to introduce |
|
|
|
1339 |
|
00:57:09,400 --> 00:57:15,000 |
|
very soon but one of the Tas will be |
|
|
|
1340 |
|
00:57:11,160 --> 00:57:16,839 |
|
doing this kind of recitation where you |
|
|
|
1341 |
|
00:57:15,000 --> 00:57:18,200 |
|
um where we go over a library so if |
|
|
|
1342 |
|
00:57:16,839 --> 00:57:19,480 |
|
you're not familiar with the library and |
|
|
|
1343 |
|
00:57:18,200 --> 00:57:21,960 |
|
you want to be more familiar with the |
|
|
|
1344 |
|
00:57:19,480 --> 00:57:23,720 |
|
library you can join this and uh then |
|
|
|
1345 |
|
00:57:21,960 --> 00:57:25,400 |
|
we'll be able to do this and this will |
|
|
|
1346 |
|
00:57:23,720 --> 00:57:28,240 |
|
cover things like |
|
|
|
1347 |
|
00:57:25,400 --> 00:57:31,039 |
|
um pie torch and sentence piece uh we're |
|
|
|
1348 |
|
00:57:28,240 --> 00:57:33,280 |
|
going to start out with hugging face um |
|
|
|
1349 |
|
00:57:31,039 --> 00:57:36,559 |
|
inference stuff like |
|
|
|
1350 |
|
00:57:33,280 --> 00:57:41,520 |
|
VM uh debugging software like |
|
|
|
1351 |
|
00:57:36,559 --> 00:57:41,520 |
|
Xeno um what were the other |
|
|
|
1352 |
|
00:57:41,960 --> 00:57:47,200 |
|
ones oh the open AI API and light llm |
|
|
|
1353 |
|
00:57:45,680 --> 00:57:50,520 |
|
other stuff like that so we we have lots |
|
|
|
1354 |
|
00:57:47,200 --> 00:57:53,599 |
|
of them planned we'll uh uh we'll update |
|
|
|
1355 |
|
00:57:50,520 --> 00:57:54,839 |
|
that um and then after class after |
|
|
|
1356 |
|
00:57:53,599 --> 00:57:58,079 |
|
almost every class we'll have a question |
|
|
|
1357 |
|
00:57:54,839 --> 00:58:00,079 |
|
quiz um and the quiz is intended to just |
|
|
|
1358 |
|
00:57:58,079 --> 00:58:02,000 |
|
you know make sure that you uh paid |
|
|
|
1359 |
|
00:58:00,079 --> 00:58:04,480 |
|
attention to the material and are able |
|
|
|
1360 |
|
00:58:02,000 --> 00:58:07,520 |
|
to answer questions about it we will aim |
|
|
|
1361 |
|
00:58:04,480 --> 00:58:09,559 |
|
to release it on the day of the course |
|
|
|
1362 |
|
00:58:07,520 --> 00:58:11,599 |
|
the day of the actual lecture and it |
|
|
|
1363 |
|
00:58:09,559 --> 00:58:14,559 |
|
will be due at the end of the following |
|
|
|
1364 |
|
00:58:11,599 --> 00:58:15,960 |
|
day of the lecture so um it will be |
|
|
|
1365 |
|
00:58:14,559 --> 00:58:18,920 |
|
three questions it probably shouldn't |
|
|
|
1366 |
|
00:58:15,960 --> 00:58:20,680 |
|
take a whole lot of time but um uh yeah |
|
|
|
1367 |
|
00:58:18,920 --> 00:58:23,400 |
|
so we'll H |
|
|
|
1368 |
|
00:58:20,680 --> 00:58:26,319 |
|
that in terms of assignments assignment |
|
|
|
1369 |
|
00:58:23,400 --> 00:58:28,640 |
|
one is going to be build your own llama |
|
|
|
1370 |
|
00:58:26,319 --> 00:58:30,200 |
|
and so what this is going to look like |
|
|
|
1371 |
|
00:58:28,640 --> 00:58:32,680 |
|
is we're going to give you a partial |
|
|
|
1372 |
|
00:58:30,200 --> 00:58:34,319 |
|
implementation of llama which is kind of |
|
|
|
1373 |
|
00:58:32,680 --> 00:58:37,960 |
|
the most popular open source language |
|
|
|
1374 |
|
00:58:34,319 --> 00:58:40,160 |
|
model nowadays and ask you to fill in um |
|
|
|
1375 |
|
00:58:37,960 --> 00:58:42,839 |
|
ask you to fill in the parts we're going |
|
|
|
1376 |
|
00:58:40,160 --> 00:58:45,920 |
|
to train a very small version of llama |
|
|
|
1377 |
|
00:58:42,839 --> 00:58:47,319 |
|
on a small data set and get it to work |
|
|
|
1378 |
|
00:58:45,920 --> 00:58:48,880 |
|
and the reason why it's very small is |
|
|
|
1379 |
|
00:58:47,319 --> 00:58:50,480 |
|
because the smallest actual version of |
|
|
|
1380 |
|
00:58:48,880 --> 00:58:53,039 |
|
llama is 7 billion |
|
|
|
1381 |
|
00:58:50,480 --> 00:58:55,359 |
|
parameters um and that might be a little |
|
|
|
1382 |
|
00:58:53,039 --> 00:58:58,400 |
|
bit difficult to train with |
|
|
|
1383 |
|
00:58:55,359 --> 00:59:00,680 |
|
resources um for assignment two we're |
|
|
|
1384 |
|
00:58:58,400 --> 00:59:04,559 |
|
going to try to do an NLP task from |
|
|
|
1385 |
|
00:59:00,680 --> 00:59:06,920 |
|
scratch and so the way this will work is |
|
|
|
1386 |
|
00:59:04,559 --> 00:59:08,520 |
|
we're going to give you an assignment |
|
|
|
1387 |
|
00:59:06,920 --> 00:59:10,880 |
|
which we're not going to give you an |
|
|
|
1388 |
|
00:59:08,520 --> 00:59:13,400 |
|
actual data set and instead we're going |
|
|
|
1389 |
|
00:59:10,880 --> 00:59:15,760 |
|
to ask you to uh perform data creation |
|
|
|
1390 |
|
00:59:13,400 --> 00:59:19,359 |
|
modeling and evaluation for a specified |
|
|
|
1391 |
|
00:59:15,760 --> 00:59:20,640 |
|
task and so we're going to tell you uh |
|
|
|
1392 |
|
00:59:19,359 --> 00:59:22,599 |
|
what to do but we're not going to tell |
|
|
|
1393 |
|
00:59:20,640 --> 00:59:26,400 |
|
you exactly how to do it but we're going |
|
|
|
1394 |
|
00:59:22,599 --> 00:59:29,680 |
|
to try to give as conrete directions as |
|
|
|
1395 |
|
00:59:26,400 --> 00:59:32,359 |
|
we can um |
|
|
|
1396 |
|
00:59:29,680 --> 00:59:34,160 |
|
yeah will you be given a parameter limit |
|
|
|
1397 |
|
00:59:32,359 --> 00:59:36,559 |
|
on the model so that's a good question |
|
|
|
1398 |
|
00:59:34,160 --> 00:59:39,119 |
|
or like a expense limit or something |
|
|
|
1399 |
|
00:59:36,559 --> 00:59:40,440 |
|
like that um I maybe actually I should |
|
|
|
1400 |
|
00:59:39,119 --> 00:59:44,240 |
|
take a break from the assignments and |
|
|
|
1401 |
|
00:59:40,440 --> 00:59:46,520 |
|
talk about compute so right now um for |
|
|
|
1402 |
|
00:59:44,240 --> 00:59:49,319 |
|
assignment one we're planning on having |
|
|
|
1403 |
|
00:59:46,520 --> 00:59:51,599 |
|
this be able to be done either on a Mac |
|
|
|
1404 |
|
00:59:49,319 --> 00:59:53,520 |
|
laptop with an M1 or M2 processor which |
|
|
|
1405 |
|
00:59:51,599 --> 00:59:57,079 |
|
I think a lot of people have or Google |
|
|
|
1406 |
|
00:59:53,520 --> 00:59:59,839 |
|
collab um so it should be like |
|
|
|
1407 |
|
00:59:57,079 --> 01:00:02,160 |
|
sufficient to use free computational |
|
|
|
1408 |
|
00:59:59,839 --> 01:00:03,640 |
|
resources that you have for number two |
|
|
|
1409 |
|
01:00:02,160 --> 01:00:06,079 |
|
we'll think about that I think that's |
|
|
|
1410 |
|
01:00:03,640 --> 01:00:08,280 |
|
important we do have Google cloud |
|
|
|
1411 |
|
01:00:06,079 --> 01:00:11,520 |
|
credits for $50 for everybody and I'm |
|
|
|
1412 |
|
01:00:08,280 --> 01:00:13,440 |
|
working to get AWS credits for more um |
|
|
|
1413 |
|
01:00:11,520 --> 01:00:18,160 |
|
but the cloud providers nowadays are |
|
|
|
1414 |
|
01:00:13,440 --> 01:00:19,680 |
|
being very stingy so um so it's uh been |
|
|
|
1415 |
|
01:00:18,160 --> 01:00:22,160 |
|
a little bit of a fight to get uh |
|
|
|
1416 |
|
01:00:19,680 --> 01:00:23,680 |
|
credits but I I it is very important so |
|
|
|
1417 |
|
01:00:22,160 --> 01:00:28,480 |
|
I'm going to try to get as as many as we |
|
|
|
1418 |
|
01:00:23,680 --> 01:00:31,119 |
|
can um and so yeah I I think basically |
|
|
|
1419 |
|
01:00:28,480 --> 01:00:32,280 |
|
uh there will be some sort of like limit |
|
|
|
1420 |
|
01:00:31,119 --> 01:00:34,480 |
|
on the amount of things you can |
|
|
|
1421 |
|
01:00:32,280 --> 01:00:36,240 |
|
practically do and so because of that |
|
|
|
1422 |
|
01:00:34,480 --> 01:00:39,920 |
|
I'm hoping that people will rely very |
|
|
|
1423 |
|
01:00:36,240 --> 01:00:43,359 |
|
heavily on pre-trained models um or uh |
|
|
|
1424 |
|
01:00:39,920 --> 01:00:46,079 |
|
yeah pre-trained models |
|
|
|
1425 |
|
01:00:43,359 --> 01:00:49,599 |
|
and yeah so that that's the the short |
|
|
|
1426 |
|
01:00:46,079 --> 01:00:52,799 |
|
story B um the second thing uh the |
|
|
|
1427 |
|
01:00:49,599 --> 01:00:54,720 |
|
assignment three is to do a survey of |
|
|
|
1428 |
|
01:00:52,799 --> 01:00:57,920 |
|
some sort of state-ofthe-art research |
|
|
|
1429 |
|
01:00:54,720 --> 01:01:00,760 |
|
resarch and do a reimplementation of |
|
|
|
1430 |
|
01:00:57,920 --> 01:01:02,000 |
|
this and in doing this again you will |
|
|
|
1431 |
|
01:01:00,760 --> 01:01:03,440 |
|
have to think about something that's |
|
|
|
1432 |
|
01:01:02,000 --> 01:01:06,359 |
|
feasible within computational |
|
|
|
1433 |
|
01:01:03,440 --> 01:01:08,680 |
|
constraints um and so you can discuss |
|
|
|
1434 |
|
01:01:06,359 --> 01:01:11,839 |
|
with your Tas about uh about the best |
|
|
|
1435 |
|
01:01:08,680 --> 01:01:13,920 |
|
way to do this um and then the final |
|
|
|
1436 |
|
01:01:11,839 --> 01:01:15,400 |
|
project is to perform a unique project |
|
|
|
1437 |
|
01:01:13,920 --> 01:01:17,559 |
|
that either improves on the state-of-the |
|
|
|
1438 |
|
01:01:15,400 --> 01:01:21,000 |
|
art with respect to whatever you would |
|
|
|
1439 |
|
01:01:17,559 --> 01:01:23,440 |
|
like to improve with this could be uh |
|
|
|
1440 |
|
01:01:21,000 --> 01:01:25,280 |
|
accuracy for sure this could be |
|
|
|
1441 |
|
01:01:23,440 --> 01:01:27,760 |
|
efficiency |
|
|
|
1442 |
|
01:01:25,280 --> 01:01:29,599 |
|
it could be some sense of |
|
|
|
1443 |
|
01:01:27,760 --> 01:01:31,520 |
|
interpretability but if it's going to be |
|
|
|
1444 |
|
01:01:29,599 --> 01:01:33,599 |
|
something like interpretability you'll |
|
|
|
1445 |
|
01:01:31,520 --> 01:01:35,440 |
|
have to discuss with us what that means |
|
|
|
1446 |
|
01:01:33,599 --> 01:01:37,240 |
|
like how we measure that how we can like |
|
|
|
1447 |
|
01:01:35,440 --> 01:01:40,839 |
|
actually say that you did a good job |
|
|
|
1448 |
|
01:01:37,240 --> 01:01:42,839 |
|
with improving that um another thing |
|
|
|
1449 |
|
01:01:40,839 --> 01:01:44,680 |
|
that you can do is take whatever you |
|
|
|
1450 |
|
01:01:42,839 --> 01:01:47,280 |
|
implemented for assignment 3 and apply |
|
|
|
1451 |
|
01:01:44,680 --> 01:01:49,039 |
|
it to a new task or apply it to a new |
|
|
|
1452 |
|
01:01:47,280 --> 01:01:50,760 |
|
language that has never been examined |
|
|
|
1453 |
|
01:01:49,039 --> 01:01:53,119 |
|
before so these are also acceptable |
|
|
|
1454 |
|
01:01:50,760 --> 01:01:54,240 |
|
final projects but basically the idea is |
|
|
|
1455 |
|
01:01:53,119 --> 01:01:55,559 |
|
for the final project you need to do |
|
|
|
1456 |
|
01:01:54,240 --> 01:01:57,480 |
|
something something new that hasn't been |
|
|
|
1457 |
|
01:01:55,559 --> 01:01:59,880 |
|
done before and create new knowledge |
|
|
|
1458 |
|
01:01:57,480 --> 01:02:04,520 |
|
with the respect |
|
|
|
1459 |
|
01:01:59,880 --> 01:02:07,640 |
|
toy um so for this the instructor is me |
|
|
|
1460 |
|
01:02:04,520 --> 01:02:09,920 |
|
um I'm uh looking forward to you know |
|
|
|
1461 |
|
01:02:07,640 --> 01:02:13,599 |
|
discussing and working with all of you |
|
|
|
1462 |
|
01:02:09,920 --> 01:02:16,119 |
|
um for TAS we have seven Tas uh two of |
|
|
|
1463 |
|
01:02:13,599 --> 01:02:18,319 |
|
them are in transit so they're not here |
|
|
|
1464 |
|
01:02:16,119 --> 01:02:22,279 |
|
today um the other ones uh Tas would you |
|
|
|
1465 |
|
01:02:18,319 --> 01:02:22,279 |
|
mind coming up uh to introduce |
|
|
|
1466 |
|
01:02:23,359 --> 01:02:26,359 |
|
yourself |
|
|
|
1467 |
|
01:02:28,400 --> 01:02:32,839 |
|
so um yeah nhir and akshai couldn't be |
|
|
|
1468 |
|
01:02:31,599 --> 01:02:34,039 |
|
here today because they're traveling |
|
|
|
1469 |
|
01:02:32,839 --> 01:02:37,119 |
|
I'll introduce them later because |
|
|
|
1470 |
|
01:02:34,039 --> 01:02:37,119 |
|
they're coming uh next |
|
|
|
1471 |
|
01:02:40,359 --> 01:02:46,480 |
|
time cool and what I'd like everybody to |
|
|
|
1472 |
|
01:02:43,000 --> 01:02:48,680 |
|
do is say um like you know what your |
|
|
|
1473 |
|
01:02:46,480 --> 01:02:53,079 |
|
name is uh what |
|
|
|
1474 |
|
01:02:48,680 --> 01:02:55,799 |
|
your like maybe what you're interested |
|
|
|
1475 |
|
01:02:53,079 --> 01:02:57,319 |
|
in um and the reason the goal of this is |
|
|
|
1476 |
|
01:02:55,799 --> 01:02:59,200 |
|
number one for everybody to know who you |
|
|
|
1477 |
|
01:02:57,319 --> 01:03:00,720 |
|
are and number two for everybody to know |
|
|
|
1478 |
|
01:02:59,200 --> 01:03:03,440 |
|
who the best person to talk to is if |
|
|
|
1479 |
|
01:03:00,720 --> 01:03:03,440 |
|
they're interested in |
|
|
|
1480 |
|
01:03:04,200 --> 01:03:09,079 |
|
particular hi uh I'm |
|
|
|
1481 |
|
01:03:07,000 --> 01:03:15,520 |
|
Aila second |
|
|
|
1482 |
|
01:03:09,079 --> 01:03:15,520 |
|
year I work on language and social |
|
|
|
1483 |
|
01:03:16,200 --> 01:03:24,559 |
|
and I'm I'm a second this year PhD |
|
|
|
1484 |
|
01:03:21,160 --> 01:03:26,799 |
|
student Grand and Shar with you I search |
|
|
|
1485 |
|
01:03:24,559 --> 01:03:28,480 |
|
is like started in the border of MP and |
|
|
|
1486 |
|
01:03:26,799 --> 01:03:31,000 |
|
computer interaction with a lot of work |
|
|
|
1487 |
|
01:03:28,480 --> 01:03:32,640 |
|
on automating parts of the developer |
|
|
|
1488 |
|
01:03:31,000 --> 01:03:35,319 |
|
experience to make it easier for anyone |
|
|
|
1489 |
|
01:03:32,640 --> 01:03:35,319 |
|
to |
|
|
|
1490 |
|
01:03:39,090 --> 01:03:42,179 |
|
[Music] |
|
|
|
1491 |
|
01:03:47,520 --> 01:03:53,279 |
|
orif |
|
|
|
1492 |
|
01:03:50,079 --> 01:03:54,680 |
|
everyone first |
|
|
|
1493 |
|
01:03:53,279 --> 01:03:57,119 |
|
year |
|
|
|
1494 |
|
01:03:54,680 --> 01:04:00,119 |
|
[Music] |
|
|
|
1495 |
|
01:03:57,119 --> 01:04:03,559 |
|
I don't like updating primar models I |
|
|
|
1496 |
|
01:04:00,119 --> 01:04:03,559 |
|
hope to not update Prim |
|
|
|
1497 |
|
01:04:14,599 --> 01:04:19,400 |
|
modelm yeah thanks a lot everyone and |
|
|
|
1498 |
|
01:04:17,200 --> 01:04:19,400 |
|
yeah |
|
|
|
1499 |
|
01:04:20,839 --> 01:04:29,400 |
|
than and so we will um we'll have people |
|
|
|
1500 |
|
01:04:25,640 --> 01:04:30,799 |
|
uh kind of have office hours uh every ta |
|
|
|
1501 |
|
01:04:29,400 --> 01:04:32,880 |
|
has office hours at a regular time |
|
|
|
1502 |
|
01:04:30,799 --> 01:04:34,480 |
|
during the week uh please feel free to |
|
|
|
1503 |
|
01:04:32,880 --> 01:04:38,400 |
|
come to their office hours or my office |
|
|
|
1504 |
|
01:04:34,480 --> 01:04:41,960 |
|
hours um I think they are visha are they |
|
|
|
1505 |
|
01:04:38,400 --> 01:04:43,880 |
|
posted on the site or okay yeah they |
|
|
|
1506 |
|
01:04:41,960 --> 01:04:47,240 |
|
they either are or will be posted on the |
|
|
|
1507 |
|
01:04:43,880 --> 01:04:49,720 |
|
site very soon um and come by to talk |
|
|
|
1508 |
|
01:04:47,240 --> 01:04:51,480 |
|
about anything uh if there's nobody in |
|
|
|
1509 |
|
01:04:49,720 --> 01:04:53,079 |
|
my office hours I'm happy to talk about |
|
|
|
1510 |
|
01:04:51,480 --> 01:04:54,599 |
|
things that are unrelated but if there's |
|
|
|
1511 |
|
01:04:53,079 --> 01:04:58,039 |
|
lots of people waiting outside or I |
|
|
|
1512 |
|
01:04:54,599 --> 01:05:00,319 |
|
might limit it to uh like um just things |
|
|
|
1513 |
|
01:04:58,039 --> 01:05:02,480 |
|
about the class so cool and we have |
|
|
|
1514 |
|
01:05:00,319 --> 01:05:04,760 |
|
Patza we'll be checking that regularly |
|
|
|
1515 |
|
01:05:02,480 --> 01:05:06,839 |
|
uh striving to get you an answer in 24 |
|
|
|
1516 |
|
01:05:04,760 --> 01:05:12,240 |
|
hours on weekdays over weekends we might |
|
|
|
1517 |
|
01:05:06,839 --> 01:05:16,000 |
|
not so um yeah so that's all for today |
|
|
|
1518 |
|
01:05:12,240 --> 01:05:16,000 |
|
are there any questions |