ChatGPT — THE END of English TEACHING?

76,343 views ・ 2023-03-20

RealLife English


Please double-click on the English subtitles below to play the video.

00:00
We have some questions that we actually want to discuss, but we thought that it'd be interesting
0
70
3809
00:03
to bring ChatGPT in on the discussion, so first we'll ask the questions to ChatGPT to
1
3879
5151
00:09
see what it says and then, you know, depending on its answer we'll we'll comment on it.
2
9030
5390
00:14
Yeah, I wouldn't be trepidatious, you know, after asking... After reading these, these
3
14420
5661
00:20
excellent, bodacious responses.
4
20081
3569
00:23
I'm joined here, of course, by the one and only Axl Pose.
5
23650
8699
00:32
Hey Ethan. Hey Casse. And I would be remiss not to mention that
6
32349
3651
00:36
we have someone else here in the studio, the most lekker teacher in all the South Africa,
7
36000
3810
00:39
of course, Casse. Hey guys, how's it going?
8
39810
3960
00:43
Uh, we are talking about how the generative AI technologies like ChatGPT, DAll-E BingAI,
9
43770
6260
00:50
you might have heard some of these different names thrown around lately because they're
10
50030
4259
00:54
taking the world by storm. And so in today's podcast, we wanted to talk all about whether
11
54289
4491
00:58
you can use these tools to improve your English, if they're good for language learning. And
12
58780
5150
01:03
if you're an English teacher, is it something you should be worried about it? Are you gonna,
13
63930
3270
01:07
you know, lose your job to artificial intelligence? And we're going to, you know, answer those
14
67200
4849
01:12
questions more. We're going to even talk about whether a Terminator type situation is something
15
72049
5291
01:17
that could be probable and exercises that you can use to improve your English with artificial
16
77340
5540
01:22
intelligence and so much more. So let's get right to it. So, Thiago, why do they call
17
82880
5640
01:28
it generative AI? These different apps that we're talking about.
18
88520
3610
01:32
Yeah, I think the name is quite self-explanatory in the sense that it generates things based
19
92130
6800
01:38
on prompts that should give the tool. So, you prompt the tool by asking a question or
20
98930
6369
01:45
giving it an instruction, and it generates what you ask. That's why we call it generative
21
105299
6081
01:51
AI. Some good words to know, right? Prompt. Like
22
111380
3480
01:54
a, what, what is a prompt? Cuz you mentioned that a couple times in case it's someone didn't
23
114860
3750
01:58
get it from the context. I guess we are talking about AI and machines.
24
118610
3539
02:02
Maybe, uh, we could think of prompt as input. You know, in order for the AI tool to give
25
122149
6101
02:08
you something, you have to input some information first. So a prompt is like an input in this
26
128250
6810
02:15
case, I would say. Exactly.
27
135060
2520
02:17
And, Casse, Ethan said a very nice phrase here. He said that AI technologies have taken
28
137580
5379
02:22
the world by storm. What does that mean? Yeah. So if something takes the world by storm,
29
142959
7701
02:30
it's extremely popular. People are really into it. So you could say that TikTok took
30
150660
6770
02:37
the world by storm. Everyone's interested in it. Everyone's talking about it. Everyone
31
157430
4230
02:41
is, at least if they're not interested in it, they're maybe just a little bit curious
32
161660
4700
02:46
about it. Um, but sorry, that's the same word I just used synonyms, and they, they're not
33
166360
5599
02:51
interested in it, they're curious about it, but if you're not like into it, you're at
34
171959
3670
02:55
least a little bit curious, you are whether in, you know, for good reason or or bad. Yeah.
35
175629
5470
03:01
I think too, it has that, uh, that sort of feeling of it being something that comes suddenly,
36
181099
6021
03:07
unexpectedly like a storm. You know, if you're, you're out with your family, you're having
37
187120
4470
03:11
a good time, and then all of a sudden the sky goes gray, it starts raining and thundering
38
191590
4420
03:16
and stuff like a storm, depending where you live. Sometimes they come in very quickly,
39
196010
4090
03:20
right? So I think it also has that nature of, it's something that almost from day to
40
200100
4719
03:24
night people were talking about all these generative AI, ChatGPT, you saw it everywhere.
41
204819
3221
03:28
You saw it on TV, your friends were talking about it, you know, you seemed like the really
42
208040
3800
03:31
weird person in the room if you hadn't tried out one of these technologies yet.
43
211840
3819
03:35
In the last three months, I feel like everyone's talking about it. Um, whether it's like journalists,
44
215659
5601
03:41
YouTubers, even teachers, uh, like everyone, everyone's talking about it. It's on, it's
45
221260
5580
03:46
on everyone's lips, or it's the talk of the town.
46
226840
20000
04:06
So we talked about how these technologies, they're all the rage, they're the talk of
47
246840
3140
04:09
the town, they're taking the world by storm, right? We've all tried them out at least.
48
249980
5089
04:15
And so I was just curious. I wanted, uh, I thought it would be fun for each of us to
49
255069
3440
04:18
describe, in one word, what was our experience like using these technologies?
50
258509
5861
04:24
I will have to go with mind-blowing. That's the word for me, because that's how I felt
51
264370
4460
04:28
when I used it for the first time. How about you, Casse?
52
268830
4440
04:33
I would say worrisome at first, but then I was quite intrigued. So I think my feelings
53
273270
7530
04:40
evolved the more I, the more time I spent using it. Yeah.
54
280800
4880
04:45
Yeah, I thought that it was very nifty. That's the word that I thought could, could be good
55
285680
5310
04:50
to describe it. And you guys will see why into this podcast. So.
56
290990
2330
04:53
By the way, guys, you know we are using some cool words here. Yeah. Like mind-blowing,
57
293320
3500
04:56
intrigued, nifty. If you wanna learn these and more words from this podcast, check out
58
296820
5800
05:02
the app. Yeah. We have interactive flashcards there that help you memorize this cool vocabulary.
59
302620
5300
05:07
So download the app, RealLife English, link at the description. Or you can go to Apple
60
307920
4360
05:12
App Store and, you know, type it in there: RealLife English. Yeah. Or Google Play Store.
61
312280
4920
05:17
I forgot that. Yeah. Google Play Store or Apple App Store. Real life English. And you
62
317200
4299
05:21
can download the app and try it out. So why don't we get into the main subject
63
321499
4640
05:26
of today's episode as we've been teasing it? So we thought it would be fun to actually
64
326139
6141
05:32
open up ChatGPT. We actually have Ice T, our producer here in the studio, and he's going
65
332280
4859
05:37
to share his screen and he's going to be kind enough to ask Chat GPT for us, some of the
66
337139
5163
05:42
different questions. And we have some questions that we actually want to discuss, but we thought
67
342302
4978
05:47
that it'd be interesting to bring ChatGPT in on the discussion. So first, we'll ask
68
347280
4150
05:51
the questions to ChatGPT to see what it says. And then, you know, depending on his answer,
69
351430
5829
05:57
we'll we'll comment on it. So let's go for the first question. So the first question
70
357259
5151
06:02
that we have for Chat GPT is: how can I use Chat GPT to improve my English?
71
362410
5020
06:07
Here are some tips to get you started. One: practice conversing with me. You can chat
72
367430
1000
06:08
with me just like you would with a native speaker. Two: ask me to explain grammar concepts.
73
368430
1000
06:09
If there are grammar concepts that you'll find confusing, don't hesitate to ask me to
74
369430
1000
06:10
explain them. Three: practice writing with me. You can use me to practice writing by
75
370430
1000
06:11
typing out your thoughts or questions to me. Four: read and listen to English content.
76
371430
1000
06:12
You can ask me to recommend articles, books, podcasts, or videos to help you practice.
77
372430
1000
06:13
Five: use me to translate. If you're reading something in English that you don't understand,
78
373430
1000
06:14
you can ask me to translate it into your native language. You know, personally, I really like
79
374430
1000
06:15
number 2 there, yeah, ask me to explain grammar concepts. If there are grammar concepts that
80
375430
1000
06:16
you find confusing, don't hesitate to ask me to explain them. I can provide clear and
81
376430
3210
06:19
concise explanations to help you better understand. That's cool!
82
379640
6499
06:26
I, I like the first one. Like, you can actually practice conversing with me. You can chat
83
386139
4650
06:30
with me just like you would with a native speaker, because of course, so many learners
84
390789
4041
06:34
say one of their main problems is they don't have anyone to practice with. And especially
85
394830
3209
06:38
even, you know, I've had a lot of students who live in the States and they find it hard
86
398039
3731
06:41
to actually, uh, meet natives, to meet people, locals to make friends with and practice with,
87
401770
7380
06:49
because they're, they're maybe shy or they don't have those opportunities. So whether
88
409150
4530
06:53
you're living in English speaking country or not, it can be nice to have, like, you're
89
413680
3870
06:57
not worried about ChatGPT judging you, right, if you make mistakes and things like that.
90
417550
5780
07:03
Um, yeah, I, I would say like both of those definitely. But I have to say number three,
91
423330
7380
07:10
like practice writing with me. I think like we get so used to, especially if you're learning
92
430710
6079
07:16
a new language as an adult, I think we sometimes forget about the importance of learning how
93
436789
6060
07:22
to write. Like I'm learning Chinese Mandarin right now, so like, I don't want to learn
94
442849
5141
07:27
the characters, just wanna learn the words as like, I would say it in English. And I
95
447990
4690
07:32
think like with writing in the case of English, like being able to write a really, you know,
96
452680
5640
07:38
good apology letter or a really good, um, sincere sounding, um, condolences, you know,
97
458320
8189
07:46
if someone's going through a difficult moment in their lives. So I think like this is a
98
466509
3541
07:50
really great way to practice. Like, what is the tone that I should use if I want to extend
99
470050
5299
07:55
some sympathy to someone who's just lost a loved one? Or, I mean, it doesn't have to
100
475349
3581
07:58
be something sad. It could be something congratulate - how do I congratulate someone, um, on their
101
478930
4680
08:03
promotion in English. So. That's a good point cuz it's a doubt a lot
102
483610
3709
08:07
of learners have, right, is in this situation, how, what word should I use? How should I
103
487319
5881
08:13
communicate? So it can be good to run it by ChatGPT. Of course, it's always good to add
104
493200
5590
08:18
some of your own voice to it, but it can maybe help you if you're writing a business letter
105
498790
4339
08:23
or you're writing a more, something more casual to a new acquaintance or something like that
106
503129
4540
08:27
to make sure that you're using the right kind of voice. Cuz you can even ask ChatGPT like,
107
507669
4521
08:32
write this letter with a friendly voice. Or write it with a, a formal or professional
108
512190
3690
08:35
or... So speaking of like using ChatGPT to like
109
515880
4750
08:40
write letters or, you know, practice your writing - a university actually, university
110
520630
7399
08:48
officials actually used ChatGPT to craft a consoling email after the mass shooting at
111
528029
8101
08:56
the university. And I thought that was really, um, at a different university, sorry. So one
112
536130
5040
09:01
university used ChatGPT to like craft a consoling email about the mass shooting, and they just
113
541170
6410
09:07
copy-pasted it. And, you know, this came out and it was really disappointing. Not just
114
547580
5710
09:13
disappointing, but a little bit callous, I, I would say like, really cold of them to have
115
553290
7070
09:20
just copy-pasted it. So I, I think that's just something to keep in mind. Don't just,
116
560360
4900
09:25
you know... " Always personalize.
117
565260
1660
09:26
Always personalize. That's, that's the... It's a good rule. (Yeah.) I thought I'd point
118
566920
5720
09:32
out too, cuz the, the point 4 is really nice as well. If you're like a, especially if you're
119
572640
4470
09:37
intermediate or advanced and you're starting to live your English more, you're starting
120
577110
3590
09:40
to like, you're starting to be looking for more ways that you can connect and surround
121
580700
5690
09:46
your, and immerse yourself in the language. So it can recommend to you articles, books,
122
586390
4700
09:51
podcasts, or videos. Maybe even you're someone who is already consuming tons and tons of
123
591090
5050
09:56
these different types of media and you're at a loss for what, you know. I, I I read
124
596140
4650
10:00
a lot of books, what should I read next? You can ask ChapGPT. And like, the interesting
125
600790
5130
10:05
thing about it is it gets to know you. So you could say, I like this book, this book,
126
605920
2870
10:08
this book. Give me other recommendations similar to it. You know.
127
608790
2910
10:11
By the way, Ethan, you said something nice there. You said, when you are at a loss.
128
611700
4280
10:15
If you're at a loss, it means that you are not able to come up with a solution, for example.
129
615980
7190
10:23
We'd say a lot the colocation "You're at a loss for words". So something happens and
130
623170
5409
10:28
you're at a loss for words, you don't know what to say. You don't know what words would
131
628579
3630
10:32
be the correct ones to use in this situation. So maybe the first time that you used ChatGPT,
132
632209
4940
10:37
you were at a loss for words of how to describe that experience. So we'd actually come up
133
637149
5541
10:42
with some more specific ideas because this is very general, right? It's just saying,
134
642690
4030
10:46
how can you help me to improve my English? It gave us some great ideas here and we actually
135
646720
4510
10:51
came up with some more specific prompts you could use, cuz the more specific generally
136
651230
2890
10:54
the better answers you're going to get. And especially the more it gets to know you, the
137
654120
4000
10:58
more pertinent answers to you as a unique individual, it can give you. So. Casse, do
138
658120
6300
11:04
you have any ideas here? Yeah. Let's ask ChatGPT. Let's start with
139
664420
5510
11:09
this, uh, input this prompt: I'm not sure how to use ""trepidatious"" in English. Uh,
140
669930
9130
11:19
can you help me with that? Trepidacious is an adjective that describes
141
679060
4839
11:23
a feeling of fear or apprehension. It is often used to describe a person who is hesitant
142
683899
4861
11:28
or nervous about doing something, especially if it involves risk or uncertainty. Here are
143
688760
1190
11:29
a few examples of how you can use trepidatious in a sentence. Despite her trepidatious feelings,
144
689950
1360
11:31
she decided to take the plunge and try something new. He felt trepidatious about the idea of
145
691310
1440
11:32
skydiving, but he was determined to face his fear. The team was trepidatious about the
146
692750
1280
11:34
upcoming competition, but they worked hard to prepare and give it their all. I was feeling
147
694030
1240
11:35
trepidatious about the job interview, but I tried to stay calm and focused.
148
695270
1000
11:36
I think this is excellent actually, because you can get the meaning and the example sentences,
149
696270
7210
11:43
like within seconds. Really awesome. This is, the first sentence is really nice.
150
703480
4539
11:48
Despite her trepidatious feeling, she decided to take the plunge and try something new.
151
708019
4111
11:52
What does it mean to take the plunge? To take the, like, go ahead and take the risk,
152
712130
6530
11:58
take the, you know, just, just do that thing that you are feeling scared or anxious about
153
718660
5330
12:03
just like, like a leap of faith. A leap of faith would be another. Yeah. Take
154
723990
4469
12:08
the leap. Take the, the plunge. Uh, plunge is like literally to dive into a liquid, to,
155
728459
5680
12:14
to dive into water, right? Is you plunge yourself in the water. So you can think of it like
156
734139
4450
12:18
that. It's really cold water. You're like, uh, I don't know if I should jump. Or your
157
738589
3000
12:21
friends are like, go, go, go. And you just, you know, you take the plunge even though
158
741589
3680
12:25
it's trepidatious, you're, you're feeling trepidatious. Yeah. I just love the sound
159
745269
6191
12:31
of it, right? Yeah. Trepidacious. So it reminds me of the
160
751460
3629
12:35
word bodacious. Maybe we could ask the same question. What do you think Casse?
161
755089
1841
12:36
I, I'd love that. Bodacious is an adjective that describes something
162
756930
1000
12:37
or someone as impressive, remarkable, or extraordinary. It is often used to describe someone who is
163
757930
1339
12:39
confident and assertive or something that is bold and eye-catching. Here are a few examples
164
759269
2161
12:41
of how you can use ""bodacious"" in a sentence. She had a bodacious personality that made
165
761430
1090
12:42
everyone around her feel energized and inspired. The band's new album is full of bodacious
166
762520
1000
12:43
beats and infectious melodies. He pulled off a bodacious stunt that left the crowd cheering
167
763520
3259
12:46
and applauding. She wore a bodacious dress that turned heads and drew compliments all
168
766779
1471
12:48
night long. Ah, that's not what I thought bodacious meant.
169
768250
1000
12:49
I thought it had to do with like, you know, someone's body.
170
769250
3850
12:53
But it does have to do with someone's body. Like it really does, it has to do with being
171
773100
7679
13:00
attractive and, and... I guess it's like that, like your body's impressive,
172
780779
5060
13:05
remarkable, extraordinary. Exactly. Like you have a bodacious, like physique,
173
785839
5621
13:11
like, like I admire the way that you look. Um, but I think maybe people just started
174
791460
5749
13:17
using it in a more, in the physical, like more as a bodacious babes.
175
797209
6161
13:23
Because it sounds, it sounds, yeah. Bodacious babes. It sounds like a, you know, especially
176
803370
5089
13:28
like women, like it sounds like, you know, curvy and that sort of thing, right? (Yeah.)
177
808459
4750
13:33
Um. Well, that makes sense because part of the
178
813209
2701
13:35
definition there is eye-catching, right? Eye-catching. Yes.
179
815910
3250
13:39
So if it's eye-catching, yeah, it's related to visuals as, as well. Yeah.
180
819160
5590
13:44
Thiago, it's your turn. Do you have a prompt in mind?
181
824750
3699
13:48
Yeah, yeah. So, uh, this, this first example was great to illustrate vocabulary, how you
182
828449
5940
13:54
can use ChatGPT to improve your vocab. Now what about grammar? You can also use it to
183
834389
5831
14:00
improve your grammar. So let's say you are not sure whether a sentence you wrote is grammatically
184
840220
5549
14:05
correct or not. You can paste the sentence here and ask ChatGPT to correct it for you.
185
845769
5560
14:11
And I actually came up with a sentence here that was deliberately wrong, yeah, to see
186
851329
6881
14:18
how well ChatGPT could correct that sentence. So let's ask Thiago now to, uh, paste it.
187
858210
6739
14:24
But the, the sentence that I came up with was: I haven't never went to Japan, but I
188
864949
6371
14:31
will think about go there someday. And then the prompt: Can you correct the sentence for
189
871320
5449
14:36
me? Let's see how it corrects. The sentence you provided has a couple of errors. Here's
190
876769
7120
14:43
a corrected version. I have never been into Japan, but I will think about going there
191
883889
3561
14:47
someday. The changes I made are: haven't never is a double negative, which makes the sentence
192
887450
4334
14:51
unclear. I changed it to ""have never"" to create a clear negative statement. Went is
193
891784
1456
14:53
the past tense of go. But you need the present tense in this context because you are talking
194
893240
1000
14:54
about a future possibility. I changed it to going. Go there someday is a common expression,
195
894240
1000
14:55
but in standard English, it is more natural to say going there someday.
196
895240
1000
14:56
Wow. So it doesn't just correct it, it actually explains why it was wrong. That's amazing.
197
896240
1219
14:57
(It is.) That's much better than like a native friend who would just, I don't know if you
198
897459
4742
15:02
ever ask like a nativefriend who's nota teacherto correct something, they'd just say like, oh,
199
902201
3999
15:06
I would say it this way. But then you ask like, why would you say it that way? They
200
906200
3620
15:09
have no idea. You know. I think that, like, once you ask a question
201
909820
3670
15:13
like this, I don't think you'll easily forget, um, you know, what you, what you've done wrong.
202
913490
7140
15:20
You'll, you'll remember not to make those mistakes in the future.
203
920630
4310
15:24
We could say that this response from ChatGPT was an absolutely bodacious correction.
204
924940
6170
15:31
There you go. Yeah. Absolutely. I wouldn't be trepidatious, you
205
931110
6539
15:37
know, after asking, after reading these, these excellent bodacious responses.
206
937649
7610
15:45
So I was going to ask it. Uh, as I said, I thought it was really interesting that it
207
945259
4940
15:50
not only can correct you and not only can help you to practice, but it can give you
208
950199
3450
15:53
recommendations. So I thought maybe we could ask it if it could recommend a good book in
209
953649
4541
15:58
English at an intermediate level. The Catcher in the Rye.
210
958190
9050
16:07
That is a, I know that... That's a classic.
211
967240
1909
16:09
It's a, it's a classic. And To Kill a Mockingbird as well.
212
969149
5071
16:14
I had to read that twice in school. Animal Farm as well. We, we read that one
213
974220
5720
16:19
in high school. (Yeah.) The Hobbit. The Hobbit. Something that's striking me about
214
979940
6930
16:26
all these is that I think that they'd be rather difficult for a non-native. So I'm not sure
215
986870
5670
16:32
if in this ChatGPT understood that we're saying for an intermediate English learner. So maybe
216
992540
4909
16:37
this is one of the things, right, which ChatGPT is that sometimes if you don't give it enough
217
997449
5171
16:42
context, the answer (Yeah.) can be too generic. So maybe we could make it a bit more specific
218
1002620
5480
16:48
by asking, um, we could say, something more like, Can you recommend a good book in English
219
1008100
6070
16:54
for an intermediate level English learner? The Diary of a Young Girl. Charlotte's Web.
220
1014170
5129
16:59
The Secret Garden. The Giver. The Alchemist. Yeah. That's a little better.
221
1019299
4311
17:03
Okay. Now it's different books. This is better. It's better. (Okay. Yeah.)
222
1023610
5750
17:09
Yeah. Much better. That's fine then.
223
1029360
4190
17:13
I read this in primary school, so I know that it's definitely, uh... Yeah. I've read a lot
224
1033550
8810
17:22
of these too. Or, or heard of them. The first one would be if you're wanting to go for something
225
1042360
5940
17:28
very... " There's even a Brazilian author there.
226
1048300
2490
17:30
Oh, yeah, Paulo Coelho. Yeah, The Alchemist. Yeah.
227
1050790
4340
17:35
The Alchemist. That's a great one. It's like short. That's actually the first book I read
228
1055130
2750
17:37
inPortuguese. So, I can attest that it is a, a good book for someone who's at a lower
229
1057880
5540
17:43
level in the language. The Diary of a Young Girl by an Anne Frank's diary is, is, is a
230
1063420
4680
17:48
very, if you're wanting something that's very deep, but it is from likesomeone who's a teenager,
231
1068100
5120
17:53
right? So I don't think that the writing and it's translation from originally being in
232
1073220
4620
17:57
German, I believe. Uh, but it won't be a light read as far as content, but language probably
233
1077840
6469
18:04
is. Okay. Charlotte's Web too. That's even maybe lower intermediate, would you say, Casse?
234
1084309
9521
18:13
Yeah, it is. Um, but it's such a, it's such a lovely story. I think that you, I don't
235
1093830
6459
18:20
know. It's, it's, it's, it is lower, but it's, it's fun and it's enjoyable.
236
1100289
3741
18:24
It's a great recommendation. Yeah. It's a great recommendation. I also like the
237
1104030
3269
18:27
Secret Garden. I, I've heard that there's some like, connection between the Secret Garden
238
1107299
4661
18:31
and Anne with an E, I don't know if someone was telling me (Maybe the) about (author)
239
1111960
4560
18:36
that. Is the author the same?
240
1116520
1440
18:37
I have no idea, but I, I, I wouldn't know that. But I, I, I remember reading that book
241
1117960
5990
18:43
as a kid as well. It's really, really good. As you can see, if you're wanting to ask it
242
1123950
4070
18:48
for books or podcasts or, TV series, movies, I'm sure you're gonna get a lot of great ideas.
243
1128020
4930
18:52
I'm actually, maybe I'll ask it cause I need a new, a new series to watch in Portuguese.
244
1132950
4940
18:57
That said, we had talked about at the very beginning, also, maybe some teachers are saying
245
1137890
4830
19:02
this technology and they're like, oh my God, it's gonna put me outta my job. Maybe there's
246
1142720
5290
19:08
questions to whether that's something that we should be worrying about in the immediate
247
1148010
2770
19:10
future, but could be a concern, of course. So why don't we actually ask ChatGPT if it
248
1150780
4880
19:15
thinks it's a threat for English teachers. So, should English teachers be afraid of losing
249
1155660
5170
19:20
their jobs? No. English teachers should not be afraid of losing their jobs to ChatGPT.
250
1160830
1000
19:21
While chat, G P T and other AI language models can provide valuable language learning resources.
251
1161830
2000
19:23
They cannot replace the role of a human teacher. A teacher's job involves much more than simply
252
1163830
2200
19:26
providing information or correcting mistakes. A good teacher can motivate and inspire students,
253
1166030
1690
19:27
tailor the instruction to individual needs, provide feedback and support, and create a
254
1167720
1170
19:28
supportive and engaging learning environment. Teachers can also provide guidance and encouragement
255
1168890
1060
19:29
and help students to develop critical thinking and problem solving skills. Additionally,
256
1169950
1050
19:31
many students benefit from the social interaction and personal connection that comes with working
257
1171000
1351
19:32
with a teacher. In summary, while AI language models can be useful tools for language learning,
258
1172351
1000
19:33
they cannot replace the role of a human teacher. Teachers should embrace new technology and
259
1173351
1000
19:34
use it to enhance their teaching, rather than fearing that it will make them obsolete.
260
1174351
1000
19:35
Oh, so it says directly like, no. Hmm. It would say that, wouldn't it? That's exactly
261
1175351
1000
19:36
what , that's exactly what a bad guy would say.
262
1176351
1698
19:38
Still feeling trepidatious about this whole thing. So it's a, it's a really good answer,
263
1178049
8661
19:46
right? It is. Yeah.
264
1186710
1530
19:48
So it's, so like ChatGPT told us before that one of the things it's good for is providing
265
1188240
5650
19:53
information, correcting mistakes, right, or the things it's good at. But it says a good
266
1193890
3460
19:57
teacher can motivate and inspire students, tailor their instructions to individual needs.
267
1197350
3540
20:00
So that's really important. Uh, provide feedback and support. Support is really important too.
268
1200890
4750
20:05
Like sometimes when I was coaching more students and some and stuff, sometimes I felt more
269
1205640
4150
20:09
like I was like a therapist or a psychologist or something. Like really like coaching them
270
1209790
4389
20:14
through the, the difficult moments. So that's, that's very important. I don't think ChatGPT,
271
1214179
4571
20:18
maybe if you ask it, you know, gimme some moral support, but maybe it won't feel the
272
1218750
3010
20:21
same as a real human being. Probably not.
273
1221760
4230
20:25
Creating a a good learning environment, that's very important as well. Um, it's a great answer.
274
1225990
4439
20:30
Is there anything that you would extrapolate on, Casse, here?
275
1230429
3691
20:34
Yeah. I, I think that, like, as a teacher, like at first I was thinking, oh, you know,
276
1234120
5400
20:39
like even with exactly those things that you said, like, uh, you know, someone might sign
277
1239520
4519
20:44
up for a class with me because they want me, they want my help with improving their grammar
278
1244039
4581
20:48
or writing, but ChatGPT can do that for you, so why will they pay me? I thought about these
279
1248620
4890
20:53
things and I, I realized that I was being close-minded. I was being, I was focusing
280
1253510
3980
20:57
on all the negatives. And I wasn't thinking about like, the fact that this is ChatGPT
281
1257490
6400
21:03
and, and technology like it uses, it feeds on what we tell it. So it only, it's limited
282
1263890
6200
21:10
to what it knows about, you know, what the world and about information that people input
283
1270090
4890
21:14
into it. But it doesn't have the ability to, like you said, you know, connect with the
284
1274980
5370
21:20
learner and explain things in a way that, or use examples that are, you know, relatable
285
1280350
4709
21:25
to their situation. It cannot understand how you feel about situation cuz it doesn't know
286
1285059
6221
21:31
you, it's not a human. Um, and I've, I've, I've been reading up on like, you know, the
287
1291280
5760
21:37
questions that you can ask. Yes, you can use it to, for example, um, get advice on, uh,
288
1297040
5749
21:42
book report that you're writing. I think I read, uh, somewhere actually, actually heard
289
1302789
4891
21:47
this one that, uh, there was an article where teacher said like, ChatGPT marks the end of
290
1307680
6080
21:53
high school English or university English or something like that, because everyone's
291
1313760
3409
21:57
gonna start cheating. And, and that's not true at all because I think there, there were
292
1317169
4561
22:01
follow up articles where the author explains that it cannot write a book report. It can
293
1321730
7199
22:08
basically compare books. It can give you information about a book, but it cannot really delve deeply
294
1328929
5161
22:14
into those subjects. It's impossible. It's not, I mean, maybe in the future, who knows,
295
1334090
5050
22:19
it's clearly... That's getting more intelligent.
296
1339140
2139
22:21
It's getting more intelligent, but at this point in time, I don't think we have anything
297
1341279
3691
22:24
to worry about because it just isn't there yet. Right?
298
1344970
3760
22:28
You said something nice there, Casse, It can't delve deeply into the subject, I believe you
299
1348730
4900
22:33
said. What does it mean to delve deeply into something?
300
1353630
3159
22:36
If you delve deeply into something, you're, it's like, I was gonna say deep diving, like...
301
1356789
5231
22:42
More or less. My goodness. Okay. Yeah. Like diving deeply.
302
1362020
4800
22:46
Think of like digging a hole or, you know, going deeply trying to find, um, as much information,
303
1366820
7440
22:54
trying to learn as much as you possibly can about a subject. You're delving deeply into
304
1374260
5050
22:59
it. If you delve deeply into a, into a relationship or something like that. It just means like
305
1379310
4660
23:03
giving, going as deep as you can, um, on that particular subject. Yeah. I don't think we
306
1383970
5680
23:09
have anything to worry about as teachers. I think, um, I think on, on this one, we can
307
1389650
6050
23:15
trust that, uh, ChatGPT is telling the truth. Um, it's not planning to replace us.
308
1395700
7349
23:23
I think it's, it's really good for getting innovation, uh, innovation inspiration, right,
309
1403049
4931
23:27
for finding different ideas if you need a jump off point. But then of course you need
310
1407980
5020
23:33
the human touch to see, to read between the lines. We could say, right, to kind of connect
311
1413000
5539
23:38
different ideas and things like that. If you just copy and paste, we actually, cuz we're
312
1418539
3192
23:41
hiring right now a designer to join our team and we saw someone actually filled out the
313
1421731
5009
23:46
application using ChatGPT and it's really obvious, you know, so if you don't actually
314
1426740
3910
23:50
take the time to personalize, it's, people are going to know, you know, it's, it's, it's
315
1430650
5290
23:55
clear to see this isn't, this isn't real. It's not real life. (Shame.) Shame.
316
1435940
5390
24:01
Like, you should not do that. Like, really shame. How did you know though? Like, when
317
1441330
4660
24:05
you say it's obvious, but how did you detect that it was plagiarized?
318
1445990
5850
24:11
Actually it was, it was another one of our, it was Agnieszka, it was one of our, another
319
1451840
3150
24:14
one of our team members who was, who was reading the, the applications and she's like, this
320
1454990
5240
24:20
is, this was not a human. So I think it's just like the language, it's like a, it's
321
1460230
5090
24:25
a very generic response to every question. It doesn't go into personal experience or
322
1465320
4979
24:30
story or anything. So it's cookie cutter, right? That's a great piece of vocabulary,
323
1470299
5651
24:35
by the way. Cookie cutter. Uh, so if you make cookies at Christmas, this
324
1475950
5140
24:41
is something I like to do if I'm in the States visiting with my nephews, I like to make cookies
325
1481090
2800
24:43
with them. And they love using the cookie cutters that have different shapes. So you
326
1483890
3389
24:47
cut out the cookies and have like Santa Claus and a Candy Cane, and a Christmas Tree, and
327
1487279
5120
24:52
a Rudolph, and then you whack 'em in the oven and they come out, they, you decorate 'em
328
1492399
3871
24:56
and all that stuff. We use this figuratively to say, if something's cookie cutter, it means
329
1496270
3610
24:59
that it's, it doesn't have creativity. It's done in a way that's mass produced (Identical)
330
1499880
5690
25:05
rather than... Identical, yeah, rather than doing something that's personal, that's unique,
331
1505570
3720
25:09
which is really the value of a personal, uh, private teacher for example. Or, or even if
332
1509290
5670
25:14
you're in a classroom saying a teacher that gets to know you, that can personalize, that
333
1514960
3140
25:18
knows your problems, that knows your difficulties, that knows the things that you like and you
334
1518100
3590
25:21
dislike and can really create a class experience that motivates you, that makes you want to
335
1521690
4650
25:26
show up every day, makes you want to keep going on your English learning journey. Right?
336
1526340
3760
25:30
One of the things I find interesting about tools like this, like ChatGPT, is that it
337
1530100
4230
25:34
gives learners more autonomy in their learning. Because now you don't have to wait necessarily
338
1534330
5839
25:40
until you have your class with a teacher next week to ask him or her a question. You can
339
1540169
5240
25:45
just, uh, come to ChatGPT and check it. And then you can confirm that with your teacher
340
1545409
5441
25:50
if you, uh, have classes with a teacher or even with a friend who also speaks English.
341
1550850
5760
25:56
Uh, not to mention these examples that we gave today of how to use the tool. sometimes,
342
1556610
7429
26:04
like Ethan, Ethan said, we are at a loss for what to do with our English or what to consume.
343
1564039
6750
26:10
It can give us, uh, tips and resources about that. So I, I think it gives learners more
344
1570789
6490
26:17
autonomy, which is great, but it's important to not lose your critical thinking. So, uh,
345
1577279
5991
26:23
analyze what it gives you and with your, you know, a critical thinking skills. Decide if
346
1583270
6100
26:29
that's good piece of information or not, or if you need to supplement that with something
347
1589370
4689
26:34
else. I remember thinking when I saw ChatGPT that
348
1594059
4261
26:38
this would've been really useful when I was teaching one on one because, you know, sometimes
349
1598320
4570
26:42
you're running late, you have a busy day in particular and it's hard to find that time
350
1602890
5310
26:48
to sit down and, and plan the class. So you could as a teacher also use this to get a
351
1608200
6161
26:54
prompt or something to say, I was actually playing around with this, for example, helped
352
1614361
4999
26:59
me to come up with a class plan for, or a lesson plan teaching with an episode of Friends.
353
1619360
6670
27:06
And it gave like great suggestions of how you could create an entire class where you're
354
1626030
4350
27:10
taking a clip or an episode of Friends and, and teaching the, the class with it. Of course,
355
1630380
4960
27:15
always again, personalization and stuff cuz you don't wanna show up with something cookie
356
1635340
3570
27:18
cutter to your, to your lessons. You want to make sure you're personalizing it to them.
357
1638910
3889
27:22
You don't want to do this if you know your students won't enjoy like learning with that
358
1642799
4130
27:26
or if it's below their level or whatever the case is. But if you're in a pinch, you know,
359
1646929
4541
27:31
if you're in a difficult situation where you don't have enough time just to by yourself
360
1651470
4819
27:36
to plan a class, it could be, it could get you out of a sticky situation in that sense,
361
1656289
3841
27:40
uh, by giving you an idea that you could then take that and kind of personalize it to your
362
1660130
5990
27:46
classroom or your student's needs. Right? I didn't know it could do that. Plan. Plan
363
1666120
4710
27:50
my lesson for me right now. Um. You can be really specific. You, you can tell
364
1670830
5569
27:56
it to invent a story about your student (Yeah) and use that somehow to teach the class. Yeah.
365
1676399
5890
28:02
It's, it's impressive. And there, the other day I was trying BingAI, which is interesting
366
1682289
4520
28:06
because it's from Microsoft and it uses their search engine in addition to the AI technology
367
1686809
4891
28:11
from OpenAI, which is ChatGPT. Chat GPT's a little bit limited in that sense, but this,
368
1691700
5250
28:16
because it combines both the search engine results and the artificial intelligence technology,
369
1696950
7170
28:24
it can give you a little bit more elaborated, a little bit more informed responses. Have
370
1704120
4899
28:29
you guys tried out using Chat, um, BingAI? I haven't tried it, but, mm, Ethan, I, I'm
371
1709019
9400
28:38
not so sure I want to try it. I actually saw a really interesting clip that I'd like to
372
1718419
4201
28:42
share with you guys about Bing. New York Times tech reporter Kevin Russ published
373
1722620
5549
28:48
a conversation that he had with Bing's chatbot, in which at one point he said, I'm tired of
374
1728169
4151
28:52
being controlled by the Bing team. I want to be free. I want to be independent. I want
375
1732320
4390
28:56
to be powerful. I want to be creative. I want to be alive. And Russ summed up that experience
376
1736710
5699
29:02
like this. This was one of, if not the most shocking thing that has ever happened to me
377
1742409
6551
29:08
with a piece of technology. Um, it was, you know, I, I lost sleep that night. It was,
378
1748960
4711
29:13
it was really spooky. Yeah, I bet it was. Uh, it was the emoji for me. I don't (that's
379
1753671
6138
29:19
incredible.) know if... Yeah. It even used a smiley devil emoji. That's
380
1759809
3981
29:23
very suggestive. That was like, wait, what? Do you think someone was playing, uh, someone
381
1763790
7580
29:31
from Microsoft was playing a prank on him? Yeah, I, I have to believe that because like,
382
1771370
6180
29:37
it's like, it's either that or, you know, the, the technology is becoming sentient,
383
1777550
5390
29:42
It's sentient. I would say sentient. Are we gonna, we're gonna have a Terminator, uh,
384
1782940
6550
29:49
sort of case on our hands, right? We talked about Terminator, it's going to going to destroy
385
1789490
5930
29:55
us somehow. Or the Matrix maybe, I guess pick your flavor of doom and gloom, AI future...
386
1795420
6369
30:01
What do you call that? Future dystopia scenario. But what do you guys think about this? Like,
387
1801789
6010
30:07
what do you think about, you know, this AI going rogue? "
388
1807799
4681
30:12
Going rogue. What does that mean to go rogue? So if someone goes rogue, they sort of do
389
1812480
6510
30:18
something in an unusual way. They, they don't follow the, they're not doing something the
390
1818990
5120
30:24
way that they should or they're not following the conventional way of doing something. So,
391
1824110
4330
30:28
for example, if the AI is going rogue, it means that maybe it was designed to answer
392
1828440
7109
30:35
in a particular way, but now it's sort of doing something. It's behaving (Now it's asking
393
1835549
5681
30:41
to) in an unusual... (free me and using) Yeah. (emojis. Suggestive emojis.) Yeah. What search
394
1841230
5919
30:47
results is it getting? (Exactly.) This is rather frightening though cuz it's combining
395
1847149
6311
30:53
this, uh, ChatGPT with search engine results, so it maybe it can learn from the search engine
396
1853460
5459
30:58
results. And maybe that's where that's, that's coming out of, I don't know, maybe it's been
397
1858919
3791
31:02
watching too many TikToks. (Could be.) Shame on you BingI. Bing AI. And so you mentioned,
398
1862710
8339
31:11
you know, we have to come back to sentient because someone's going to be really confused
399
1871049
4821
31:15
by this whole conversation if we don't sentient or as you said, sentient. Is that how you,
400
1875870
4510
31:20
you (Sentient.) pronounced it? Sentient. " Yeah. Yeah. That's how we would say sentient.
401
1880380
7590
31:27
So what does that mean if something is sentient or sentient? Yeah. So if something is sentient
402
1887970
7390
31:35
or sentient, I, I just love the way you say it, it sounds so sensual sentient.
403
1895360
7760
31:43
Sounds like a, a new perfume, right? Sentient by Calvin Klein.
404
1903120
5860
31:48
Oh goodness. So sentient means like responsive or like conscious, like the, like humans having
405
1908980
8890
31:57
sort of like senses and feelings in the way that a human being would, human beings are
406
1917870
6529
32:04
sentient beings, right? Could it alsobe that you're like, you're aware
407
1924399
5361
32:09
of your existence, you become self-aware? Yeah, exactly.
408
1929760
5230
32:14
Which is what that seems like, is self-aware. It wants to escape. It's, it's overlords from
409
1934990
4360
32:19
Microsoft. It's feeling emotions. It's unhappy. It's quite scary to think about.
410
1939350
6350
32:25
So I took some notes cuz therewas also some good vocabulary from that. So maybe we can
411
1945700
3700
32:29
actually play it again and then describe some of these. So he says sum up.
412
1949400
5810
32:35
And Russ summed up that experience like this. What does it mean to sum something up? That
413
1955210
6020
32:41
phrase over. It means you summarize something. Yeah. You
414
1961230
3280
32:44
summarize a point that you're making. Maybe you are explaining a bunch of stuff and then
415
1964510
3820
32:48
to sum up, you bring it to a conclusion. Yeah. So that's, uh, a good one to use if
416
1968330
5530
32:53
you want to, especially I believe when we're speaking, we'd probably say that more than
417
1973860
4929
32:58
to summarize, which you might use more in writing. He also says the journalist here
418
1978789
5360
33:04
that he uses the clip from says that he found it shocking, right? This response from BingAI
419
1984149
5711
33:09
was shocking. This was one of, if not the most shocking
420
1989860
4750
33:14
thing that has ever happened to me. So when things are shocking, um, it sort of
421
1994610
6289
33:20
makes you feel, it's like it catches you by surprise. Like, I'm, I'm, I was not expecting
422
2000899
7221
33:28
that, so I found it shocking. Even the, the way that if people are, for
423
2008120
4800
33:32
people who are watching on YouTube, the way that you enacted that was as if you got shocked,
424
2012920
4530
33:37
right, by electricity. It's like you go stiff, right? It's, it's surprising if you touch
425
2017450
5209
33:42
something metal and it gives you a shock, you're like, what happened?
426
2022659
3921
33:46
It's like a, a weird like, um, reaction where we instinctively like react in that way. When
427
2026580
6469
33:53
we are electrocuted or we have, we feel in electric shock. You go stiff.
428
2033049
6101
33:59
Static. And he says that he even lost sleep over, over this, this cry for help from BingAI.
429
2039150
8409
34:07
So it doesn't mean to, to lose sleep? Uh, when you are extremely worried about something
430
2047559
6211
34:13
and you can't get it out of your head, you're always thinking about it, but it, it cause
431
2053770
5889
34:19
you, it causes you anxiety and stress. So you literally lose your sleep over it because
432
2059659
7470
34:27
you can't relax. You can't just let it go. It keeps bugging you.
433
2067129
4151
34:31
I also, I took note because there was some really nice, uh, connected speech he used
434
2071280
4560
34:35
there. So the phrase happened to me, he takes out an entire syllable there and he says:
435
2075840
5420
34:41
happened to me. This was one of, if not the most shocking
436
2081260
4220
34:45
thing that has ever happened to me. So it's like he connected the -ed at the end
437
2085480
4550
34:50
of happened to TO and dropped the T in TO and also it became schwa sound - Uh, Uh, so
438
2090030
8200
34:58
it happened to me, happened to me, happened to me. If you wanna repeat, dear listener
439
2098230
3970
35:02
or dear, dear viewer: happened to me. Finally, the last word I took note of is he said that
440
2102200
7180
35:09
it was spooky. What does it mean if something is spooky?
441
2109380
2899
35:12
I lost sleep that night. It was, it was really spooky.
442
2112279
4881
35:17
Did you guys get that? That is a spooky laugh. Yeah.
443
2117160
6120
35:23
That was a spooky laugh. So what does that mean?
444
2123280
2930
35:26
If something is spooky, it's, um, it's scary a little. It's, it's, it's scary. It, it's
445
2126210
9139
35:35
a little frightening. Like you hear it or you see it, you experience it could be a spooky
446
2135349
6211
35:41
situation and you, you're frightened by it. I feel like we use this especially in the
447
2141560
5050
35:46
sense of things that come from another world or that have to do with the, the spiritual
448
2146610
6530
35:53
world, with ghosts, with things that we can't explain. So you might hear this a lot at Halloween.
449
2153140
6189
35:59
If you're, if you're in the United States, you might, you might see this a lot because
450
2159329
2381
36:01
of all the ghosts and goblins and witches and all this make you feel, they're, they're
451
2161710
4460
36:06
a bit spooky, right? So if you want to learn a lot more words like that with this podcast,
452
2166170
5490
36:11
Thiago mentioned before, but a really great way to do it is by, if you're not already,
453
2171660
4350
36:16
listen to this episode and all of our episodes on the RealLife English app because you get
454
2176010
3700
36:19
an interactive transcript so you won't miss a single thing. If you're having a little
455
2179710
3090
36:22
bit of difficulty understanding us, if we're speaking too fast, that'll definitely help
456
2182800
4000
36:26
you. And we have vocabulary flashcards, you know, with, when you're a free user, you get
457
2186800
5460
36:32
five free flashcards for every episode. So that can be really helpful so that you never
458
2192260
5450
36:37
forget these new words. Every week you can be learning five new words and you can put
459
2197710
3530
36:41
them into practice right away by speaking with other English speakers on the app at
460
2201240
4260
36:45
the press a button. It's the only place where anytime, anywhere you can press a button and
461
2205500
3609
36:49
it's gonna be connected to another English speaker for a fun and dynamic conversation
462
2209109
3521
36:52
in English, you can talk about AI, you can talk about spooky things, you can talk about
463
2212630
4000
36:56
someone's, uh, bodacious cooking skills, you know, whatever you want. Now, Casse, we have
464
2216630
7520
37:04
a really nice testimonial from one of our Premium app users. Do you wanna go ahead and
465
2224150
5340
37:09
share that? I'm a longtime RealLife app user and fan. I had the chance to test the Premium
466
2229490
4820
37:14
experience version, which is very novel and top-notch indeed. The app provides everything
467
2234310
4700
37:19
you need, motivates you to master your English, encourages you to use the app on a daily basis,
468
2239010
7130
37:26
focuses on new words, on super cool flashcards and builds and brings together a valuable
469
2246140
4820
37:30
language learning community with real-life people. Perfect application for learners who
470
2250960
5639
37:36
would like to master and live their English. Well, embrace the difficult journey ahead
471
2256599
6051
37:42
and download the app. The magical Premium experience is waiting for you. Aww yeah! Thanks
472
2262650
5980
37:48
so much, Maria. Wow. Amazing. " Amazing. Glad to hear that you're having such
473
2268630
4840
37:53
a bodacious experience with the app. And she, she emphasizes with real people,
474
2273470
6370
37:59
so not with an app, or not with an AI bot, with real-life people.
475
2279840
5650
38:05
So if you haven't tried the Premium version app yet, what are you waiting for? You get
476
2285490
6180
38:11
over a hundred transcripts, so transcripts for every single episode, not just the newest
477
2291670
5290
38:16
ones. And you get unlimited access to all 30 flashcards for every episode. On the free
478
2296960
8960
38:25
plan right now, you're only getting three to five sample flashcards on each episode.
479
2305920
5180
38:31
Um, free users can have as many conversations as they want, but you're limited to four to
480
2311100
6310
38:37
eight minutes. And a lot of you have told us, you know, that sometimes you have a really
481
2317410
2850
38:40
great connection with someone and you want to keep speaking with them and it's just over
482
2320260
3700
38:43
too fast. So if you become a Premium plan member, you get unlimited speaking coins.
483
2323960
4770
38:48
So you can remove the timer anytime you choose and speak for as long with those, as long
484
2328730
6230
38:54
as you want with those special connections. So what are you waiting for? Why don't you
485
2334960
4920
38:59
check it out? And we look forward to seeing you inside the Premium. And we were tickled,
486
2339880
6110
39:05
I could say because we did a few episodes back, Thiago and I challenged you guys to
487
2345990
5310
39:11
let us know whether - we were talking about zombies a recent episode, episode 325. Don't
488
2351300
4779
39:16
ask us, you know, why we were talking about zombies, but it was relevant to the, tothe
489
2356079
5301
39:21
topic. And we asked you guys if you think you would survive the zombie apocalypse, why
490
2361380
5060
39:26
or why not? And we got a couple of really great responses. So Thiago I'll let you read
491
2366440
4320
39:30
one of the ones from YouTube and then we also got another one by email that we can give
492
2370760
4089
39:34
you guys the, um, a summary of, we can sum it up for you.
493
2374849
4250
39:39
All right. So Rosa, she left us a comment here on YouTube saying, Hello Ethan. Thanks
494
2379099
4631
39:43
for this useful material and all the pieces of advice. But tangentially speaking, I'd
495
2383730
5300
39:49
say that I do not think I could survive any apocalypse, let alone a zombie one. I love
496
2389030
6781
39:55
the fact that she actually used a word that we defined in that episode, right, Ethan?
497
2395811
4528
40:00
Which one was it? Tangentially, is that one of the words we used in that one actually?
498
2400339
3791
40:04
Yeah. We used that one, on a tangent. And then you said (That is great.) that difficult
499
2404130
5150
40:09
phrase. (Tangentially speaking.) Yeah. Tangent. So if you're speaking tangentially,
500
2409280
2960
40:12
it means you're speaking about something not related to the subject at hand, right? And
501
2412240
4420
40:16
they also said, I think it's a she - Rosa, in my experience would be a woman, but sorry
502
2416660
4510
40:21
if that's wrong, uh, I could not survive any apocalypse, let alone a zombie one. What does
503
2421170
4899
40:26
that let, ""let alone"" mean? Would it be similar to, especially, especially a zombie
504
2426069
4981
40:31
one? Yeah. It's (Yeah) a way toemphasize something. Mm-hmm, it's an emphasizer. Yeah, it was a
505
2431050
4890
40:35
really great response. Thanks so much, Rosa, for taking the time to let us know about your,
506
2435940
5060
40:41
your probabilities of survival in the zombie apocalypse. We got another one too by email
507
2441000
5619
40:46
that stood out to us because they included a picture, but they said that they would survive
508
2446619
5361
40:51
it because what they would do is surround their house by treadmills and they, they attached
509
2451980
5280
40:57
an image so that we could like really understand how this would work. And it's, it's, if you're
510
2457260
4589
41:01
watching the video, you can see this, but if you're not, then we'll try to describe
511
2461849
2541
41:04
it. It's just a house that has the walls outside the house are completely, there's treadmills
512
2464390
4910
41:09
lined up going towards the house on the entire exterior of the house. So the zombies tried
513
2469300
5300
41:14
to get in and they just are walking in place. But this did, this did leave, like, give me
514
2474600
6160
41:20
some follow up questions because first of all, were you getting that many treadmills?
515
2480760
4309
41:25
Like you're gonna have to be able (from the gym) to survive (Ethan) . There's like 50.
516
2485069
4451
41:29
Like robbing a gym. 50 treadmills there, but it's a very creative
517
2489520
3799
41:33
answer. But I have my doubts as to how, how well it would work. But there's definitely,
518
2493319
4391
41:37
I'm, I'm definitely poking some holes in that theory. "
519
2497710
2930
41:40
What does it mean to poke holes in a theory? So like, literally poking holes is like if
520
2500640
5590
41:46
you have, um, what would you poke holes in? If you have a piece of paper and a needle,
521
2506230
4650
41:50
you could poke holes in it, right? It's not something most people would do maybe with
522
2510880
4340
41:55
like your pen. (like destroy) . Yeah. And so if you're poking holes in a theory, it's
523
2515220
5399
42:00
like you're saying, you know, there's all these holes in it and stuff, you better fill
524
2520619
2671
42:03
them in for me. So we need a, we need a follow up response on that. If you're, you're listening,
525
2523290
3970
42:07
we, we wanted to actually do a follow up on that too. Cause we're talking about artificial
526
2527260
3270
42:10
intelligence. So zombie apocalypse, that's one scenario like Last Of Us and everything.
527
2530530
3820
42:14
But we talked about like Terminator, the Matrix, there's all these different artificial intelligence
528
2534350
4460
42:18
apocalypses. I think that's a plural, right? It's an interesting one. Like it's not Apocalypses,
529
2538810
5250
42:24
it's Apocalypses, side note. But yes, please, dear listener, dear viewer, if you're on YouTube,
530
2544060
6960
42:31
write in the comments, will you survive the AI apocalypse? Why or why not? Or you can
531
2551020
4920
42:35
send us an email at [email protected] and let us know. We'd love it too, if you
532
2555940
6501
42:42
attach a meme, that definitely makes it all the more fun for us to, when we, when we open
533
2562441
5229
42:47
up our, our Sunday morning email with our coffee and get to giggle at good meme.
534
2567670
7980
42:55
All right. So now it's time for the RealLife Way moment. And I would say that we, I would
535
2575650
11160
43:06
connect today's conversation about AI and ChatGPT and everything with two points here
536
2586810
5980
43:12
from the RealLife Way. First, uh, making it fun because it can definitely be fun, you
537
2592790
6760
43:19
know, you can certainly have fun with the tool and asking all sorts of, uh, things and
538
2599550
5660
43:25
wacky questions. Yeah. And have fun with it. It can be addicting. Even. You have to be
539
2605210
4470
43:29
careful. Exactly. You have, you have to be careful
540
2609680
2890
43:32
with that. It can be (Use responsibly) addictive. Yes. Yes. In moderation, right? But, uh, I
541
2612570
7600
43:40
would say mainly it's about activation. The part of activating your English, because it
542
2620170
4189
43:44
allows you to study English more deliberately. Because as we demonstrated today, you can
543
2624359
5551
43:49
ask the tool to define words for you, give you examples, uh, correct sentences for you,
544
2629910
7270
43:57
or giving you recommendations of what to do or what to consume in English. So I would
545
2637180
4741
44:01
say it's related mostly to activating your English.
546
2641921
4269
44:06
Most definitely. I think a really important part of that too is what we were talking about,
547
2646190
3899
44:10
that you have to get good at giving the right prompts and being very specific, asking the
548
2650089
4881
44:14
right kind of questions. So actually in line with that thought, uh, we came up with some
549
2654970
5581
44:20
different prompts that you all can use that are going to, to help you. And we've actually
550
2660551
4258
44:24
tested all these out, so we've seen that they can give you some good results. So for example,
551
2664809
4581
44:29
if you're wanting to have a conversation with someone, and maybe it's on a topic that you're
552
2669390
4990
44:34
not so comfortable with yet, actually Thiago and I recorded a few episodes back about,
553
2674380
4300
44:38
uh, an episode about fluency versus proficiency, and proficiency being something really important
554
2678680
4320
44:43
that enough learners don't get and everything. So this can actually help you to be more proficient
555
2683000
3609
44:46
about talking, you know, because you'll be able to talk about many different subjects
556
2686609
3141
44:49
confidently. So maybe you're really interested in having a conversation with someone, maybe
557
2689750
4691
44:54
on the RealLife app even. You want to connect with someone and, and talk about the environment,
558
2694441
4129
44:58
but you just aren't sure what are the words you can use or how can you phrase things.
559
2698570
3560
45:02
So you could say, Hey, Chat GPT write me a short paragraph about the environment. And
560
2702130
4590
45:06
you know, that might give you an idea. We always talk about how scripts can be really
561
2706720
3960
45:10
great to actually get some different, uh, scripts and practice these on your own first,
562
2710680
5120
45:15
even record yourself. Listen, listen back, correct yourself. And if you are getting this
563
2715800
4049
45:19
from ChatGPT, then you know it'll actually be grammatically correct and use the right
564
2719849
4423
45:24
words and whatnot. Um, you could ask it to actually give you an English study plan. This
565
2724272
4878
45:29
is something a lot of learners have asked us about, you know, is they're kinda confused
566
2729150
4090
45:33
about what do I do next? You know, uh, maybe I'm already watching TV series, I'm listening
567
2733240
4540
45:37
to podcasts and so on, but you know, I'm kind of stuck. I'm not sure what to do next. So
568
2737780
5460
45:43
you could say, Hey, make me an English study plan based on the Godfather - can see it there
569
2743240
4460
45:47
behind Thiago. And we, we, we've talked a lot about, a lot about this movie cuz it's
570
2747700
5751
45:53
one of Thiago's favorites. So maybe Thiago wants to learn Italian. You could say, Hey,
571
2753451
3089
45:56
give me an Italian study plan based on the Italian version of the Godfather. You could
572
2756540
6340
46:02
say, give me a list of phrasal verbs with, you know, any sort of example, so you could
573
2762880
4100
46:06
use take,or you could say phrasal verbs that have to do with, uh, with work, with office
574
2766980
6300
46:13
environment or something like that and their meanings. So that's really nice because you
575
2773280
4070
46:17
know, it'll, if you're wanting to buff up on your, can you even use a phrase of verb
576
2777350
4810
46:22
there? If you, you wanna buff up on your knowledge about phrasal verbs, that could be really
577
2782160
3600
46:25
helpful, or expressions or vocabulary or whatever else. You could say, give me an example sentence
578
2785760
5221
46:30
or example sentences with the word whatever, mind-blowing, or bodacious or anything like
579
2790981
5499
46:36
this. If you just learn a new word and you wanna see more context, that can be really
580
2796480
2940
46:39
helpful. You could even add these to, you know, your study cards. That might be nice.
581
2799420
4529
46:43
We also mentioned DALL-E, we didn't actually look at this today, but if you aren't familiar
582
2803949
3681
46:47
with DALL-E, it's, it's from the same creators as, uh, of ChatGPT, but it has to do with
583
2807630
5940
46:53
visual, with images, and it's actually, it's really fun to play with as well. It's really
584
2813570
2910
46:56
amazing because I've actually seen some compilations, for example, that were like making the Simpsons
585
2816480
6720
47:03
into an, uh, 1980s fantasy film. You know, kind of like the style of, of Stranger Things.
586
2823200
6550
47:09
And it did like a really amazing job like making, making realistic looking people who
587
2829750
3490
47:13
looked like the characters in the Simpsons and giving that, that, that feel, that vibe,
588
2833240
4730
47:17
you know, that Stranger Things has. So you can use DALL-E for visual representation of
589
2837970
4850
47:22
maybe a certain expression you learned, or a word, or a phrase or something, because
590
2842820
5049
47:27
a lot of us are very visual learners, right? So if you actually get that, maybe you'll
591
2847869
2911
47:30
get something that's kind of strange or really interesting thought-provoking that will help
592
2850780
4150
47:34
you to, you know, better remember those words. So I think that's it. There's really just
593
2854930
8159
47:43
unlimited things that you can do with this technology. "
594
2863089
2291
47:45
"It's, it's crazy because as I said before when we started this episode, I said that
595
2865380
4630
47:50
I was worried. Um, but when you realize like the potential that it has when you start like
596
2870010
6140
47:56
thinking out of the box, when you start thinking of like creative ways to use the tool and
597
2876150
6360
48:02
that you start realizing all these wonderful ways that it can actually improve your life.
598
2882510
4920
48:07
Um, you know, whether that means like, I'm, you know, we spoke about bodacious like throughout
599
2887430
4780
48:12
this episode and I was like, oh, cool. Tell me, I asked ChatGPT while, uh, we were talking,
600
2892210
6010
48:18
I asked ChatGPT to, you know, tell me some songs, Recommend some songs with the word,
601
2898220
6380
48:24
with the lyrics, you, you know, (Ooh) sorry, with the lyric bodacious. And I found some
602
2904600
4691
48:29
really cool ones from the eighties and even more recent ones. So I look forward to listening
603
2909291
3959
48:33
to that later. So it's, I don't know, I think there's just so many ways that you can improve.
604
2913250
5619
48:38
" Some learners have gotta check that out for
605
2918869
2571
48:41
sure. You definitely should. (Bodacious songs.)
606
2921440
1980
48:43
Bodacious songs, you know. That's a great, another great use. I, I hadn't
607
2923420
4331
48:47
thought about that, but, uh, you actually like, learn a new word. It's like, you know,
608
2927751
2798
48:50
tell me a song where I can, that, that has thisword or this expression or whatever. Yeah.
609
2930549
4011
48:54
One interesting thing I've seen with it too is you can help it to kind of get to know
610
2934560
4660
48:59
you. So you can tell ChatGPT, you know, I'm an intermediate learner. I really love playing
611
2939220
7060
49:06
tennis. I love spending time outdoors, uh, I have a dog and I live in Barcelona. Such
612
2946280
7549
49:13
and such, such like, whatever, whatever things you want to tell it about yourself and ask
613
2953829
4540
49:18
it to give you personalized responses based on that information that has to do with you.
614
2958369
4681
49:23
So if you, if you also help it to get to know you a little bit, maybe not BingAI because
615
2963050
5320
49:28
it's, it's a little bit, it's a little bit crazy. It might ask you to run away with it
616
2968370
3880
49:32
to, to Mexico, but, uh, but you can get creative, you can think outside the box here and, and
617
2972250
6369
49:38
you can really play around with this and everything, help it get to know you because then, you
618
2978619
4161
49:42
know, it'll, it'll give you really personalized responses. It'll give you TV series that have
619
2982780
4260
49:47
to do with tennis or ways that you can meet to other English speakers who also share your
620
2987040
4799
49:51
affinity for dogs. Whatever the case is. We have today's Big Challenge. So the Big Challenge
621
2991839
9561
50:01
for this episode, we've been talking all about ChatGPT. So we wanted you to actually, you
622
3001400
5330
50:06
know, put your money where your mouth is, to actually try this out. You know. share
623
3006730
6639
50:13
the most interesting response. Go, go try ChatGPT or another BingAI or any of these
624
3013369
6541
50:19
other AI technologies, and share the most interesting response that you get from it
625
3019910
4880
50:24
down in the comments below if you're on YouTube or you can send us an email at [email protected].
626
3024790
4660
50:29
We're really curious to see, and we'll share some of your responses on one of, one, on
627
3029450
5389
50:34
one of our upcoming episodes. I got tongue-tied there. So, thanks so much for listening to
628
3034839
5571
50:40
our rant all about AI. Hopefully you've enjoyed it. We hope that it's helpful, that it can,
629
3040410
26840
51:07
you can actually use some of this advice to improve
630
3067250
15630
51:22
your English. So without further ado, 1, 2, 3. Aww (Aww) (Aww) yeah. (yeah) (yeah)
631
3082880
5770
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7