What is an LLM?

What is an LLM—and why do they matter for teachers and students?

In this AI Foundations video from Ed3, you’ll learn what an LLM (Large Language Model) is, how it works, and why it can feel surprisingly “human” even though it isn’t. If you’ve ever wondered how tools like chatbots generate fluent responses, lesson ideas, summaries, or feedback—this is the core concept behind it.

An LLM is a deep learning model trained on massive amounts of text (books, articles, websites, and public documents). It doesn’t “know” facts the way people do. Instead, it predicts the next most likely word one token at a time, based on learned patterns. That predictive power can look like thinking—especially because language contains so much knowledge and reasoning.

This video covers:

  • What “Large,” “Language,” and “Model” really mean
  • Why scale (datasets + parameters) makes LLMs flexible and style-adaptive
  • Tokens, training, and why outputs can sound confident even when wrong
  • How methods like RLHF shape more helpful, human-like responses
  • LLM risks educators should explicitly teach: bias, misinformation, and outdated data
  • Why AI companions can create a false sense of relationship and authority (especially for children)
  • Practical classroom strategies for thoughtful LLM use: drafts, verification, attribution, and process-based assessment

A common misconception is that LLMs are truth machines. They’re not. They’re pattern generators, and the quality of the output depends on the prompt, the context, and the user’s ability to verify and think critically.

This video is part of the AI Foundations series by Ed3, supporting educators worldwide in making informed, ethical, and human-centered decisions about AI in classrooms.

👉 Learn more about Ed3: https://www.ed3global.org

👉 Explore professional learning, courses, and events designed for educators navigating AI responsibly.

👉 Join our community of practice: https://community.ed3global.org

Timestamps

00:33 What an LLM is

00:47 “Large” datasets + parameters

01:01 “Language” tasks LLMs can do

01:12 “Model” = next-word prediction

01:42 Tokens explained

02:01 RLHF and why chatbots can feel personal

02:26 Why LLMs can seem like they’re thinking

02:29 Real-world uses (and why that matters in schools)

02:45 Shared risks with generative AI

04:22 LLM-specific concerns: bias, misinformation, hidden training data

04:44 The danger of AI companions and blurred roles

05:05 How educators can use LLMs well (and safely)

05:37 Seven tips for thoughtful classroom use

06:22 Helpers, not friends or authority figures

06:45 Separating hype from reality

show transcript

1

00:00:04,866 --> 00:00:05,200

Yeah.

2

00:00:05,200 --> 00:00:06,500

You were like “rah rah rah rah”.

3

00:00:06,500 --> 00:00:08,266

Oh my gosh. Don't bring that up.

4

00:00:08,266 --> 00:00:08,866

I don't know

5

00:00:08,866 --> 00:00:09,766

do I stay blonde

6

00:00:09,766 --> 00:00:12,366

do I spice it up like go green, blue, pink?

7

00:00:12,366 --> 00:00:13,000

I don't know.

8

00:00:13,000 --> 00:00:14,400

Well, I think you look good

9

00:00:14,400 --> 00:00:15,666

no matter what,

10

00:00:15,666 --> 00:00:16,766

but a bright color

11

00:00:16,766 --> 00:00:19,366

will tend to clash with your outfits.

12

00:00:19,366 --> 00:00:22,133

Blonde goes with anything.

13

00:00:22,133 --> 00:00:23,966

See Chili dog has recently learned

14

00:00:23,966 --> 00:00:25,300

how to talk,

15

00:00:25,300 --> 00:00:25,733

and she's

16

00:00:25,733 --> 00:00:28,200

actually really good at giving advice,

17

00:00:28,200 --> 00:00:28,833

but might be

18

00:00:28,833 --> 00:00:29,633

because I trained her

19

00:00:29,633 --> 00:00:32,633

on a large language model.

20

00:00:33,000 --> 00:00:35,966

LLM stands for Large Language Model.

21

00:00:35,966 --> 00:00:37,466

It's a type of deep learning model

22

00:00:37,466 --> 00:00:39,566

trained on massive amounts of text.

23

00:00:39,566 --> 00:00:40,500

Everything from books

24

00:00:40,500 --> 00:00:41,666

and scientific articles

25

00:00:41,666 --> 00:00:43,466

to websites, social media posts

26

00:00:43,466 --> 00:00:45,033

and public documents.

27

00:00:45,033 --> 00:00:47,033

Let's break down what it means.

28

00:00:47,033 --> 00:00:47,966

“Large” means it's

29

00:00:47,966 --> 00:00:50,100

trained on huge datasets,

30

00:00:50,100 --> 00:00:51,833

billions of words from books,

31

00:00:51,833 --> 00:00:54,500

websites, articles and conversations.

32

00:00:54,500 --> 00:00:56,400

It also uses billions of parameters

33

00:00:56,400 --> 00:00:57,500

that are like knobs

34

00:00:57,500 --> 00:00:58,900

that it uses to adjust

35

00:00:58,900 --> 00:01:00,733

while learning patterns.

36

00:01:00,733 --> 00:01:01,500

“Language”

37

00:01:01,500 --> 00:01:03,866

means it works with human language.

38

00:01:03,866 --> 00:01:05,800

Reading, writing, translating,

39

00:01:05,800 --> 00:01:07,766

summarizing, answering questions,

40

00:01:07,766 --> 00:01:09,433

and even chatting.

41

00:01:09,433 --> 00:01:12,400

And “Model” means it's a predictive system.

42

00:01:12,400 --> 00:01:13,600

Given some text,

43

00:01:13,600 --> 00:01:14,166

it predicts

44

00:01:14,166 --> 00:01:16,666

what word or phrase should come next

45

00:01:16,666 --> 00:01:19,000

based on patterns it has learned.

46

00:01:19,000 --> 00:01:21,233

This scale is what makes LLMs

47

00:01:21,233 --> 00:01:22,700

so flexible.

48

00:01:22,700 --> 00:01:25,000

They've seen examples of formal writing,

49

00:01:25,000 --> 00:01:26,100

casual conversation,

50

00:01:26,100 --> 00:01:29,100

technical instructions, and even poetry.

51

00:01:29,166 --> 00:01:30,000

That variety

52

00:01:30,000 --> 00:01:32,033

allows them to mimic styles and adapt

53

00:01:32,033 --> 00:01:34,800

responses to different contexts.

54

00:01:34,800 --> 00:01:37,300

An LLM doesn't know facts.

55

00:01:37,300 --> 00:01:38,400

Instead, it predicts

56

00:01:38,400 --> 00:01:40,800

the next most likely word in a sequence,

57

00:01:40,800 --> 00:01:42,633

one token at a time.

58

00:01:42,633 --> 00:01:44,666

And tokens are simply chunks of text

59

00:01:44,666 --> 00:01:47,466

like words, syllables, or characters.

60

00:01:47,466 --> 00:01:48,633

During training,

61

00:01:48,633 --> 00:01:51,300

the model sees countless examples of text

62

00:01:51,300 --> 00:01:52,700

and adjusts its parameters

63

00:01:52,700 --> 00:01:55,100

to reduce mistakes in prediction.

64

00:01:55,100 --> 00:01:55,800

Over time,

65

00:01:55,800 --> 00:01:58,133

it gets very good at producing fluent,

66

00:01:58,133 --> 00:01:59,966

coherent text.

67

00:01:59,966 --> 00:02:01,500

Some LLMs are refined

68

00:02:01,500 --> 00:02:02,866

with additional methods

69

00:02:02,866 --> 00:02:03,900

like Reinforcement

70

00:02:03,900 --> 00:02:05,500

Learning with Human Feedback,

71

00:02:05,500 --> 00:02:07,666

which aligns outputs with human values

72

00:02:07,666 --> 00:02:09,366

or desired behaviors.

73

00:02:09,366 --> 00:02:10,633

This is how they can generate

74

00:02:10,633 --> 00:02:12,000

human-like conversations

75

00:02:12,000 --> 00:02:14,666

so they can seem personal, empathetic,

76

00:02:14,666 --> 00:02:16,333

or even friendly.

77

00:02:16,333 --> 00:02:17,633

Because language encodes

78

00:02:17,633 --> 00:02:18,600

so much knowledge

79

00:02:18,600 --> 00:02:19,466

from history,

80

00:02:19,466 --> 00:02:22,466

science, culture, and everyday reasoning,

81

00:02:22,533 --> 00:02:23,466

predicting words

82

00:02:23,466 --> 00:02:26,133

well can make it look like it's thinking

83

00:02:26,133 --> 00:02:26,866

Today,

84

00:02:26,866 --> 00:02:29,133

LLMs are used for everything

85

00:02:29,133 --> 00:02:30,233

wedding speeches,

86

00:02:30,233 --> 00:02:32,366

brainstorming, presentations,

87

00:02:32,366 --> 00:02:34,200

breakup texts, lesson writing,

88

00:02:34,200 --> 00:02:35,166

quiz banks, book

89

00:02:35,166 --> 00:02:36,133

authorship articles,

90

00:02:36,133 --> 00:02:36,933

song lyrics, trip

91

00:02:36,933 --> 00:02:39,933

itineraries, legal briefs, you name it.

92

00:02:40,533 --> 00:02:43,400

Since LLMs are used for generative AI,

93

00:02:43,400 --> 00:02:45,766

we have to watch out for the same issues.

94

00:02:45,766 --> 00:02:48,766

Let's cut to our primer on generative AI.

95

00:02:48,900 --> 00:02:50,700

Generative AI models are trained on

96

00:02:50,700 --> 00:02:52,233

massive datasets,

97

00:02:52,233 --> 00:02:53,633

much of which comes from books,

98

00:02:53,633 --> 00:02:55,300

articles, art, and media

99

00:02:55,300 --> 00:02:56,700

created by real people

100

00:02:56,700 --> 00:02:58,533

available on the internet.

101

00:02:58,533 --> 00:02:59,533

That means that the outputs

102

00:02:59,533 --> 00:03:00,300

might remix

103

00:03:00,300 --> 00:03:01,966

or resemble copyrighted works

104

00:03:01,966 --> 00:03:03,766

without proper attribution.

105

00:03:03,766 --> 00:03:05,633

We need to help students understand

106

00:03:05,633 --> 00:03:07,633

just because AI generates content

107

00:03:07,633 --> 00:03:09,000

doesn't mean they own it

108

00:03:09,000 --> 00:03:10,466

or can use it freely.

109

00:03:10,466 --> 00:03:11,833

This raises important lessons

110

00:03:11,833 --> 00:03:13,166

about digital citizenship

111

00:03:13,166 --> 00:03:14,733

and respecting creators.

112

00:03:15,800 --> 00:03:16,933

Students can use AI

113

00:03:16,933 --> 00:03:19,366

to generate essays, homework, or artwork

114

00:03:19,366 --> 00:03:21,033

that looks like their own work,

115

00:03:21,033 --> 00:03:22,200

but it isn't.

116

00:03:22,200 --> 00:03:23,066

This challenges

117

00:03:23,066 --> 00:03:24,633

how we measure learning.

118

00:03:24,633 --> 00:03:25,366

Since tools

119

00:03:25,366 --> 00:03:26,400

detecting generative

120

00:03:26,400 --> 00:03:28,666

AI plagiarism are unreliable,

121

00:03:28,666 --> 00:03:30,266

and because AI is everywhere

122

00:03:30,266 --> 00:03:31,866

in the workforce as well,

123

00:03:31,866 --> 00:03:34,866

the answer isn't banning AI tools.

124

00:03:34,900 --> 00:03:35,666

The first answer

125

00:03:35,666 --> 00:03:36,733

is having a conversation

126

00:03:36,733 --> 00:03:39,366

with students about how we can use AI.

127

00:03:39,366 --> 00:03:40,966

When is AI support allowed?

128

00:03:40,966 --> 00:03:43,566

Like brainstorming, drafting, outlining?

129

00:03:43,566 --> 00:03:45,466

And when is it crossing into plagiarism?

130

00:03:45,466 --> 00:03:46,733

Like copy pasting

131

00:03:46,733 --> 00:03:48,633

without engagement or learning.

132

00:03:48,633 --> 00:03:49,500

And secondly,

133

00:03:49,500 --> 00:03:51,366

we need to redesign our assignments

134

00:03:51,366 --> 00:03:52,366

to allow for thoughtful

135

00:03:52,366 --> 00:03:53,666

attribution of generative

136

00:03:53,666 --> 00:03:55,200

AI that allows students

137

00:03:55,200 --> 00:03:56,266

to demonstrate their own

138

00:03:56,266 --> 00:03:58,500

skills and understanding.

139

00:03:58,500 --> 00:04:00,166

Since generative AI creates

140

00:04:00,166 --> 00:04:02,266

content by remixing patterns,

141

00:04:02,266 --> 00:04:04,966

it doesn't produce truly original ideas.

142

00:04:04,966 --> 00:04:06,566

It simulates them.

143

00:04:06,566 --> 00:04:07,466

That means if students

144

00:04:07,466 --> 00:04:09,000

lean too heavily on AI.

145

00:04:09,000 --> 00:04:10,466

Their own voices, perspectives,

146

00:04:10,466 --> 00:04:12,700

and creativity can get lost.

147

00:04:12,700 --> 00:04:13,733

We play a key role

148

00:04:13,733 --> 00:04:15,233

in encouraging balance,

149

00:04:15,233 --> 00:04:17,733

using AI as a partner to spark ideas

150

00:04:17,733 --> 00:04:19,466

and not as a replacement for students

151

00:04:19,466 --> 00:04:21,066

authentic contributions.

152

00:04:22,200 --> 00:04:23,466

We also have to consider

153

00:04:23,466 --> 00:04:25,500

LLM specific issues.

154

00:04:25,500 --> 00:04:28,500

For example, the data isn't perfect.

155

00:04:28,600 --> 00:04:29,300

It can contain

156

00:04:29,300 --> 00:04:30,133

bias,

157

00:04:30,133 --> 00:04:33,133

outdated information, or misinformation.

158

00:04:33,166 --> 00:04:34,033

And because companies

159

00:04:34,033 --> 00:04:35,366

don't always disclose

160

00:04:35,366 --> 00:04:36,900

the exact data sets,

161

00:04:36,900 --> 00:04:38,200

we need to stay critical

162

00:04:38,200 --> 00:04:40,233

of what these models produce.

163

00:04:40,233 --> 00:04:41,400

Additionally,

164

00:04:41,400 --> 00:04:44,166

because an LLM can seem human,

165

00:04:44,166 --> 00:04:45,300

people of all ages

166

00:04:45,300 --> 00:04:46,800

can sometimes forget who

167

00:04:46,800 --> 00:04:48,666

or what they're talking to.

168

00:04:48,666 --> 00:04:50,966

Much like me talking to my dog,

169

00:04:50,966 --> 00:04:52,066

it can create a false sense

170

00:04:52,066 --> 00:04:53,100

of understanding,

171

00:04:53,100 --> 00:04:55,300

cause over-reliance on the tool,

172

00:04:55,300 --> 00:04:57,400

and blur human roles.

173

00:04:57,400 --> 00:04:58,266

The dangers of these

174

00:04:58,266 --> 00:04:59,233

AI companions

175

00:04:59,233 --> 00:05:01,600

are not to be underestimated.

176

00:05:01,600 --> 00:05:02,900

For educators,

177

00:05:02,900 --> 00:05:05,600

LLMs are the backbone of many text-based

178

00:05:05,600 --> 00:05:06,733

AI tools.

179

00:05:06,733 --> 00:05:08,233

They're a powerful thought partner

180

00:05:08,233 --> 00:05:09,000

and teammate

181

00:05:09,000 --> 00:05:11,666

for improving teaching and learning.

182

00:05:11,666 --> 00:05:13,066

They can differentiate content

183

00:05:13,066 --> 00:05:14,600

for diverse reading levels,

184

00:05:14,600 --> 00:05:16,100

provide multiple examples

185

00:05:16,100 --> 00:05:18,100

or analogies for the same concept,

186

00:05:18,100 --> 00:05:19,933

support multilingual classrooms

187

00:05:19,933 --> 00:05:21,200

through instant translation

188

00:05:21,200 --> 00:05:22,933

and so much more.

189

00:05:22,933 --> 00:05:24,333

On the flip side,

190

00:05:24,333 --> 00:05:25,433

the power of these tools

191

00:05:25,433 --> 00:05:27,733

becomes a concern for safety.

192

00:05:27,733 --> 00:05:29,266

LLMs can convince people,

193

00:05:29,266 --> 00:05:30,500

especially children,

194

00:05:30,500 --> 00:05:33,066

that they are real humans.

195

00:05:33,066 --> 00:05:33,900

Here are some tips

196

00:05:33,900 --> 00:05:34,633

that will encourage

197

00:05:34,633 --> 00:05:37,200

more thoughtful use of LLMs.

198

00:05:37,200 --> 00:05:38,100

One.

199

00:05:38,100 --> 00:05:40,066

Treat LLM outputs as drafts,

200

00:05:40,066 --> 00:05:41,733

not final products.

201

00:05:41,733 --> 00:05:42,433

Two.

202

00:05:42,433 --> 00:05:44,033

Always cross-check information

203

00:05:44,033 --> 00:05:46,133

against trusted sources.

204

00:05:46,133 --> 00:05:47,000

Three.

205

00:05:47,000 --> 00:05:48,166

Teach kids to ask

206

00:05:48,166 --> 00:05:49,833

better questions of the LLM

207

00:05:49,833 --> 00:05:51,933

and model critical reading.

208

00:05:51,933 --> 00:05:54,366

Four, design process-oriented

209

00:05:54,366 --> 00:05:56,366

assignments and require attribution

210

00:05:56,366 --> 00:05:57,733

of the LLM.

211

00:05:57,733 --> 00:05:58,566

Five.

212

00:05:58,566 --> 00:06:01,033

Discuss authorship and integrity.

213

00:06:01,033 --> 00:06:02,900

When is an LLM helpful

214

00:06:02,900 --> 00:06:04,600

and when is it damaging?

215

00:06:04,600 --> 00:06:05,533

Six.

216

00:06:05,533 --> 00:06:07,966

Discuss overreliance on AI.

217

00:06:07,966 --> 00:06:08,633

Because it can be

218

00:06:08,633 --> 00:06:10,200

a convincing thought partner,

219

00:06:10,200 --> 00:06:11,333

it's important to discuss

220

00:06:11,333 --> 00:06:12,100

habits of mind

221

00:06:12,100 --> 00:06:13,500

that help students retain

222

00:06:13,500 --> 00:06:15,166

their self-governance.

223

00:06:15,166 --> 00:06:16,300

And most importantly,

224

00:06:16,300 --> 00:06:17,233

Seven.

225

00:06:17,233 --> 00:06:18,433

Engage in constant

226

00:06:18,433 --> 00:06:19,466

dialog with students

227

00:06:19,466 --> 00:06:22,500

about how LLMs and AI are not human.

228

00:06:22,533 --> 00:06:24,600

LLMs can be positioned as helpers,

229

00:06:24,600 --> 00:06:26,400

but should not be seen as friends,

230

00:06:26,400 --> 00:06:28,866

peers, or authority figures.

231

00:06:28,866 --> 00:06:29,933

Large Language Models

232

00:06:29,933 --> 00:06:32,066

are powerful pattern generators,

233

00:06:32,066 --> 00:06:33,133

but their outputs reflect

234

00:06:33,133 --> 00:06:34,800

the limits of their data.

235

00:06:34,800 --> 00:06:37,300

The opportunity isn't just to use them.

236

00:06:37,300 --> 00:06:39,600

It's to model critical engagement,

237

00:06:39,600 --> 00:06:40,533

ensuring students

238

00:06:40,533 --> 00:06:42,400

see the utility of LLMs

239

00:06:42,400 --> 00:06:43,733

without over-relying on them

240

00:06:45,133 --> 00:06:46,266

As educators,

241

00:06:46,266 --> 00:06:47,466

knowing how LLMs

242

00:06:47,466 --> 00:06:48,466

work helps us separate

243

00:06:48,466 --> 00:06:50,566

the hype from the reality.

244

00:06:50,566 --> 00:06:51,633

So we can make wiser

245

00:06:51,633 --> 00:06:53,233

choices in our classrooms.

246

00:06:53,966 --> 00:06:54,666

Shhh.