00:00:03.782 --> 00:00:04.062

Hello,

00:00:04.243 --> 00:00:05.522

and welcome to the TeacherCast

00:00:05.644 --> 00:00:06.464

Educational Network.

00:00:06.484 --> 00:00:07.564

My name is Jeff Bradbury,

00:00:07.623 --> 00:00:09.605

and welcome to Digital Learning Today.

00:00:10.025 --> 00:00:10.845

On today's episode,

00:00:10.865 --> 00:00:11.605

we're going to be talking

00:00:11.804 --> 00:00:13.766

all about artificial intelligence.

00:00:14.326 --> 00:00:15.785

We're going to be defining what it is,

00:00:15.805 --> 00:00:17.227

how school districts are

00:00:17.466 --> 00:00:18.766

currently taking advantage of it,

00:00:19.106 --> 00:00:20.327

and how you can safely

00:00:20.588 --> 00:00:22.047

deploy it in your classroom.

00:00:22.347 --> 00:00:24.007

I have a fantastic guest on today,

00:00:24.428 --> 00:00:26.609

Dr. Jane Lammers from Edmentum.

00:00:26.949 --> 00:00:28.190

Dr. Lammers, how are you today?

00:00:28.370 --> 00:00:29.289

Welcome to TeacherCast.

00:00:29.974 --> 00:00:30.795

Thank you, Jeff.

00:00:30.875 --> 00:00:31.736

I'm doing great.

00:00:31.815 --> 00:00:32.475

And I'm really looking

00:00:32.536 --> 00:00:33.737

forward to talking with you

00:00:33.798 --> 00:00:35.079

about this subject that

00:00:35.418 --> 00:00:36.981

everybody is talking about.

00:00:37.401 --> 00:00:39.402

I am so excited to have you on.

00:00:39.603 --> 00:00:39.823

You know,

00:00:39.862 --> 00:00:40.683

we've been talking about

00:00:40.743 --> 00:00:42.145

artificial intelligence here.

00:00:42.204 --> 00:00:43.447

It seems like forever.

00:00:44.408 --> 00:00:46.448

Every single time we have a guest come on,

00:00:46.869 --> 00:00:48.892

it's the topic that we have to bring up.

00:00:49.271 --> 00:00:50.613

But I'm excited about having

00:00:50.654 --> 00:00:51.555

you on today because we've

00:00:51.595 --> 00:00:52.597

never really had a chance

00:00:52.637 --> 00:00:55.000

to really start from ground zero,

00:00:55.081 --> 00:00:56.523

start from the beginning here.

00:00:57.064 --> 00:00:58.625

Before we get into those fun questions,

00:00:58.665 --> 00:00:59.487

tell us a little bit about

00:00:59.506 --> 00:01:00.429

yourself and what's

00:01:00.469 --> 00:01:01.731

happening these days at Edmentum.

00:01:02.911 --> 00:01:04.173

Well, thanks for that invitation.

00:01:04.313 --> 00:01:05.635

So, Jeff,

00:01:05.894 --> 00:01:09.298

I was a teacher educator for 15 years.

00:01:09.498 --> 00:01:10.819

I worked in higher education.

00:01:11.540 --> 00:01:13.162

I was a tenured professor at

00:01:13.201 --> 00:01:14.602

the University of Rochester.

00:01:14.623 --> 00:01:16.424

I was a Fulbright scholar

00:01:16.465 --> 00:01:17.766

who got to travel to

00:01:17.846 --> 00:01:19.328

Indonesia just before the

00:01:19.388 --> 00:01:20.709

pandemic and conduct

00:01:20.769 --> 00:01:22.490

research with a partner

00:01:22.531 --> 00:01:24.313

down there on the digital

00:01:24.332 --> 00:01:27.034

literacy practices of Indonesian youth.

00:01:28.557 --> 00:01:30.037

Had a lot of fun doing that

00:01:30.778 --> 00:01:33.700

and was I ran an English

00:01:33.740 --> 00:01:35.123

teacher preparation program

00:01:35.162 --> 00:01:36.584

and also advised doctoral

00:01:36.623 --> 00:01:38.165

students and really enjoyed

00:01:38.286 --> 00:01:39.966

studying how young people

00:01:40.167 --> 00:01:41.709

use technology for their

00:01:41.909 --> 00:01:44.070

own interest driven learning purposes.

00:01:44.631 --> 00:01:46.552

So that's where I kind of came from.

00:01:46.572 --> 00:01:48.295

I was an English teacher

00:01:48.415 --> 00:01:49.936

prior to going into higher ed.

00:01:50.757 --> 00:01:51.977

But as with many folks,

00:01:52.097 --> 00:01:54.280

the pandemic changed things, right?

00:01:55.081 --> 00:01:57.662

And I wanted to be closer to family.

00:01:57.862 --> 00:02:00.245

I wanted to have greater impact.

00:02:00.305 --> 00:02:01.727

There was something about

00:02:02.266 --> 00:02:05.290

the pandemic and the remote

00:02:05.370 --> 00:02:07.131

emergency instruction that

00:02:07.212 --> 00:02:08.673

happened as a result that

00:02:08.872 --> 00:02:12.056

really put what I had been studying,

00:02:12.195 --> 00:02:13.477

which was how are people

00:02:13.637 --> 00:02:14.918

using technologies,

00:02:15.998 --> 00:02:17.579

into the forefront and into

00:02:17.598 --> 00:02:18.419

the conversation.

00:02:18.460 --> 00:02:19.719

And then an opportunity came

00:02:19.860 --> 00:02:21.620

up and I am now the

00:02:21.860 --> 00:02:24.580

Director of Learning Design at Edmentum.

00:02:25.420 --> 00:02:29.421

We are a K-12 digital curriculum provider.

00:02:29.921 --> 00:02:31.882

We aim to be the premier

00:02:31.961 --> 00:02:33.143

learning acceleration

00:02:33.242 --> 00:02:34.983

company that helps get

00:02:35.502 --> 00:02:37.423

young people all across the

00:02:37.462 --> 00:02:39.743

country and in countries around the world

00:02:40.484 --> 00:02:42.566

back up to speed and beyond

00:02:42.825 --> 00:02:43.507

in their learning.

00:02:43.546 --> 00:02:45.568

And we use technology to do

00:02:45.628 --> 00:02:46.911

that in a variety of ways.

00:02:46.991 --> 00:02:48.671

We have intervention programs.

00:02:48.752 --> 00:02:50.875

We also have fully online

00:02:50.935 --> 00:02:54.237

courses and a fully online academy.

00:02:54.699 --> 00:02:57.260

And so my job is to make

00:02:57.360 --> 00:02:58.862

sure that what we know

00:02:58.962 --> 00:03:00.724

about good learning is

00:03:00.805 --> 00:03:03.487

built into the design of our products.

00:03:04.931 --> 00:03:06.592

And when you say good learning,

00:03:06.612 --> 00:03:08.155

how do you define that,

00:03:08.615 --> 00:03:09.596

especially in 2024?

00:03:09.596 --> 00:03:11.918

What does good learning look like?

00:03:12.437 --> 00:03:13.558

What does good learning look like?

00:03:13.598 --> 00:03:13.919

Well,

00:03:14.580 --> 00:03:17.943

engagement is on top of mind for most

00:03:17.983 --> 00:03:19.525

of the educators I talk to

00:03:19.625 --> 00:03:21.066

and most of the school district folks.

00:03:21.186 --> 00:03:23.147

How do we actually get young

00:03:23.228 --> 00:03:25.931

people to be and remain

00:03:26.050 --> 00:03:27.491

engaged in the learning

00:03:27.532 --> 00:03:28.652

that happens in their

00:03:28.712 --> 00:03:29.995

formal schooling environments,

00:03:30.474 --> 00:03:32.116

especially after the years

00:03:32.257 --> 00:03:33.318

of disengagement?

00:03:34.199 --> 00:03:36.159

And what I would argue is

00:03:36.580 --> 00:03:38.081

couple that with all of the

00:03:38.141 --> 00:03:39.364

engagement that they get in

00:03:39.503 --> 00:03:40.884

other kinds of social media

00:03:41.324 --> 00:03:42.586

and other kinds of social

00:03:42.667 --> 00:03:43.527

learning spaces.

00:03:44.935 --> 00:03:47.679

Young people need engaging learning.

00:03:47.758 --> 00:03:48.520

So that's the first

00:03:48.560 --> 00:03:49.681

definition of good learning.

00:03:50.163 --> 00:03:51.223

It's engaging learning.

00:03:51.425 --> 00:03:52.967

It meets learners where they

00:03:53.187 --> 00:03:55.349

are and then helps to

00:03:55.670 --> 00:03:57.293

leverage what they already

00:03:57.353 --> 00:03:59.276

do know and get them to

00:03:59.395 --> 00:04:00.176

where they need to be.

00:04:01.344 --> 00:04:02.467

I love that definition.

00:04:03.207 --> 00:04:04.650

Just a few hours before we

00:04:04.669 --> 00:04:05.450

did this recording,

00:04:05.570 --> 00:04:06.953

I brought home a Google

00:04:06.973 --> 00:04:08.735

Sheets project that I'm

00:04:08.756 --> 00:04:09.516

going to be giving in my

00:04:09.556 --> 00:04:10.358

middle school soon.

00:04:11.118 --> 00:04:12.662

And my middle school kids

00:04:12.681 --> 00:04:14.463

know that before they get any assignment,

00:04:14.503 --> 00:04:16.627

it has to pass a series of three tests.

00:04:17.168 --> 00:04:18.189

And those tests, of course,

00:04:18.228 --> 00:04:18.870

are my triplets.

00:04:19.269 --> 00:04:19.730

So tonight,

00:04:19.769 --> 00:04:20.990

my triplets were doing these

00:04:21.031 --> 00:04:22.012

Google Sheets homeworks.

00:04:22.612 --> 00:04:23.413

They're in fourth grade,

00:04:23.432 --> 00:04:24.634

but they were doing the

00:04:24.673 --> 00:04:25.654

middle school level work.

00:04:26.136 --> 00:04:27.216

And just as you were saying,

00:04:27.416 --> 00:04:28.737

meet the kids where they are,

00:04:28.817 --> 00:04:30.238

give them something engaging,

00:04:30.779 --> 00:04:33.422

and just sit back and watch what it is.

00:04:33.802 --> 00:04:35.083

Adventum, of course, is in 43,000 schools,

00:04:35.103 --> 00:04:35.624

hitting 420,000 educators

00:04:35.644 --> 00:04:36.564

and 5.2 million students in

00:04:36.624 --> 00:04:36.944

all 50 states.

00:04:43.550 --> 00:04:45.211

I'm looking forward to this conversation.

00:04:45.230 --> 00:04:46.350

You want to just dive right into this?

00:04:46.891 --> 00:04:47.391

Absolutely.

00:04:47.451 --> 00:04:47.730

Let's go.

00:05:04.516 --> 00:05:05.338

Some people think that

00:05:05.377 --> 00:05:08.300

artificial intelligence is only chat GPT.

00:05:08.480 --> 00:05:09.821

We've got different terms, right?

00:05:09.860 --> 00:05:11.362

We've got generative AI.

00:05:11.442 --> 00:05:12.882

We've got design AI.

00:05:12.942 --> 00:05:15.245

We've got text-based AI.

00:05:15.305 --> 00:05:17.826

We've got AI in different

00:05:17.886 --> 00:05:19.346

products like Canva and

00:05:19.427 --> 00:05:21.369

Adobe and Microsoft and all

00:05:21.389 --> 00:05:22.350

of these different things.

00:05:22.850 --> 00:05:23.930

So I'm going to ask you,

00:05:23.951 --> 00:05:24.831

I'm going to put you on the

00:05:24.870 --> 00:05:25.591

hot seat here.

00:05:26.271 --> 00:05:27.552

Millions of educators have

00:05:27.593 --> 00:05:28.894

just stopped their cars and

00:05:28.934 --> 00:05:29.935

pulled over to the side.

00:05:30.555 --> 00:05:30.855

Jane?

00:05:31.716 --> 00:05:35.000

How do you define the term generative AI?

00:05:36.201 --> 00:05:38.625

When I'm talking about generative AI,

00:05:38.646 --> 00:05:41.670

I am talking about any of

00:05:41.790 --> 00:05:44.733

the tools that will generate

00:05:45.918 --> 00:05:48.021

new content because they are

00:05:48.182 --> 00:05:49.302

using the artificial

00:05:49.362 --> 00:05:50.524

intelligence that they have

00:05:50.543 --> 00:05:54.588

been programmed with to look for patterns,

00:05:55.028 --> 00:05:57.911

to call from whatever large

00:05:57.971 --> 00:05:59.372

language model usually,

00:05:59.574 --> 00:06:01.636

so whatever big batch of

00:06:01.716 --> 00:06:05.639

data they were given, to give the user

00:06:06.961 --> 00:06:09.502

a response or a creation,

00:06:09.523 --> 00:06:10.322

because it could be

00:06:10.442 --> 00:06:11.624

image-based if we're

00:06:11.663 --> 00:06:12.704

talking about something

00:06:12.744 --> 00:06:16.286

like DALI or mid-journey or in Canva,

00:06:16.545 --> 00:06:18.666

could be design-based, right?

00:06:19.067 --> 00:06:20.827

But it gives the user what

00:06:20.908 --> 00:06:22.548

it thinks it wants.

00:06:23.709 --> 00:06:25.370

So that's important to know

00:06:25.891 --> 00:06:28.192

that it provides what it

00:06:28.291 --> 00:06:29.612

thinks you want based on

00:06:29.653 --> 00:06:31.173

the prompt you gave it and

00:06:31.233 --> 00:06:34.815

based on its training and in its model.

00:06:35.891 --> 00:06:37.411

That sounds like my triplets.

00:06:37.451 --> 00:06:38.572

Let me see what we can get a

00:06:38.632 --> 00:06:39.291

couple of things here.

00:06:39.992 --> 00:06:44.713

If I go on to a Google search and I say,

00:06:44.733 --> 00:06:46.994

I need a recipe for cookies.

00:06:47.593 --> 00:06:49.194

Is that generative AI?

00:06:49.274 --> 00:06:50.274

I'm putting in a prompt.

00:06:50.714 --> 00:06:51.716

It's giving me something.

00:06:52.036 --> 00:06:53.156

Is that generative AI?

00:06:53.762 --> 00:06:55.463

No, that's a search, right?

00:06:55.644 --> 00:06:57.245

And artificial intelligence

00:06:57.305 --> 00:06:58.766

might be involved in the search,

00:06:59.086 --> 00:07:01.086

but that's not necessarily generative AI.

00:07:01.146 --> 00:07:02.608

We'll say like when you're

00:07:02.668 --> 00:07:03.649

starting to type,

00:07:04.168 --> 00:07:05.670

give me a recipe and you

00:07:05.730 --> 00:07:07.850

see all the stuff that pops up below it,

00:07:08.350 --> 00:07:09.732

that is generative AI

00:07:09.752 --> 00:07:11.012

because here it is using

00:07:11.052 --> 00:07:12.713

its training to make a

00:07:12.754 --> 00:07:14.014

prediction to give you what

00:07:14.055 --> 00:07:15.314

it thinks you want, right?

00:07:15.435 --> 00:07:16.536

But it's just filling in the

00:07:16.596 --> 00:07:17.456

search right there.

00:07:19.098 --> 00:07:20.459

But what I'm talking about

00:07:20.500 --> 00:07:21.240

when I talk about

00:07:21.281 --> 00:07:23.382

generative AI is when

00:07:23.401 --> 00:07:24.742

you're using a tool like

00:07:24.843 --> 00:07:26.142

Microsoft Copilot,

00:07:26.262 --> 00:07:27.843

who any of the schools who

00:07:28.165 --> 00:07:29.785

are on Microsoft tools,

00:07:30.245 --> 00:07:32.307

they likely now have access,

00:07:32.466 --> 00:07:34.108

whether they use it or ignore it,

00:07:34.747 --> 00:07:36.428

to Copilot,

00:07:36.749 --> 00:07:39.091

to have a chat feature that

00:07:39.211 --> 00:07:41.291

they can put in and ask a question.

00:07:41.312 --> 00:07:44.233

And when you ask that program a question,

00:07:45.293 --> 00:07:46.595

Unlike a Google search,

00:07:46.975 --> 00:07:48.714

you'll get a different sort

00:07:48.754 --> 00:07:50.216

of generative answer.

00:07:50.315 --> 00:07:52.276

You'll get a text-based answer,

00:07:52.716 --> 00:07:53.956

often with different sources.

00:07:54.377 --> 00:07:55.798

What Google gives you is a

00:07:55.978 --> 00:07:58.798

list of what it thinks are

00:07:58.879 --> 00:08:01.199

your most likely or best

00:08:01.319 --> 00:08:04.661

paid for choices to answer your question.

00:08:04.841 --> 00:08:06.242

And you then have to go out

00:08:06.302 --> 00:08:07.343

and look at the site that

00:08:07.382 --> 00:08:08.062

it links to you.

00:08:08.742 --> 00:08:10.163

What the generative AI tool

00:08:10.204 --> 00:08:11.324

will do with your question

00:08:11.464 --> 00:08:13.706

is it will create text that

00:08:13.747 --> 00:08:15.228

it thinks answers your question,

00:08:15.307 --> 00:08:17.050

pulling and synthesizing

00:08:17.449 --> 00:08:18.771

from a variety of sources.

00:08:19.552 --> 00:08:19.891

And, you know,

00:08:19.932 --> 00:08:21.052

while you're giving me that

00:08:21.132 --> 00:08:22.533

amazing answer, of course,

00:08:22.593 --> 00:08:24.935

I go on to Microsoft Copilot and I say,

00:08:25.615 --> 00:08:26.896

start a knock knock joke.

00:08:26.997 --> 00:08:28.358

And of course, it says knock knock.

00:08:29.175 --> 00:08:29.754

Who's there?

00:08:30.375 --> 00:08:30.834

Banana.

00:08:31.295 --> 00:08:32.495

So this is where artificial

00:08:32.556 --> 00:08:33.576

intelligence is, right?

00:08:34.775 --> 00:08:36.817

If I go into these different programs,

00:08:36.856 --> 00:08:38.256

we know that there's, as you mentioned,

00:08:38.317 --> 00:08:40.697

a variety of different kinds.

00:08:40.977 --> 00:08:43.577

I think the two biggies that are out there,

00:08:44.118 --> 00:08:47.698

ChatGPT and Microsoft Copilot.

00:08:48.019 --> 00:08:48.798

And I would even throw a

00:08:48.879 --> 00:08:49.600

third one in there.

00:08:49.879 --> 00:08:52.100

Google has their BARD slash,

00:08:52.220 --> 00:08:53.841

they're not calling it Gemini, right?

00:08:54.160 --> 00:08:54.400

Right.

00:08:54.640 --> 00:08:54.780

And...

00:08:56.380 --> 00:08:58.542

The scary part is these

00:08:58.642 --> 00:09:00.043

things are now starting to

00:09:00.082 --> 00:09:02.043

be turned on at the admin

00:09:02.063 --> 00:09:04.384

level for school districts.

00:09:05.445 --> 00:09:07.686

This is not the AI world and

00:09:07.765 --> 00:09:09.826

then the school world.

00:09:09.947 --> 00:09:11.447

Microsoft is now every

00:09:11.508 --> 00:09:13.288

single day putting out videos going,

00:09:13.749 --> 00:09:16.049

here is Copilot with PowerPoint.

00:09:16.090 --> 00:09:17.630

Here is Copilot with Outlook.

00:09:18.270 --> 00:09:19.792

I got to be transparent.

00:09:19.892 --> 00:09:20.971

I'm personally one of those

00:09:21.072 --> 00:09:22.173

ones that are paying 30

00:09:22.173 --> 00:09:23.833

bucks a month for Copilot.

00:09:24.614 --> 00:09:25.575

I love it.

00:09:26.075 --> 00:09:27.677

I love the fact that I can sit,

00:09:27.736 --> 00:09:28.397

if nothing else,

00:09:28.397 --> 00:09:30.359

30 bucks a month is paying

00:09:30.438 --> 00:09:31.899

for me to look at a strand

00:09:31.919 --> 00:09:34.201

of emails and have it read

00:09:34.241 --> 00:09:35.683

the emails and give me like

00:09:35.724 --> 00:09:38.326

a three sentence synopsis

00:09:38.745 --> 00:09:40.687

of what the entire email thread is.

00:09:40.707 --> 00:09:43.049

I absolutely love that.

00:09:43.682 --> 00:09:45.144

Yeah, it's a time saver.

00:09:45.325 --> 00:09:46.525

Huge time saver.

00:09:46.947 --> 00:09:48.628

I'm still trying to figure things out.

00:09:48.729 --> 00:09:49.789

Last night I was doing a

00:09:49.870 --> 00:09:50.850

chat with a friend who was

00:09:50.890 --> 00:09:52.873

at a Rangers game and I said, you know,

00:09:53.153 --> 00:09:54.674

please take this picture

00:09:54.855 --> 00:09:55.956

and put it in a Rangers

00:09:55.996 --> 00:09:57.437

jersey and put the Stanley Cup.

00:09:57.999 --> 00:09:58.538

I was doing the

00:09:58.578 --> 00:10:01.342

designer.microsoft.com thing and

00:10:02.503 --> 00:10:04.024

We were just having fun with it, right?

00:10:04.283 --> 00:10:04.504

Right.

00:10:05.284 --> 00:10:08.086

And let me stop you there, Jeff,

00:10:08.187 --> 00:10:10.028

because what you're doing

00:10:10.769 --> 00:10:12.769

is exactly what I'm trying

00:10:12.809 --> 00:10:15.751

to advocate for to school leaders.

00:10:16.091 --> 00:10:17.173

You're playing with it.

00:10:17.293 --> 00:10:18.673

You're getting your hands in there.

00:10:18.714 --> 00:10:20.816

You're trying different use cases.

00:10:21.115 --> 00:10:22.096

The use case may be

00:10:22.177 --> 00:10:23.177

entertaining your friend.

00:10:23.496 --> 00:10:25.298

The use case may be digging

00:10:25.318 --> 00:10:25.759

through your emails.

00:10:25.798 --> 00:10:27.559

email and saving yourself time.

00:10:27.899 --> 00:10:29.341

The use case may be for our

00:10:29.380 --> 00:10:30.601

teacher friends listening,

00:10:31.123 --> 00:10:32.503

designing a lesson plan or

00:10:32.563 --> 00:10:35.065

giving a student sample to

00:10:35.264 --> 00:10:36.865

meet the needs of their students, right?

00:10:37.226 --> 00:10:39.447

You've spent time to play

00:10:39.687 --> 00:10:41.590

and figure out where it

00:10:41.750 --> 00:10:43.110

could be useful for you.

00:10:43.951 --> 00:10:45.091

And that's what we're

00:10:45.211 --> 00:10:46.751

advocating for our

00:10:47.251 --> 00:10:48.491

education partners to do.

00:10:49.072 --> 00:10:50.293

So at Edmentum,

00:10:50.552 --> 00:10:52.052

we ran a series of

00:10:52.113 --> 00:10:54.092

experiments to try to

00:10:54.253 --> 00:10:56.594

figure out how would we want to advise,

00:10:56.734 --> 00:10:57.833

especially last summer,

00:10:58.014 --> 00:10:59.414

everybody was talking about it.

00:10:59.855 --> 00:11:01.355

School districts had shut it down.

00:11:01.735 --> 00:11:02.855

We're trying to figure out

00:11:03.014 --> 00:11:05.076

what we could suggest to

00:11:05.296 --> 00:11:07.096

to our partners and so we

00:11:07.136 --> 00:11:08.136

got in there and started

00:11:08.197 --> 00:11:09.658

running experiments and

00:11:09.738 --> 00:11:11.077

that's the kind of thing

00:11:11.118 --> 00:11:12.457

that we learned is that

00:11:12.859 --> 00:11:14.379

teachers need or school

00:11:14.399 --> 00:11:15.899

district leaders need to

00:11:16.200 --> 00:11:17.360

get their hands in it try

00:11:17.399 --> 00:11:18.500

different tools see how

00:11:18.541 --> 00:11:19.961

they work so they can

00:11:20.041 --> 00:11:21.522

figure out where it might

00:11:21.562 --> 00:11:22.381

be useful for them

00:11:23.673 --> 00:11:24.615

I'd like to have this

00:11:24.695 --> 00:11:26.034

conversation from a couple

00:11:26.134 --> 00:11:27.075

different chairs.

00:11:27.696 --> 00:11:28.395

I'll try to tell you the

00:11:28.436 --> 00:11:29.596

chair I'm doing the question from.

00:11:29.635 --> 00:11:29.936

Right now,

00:11:29.956 --> 00:11:30.956

I want to do this from the tech

00:11:30.976 --> 00:11:31.636

director chair.

00:11:32.297 --> 00:11:33.238

When I'm working with a

00:11:33.278 --> 00:11:34.918

company and they say they

00:11:34.977 --> 00:11:37.339

now are using artificial intelligence,

00:11:38.178 --> 00:11:39.340

I know as a tech director,

00:11:40.220 --> 00:11:41.740

I need to have a privacy

00:11:41.780 --> 00:11:43.042

agreement signed with that

00:11:43.121 --> 00:11:45.283

company for my users to log in?

00:11:45.844 --> 00:11:47.403

Do I also need to ask

00:11:47.504 --> 00:11:49.725

questions such as where is

00:11:49.806 --> 00:11:51.767

that company getting their

00:11:51.826 --> 00:11:53.107

artificial intelligence

00:11:53.368 --> 00:11:54.769

originally sourced from?

00:11:55.129 --> 00:11:57.129

And do I need to worry about

00:11:57.190 --> 00:11:58.490

having a privacy agreement

00:11:58.530 --> 00:11:59.511

with that source?

00:12:01.942 --> 00:12:03.826

I think what you're hitting

00:12:03.926 --> 00:12:05.948

on with that question there, Jeff,

00:12:06.009 --> 00:12:08.172

that is on the minds of

00:12:08.272 --> 00:12:10.894

every tech director and the

00:12:10.955 --> 00:12:12.918

legal folks in districts, right,

00:12:13.359 --> 00:12:15.782

is how do the data that get

00:12:15.902 --> 00:12:17.945

put into an AI get used?

00:12:18.945 --> 00:12:21.067

So one of the benefits of

00:12:21.226 --> 00:12:23.769

using a Microsoft co-pilot, for example,

00:12:24.370 --> 00:12:26.231

is the way that it's

00:12:26.312 --> 00:12:28.693

attached to any enterprise

00:12:28.833 --> 00:12:30.975

is it protects the privacy of the data.

00:12:31.116 --> 00:12:33.538

That data doesn't get fed into the model.

00:12:33.918 --> 00:12:34.899

But the important thing for

00:12:34.960 --> 00:12:37.221

teachers to know, if, for example,

00:12:37.282 --> 00:12:39.303

the only generative AI that

00:12:39.344 --> 00:12:40.745

they think of is chat GPT,

00:12:42.427 --> 00:12:43.307

What they need to know is

00:12:43.347 --> 00:12:44.889

that ChatGPT will take

00:12:45.110 --> 00:12:46.392

anything that you input

00:12:46.511 --> 00:12:48.433

into it and it starts to

00:12:48.614 --> 00:12:49.975

use it to train the model.

00:12:50.755 --> 00:12:52.378

So the question from a tech

00:12:52.418 --> 00:12:54.921

director seat is probably yes.

00:12:55.100 --> 00:12:56.962

They need to figure out where...

00:12:58.432 --> 00:13:01.874

what a company is using and which data,

00:13:02.033 --> 00:13:03.615

like whether or not the data gets shared,

00:13:04.176 --> 00:13:05.235

you're safer if they're

00:13:05.275 --> 00:13:06.756

using Microsoft Copilot.

00:13:07.157 --> 00:13:10.019

And there's also almost always,

00:13:10.178 --> 00:13:10.860

as I've seen it,

00:13:11.000 --> 00:13:12.620

data sharing agreements or

00:13:12.780 --> 00:13:15.442

not that protect the privacy.

00:13:15.522 --> 00:13:17.302

So even as an ed tech company,

00:13:18.364 --> 00:13:20.004

all of the same rules and

00:13:20.044 --> 00:13:22.986

regulations for protecting

00:13:23.086 --> 00:13:25.207

student data apply to us as

00:13:25.268 --> 00:13:26.509

they do to a school district.

00:13:26.849 --> 00:13:28.953

So we can't use and share

00:13:29.094 --> 00:13:31.980

and email and feed into a

00:13:32.201 --> 00:13:34.004

system any student data.

00:13:35.304 --> 00:13:36.586

when a tech director is

00:13:36.686 --> 00:13:38.427

looking at an application

00:13:38.927 --> 00:13:40.488

or when an application

00:13:40.548 --> 00:13:42.288

comes to a tech director that says, hey,

00:13:42.349 --> 00:13:44.150

now we have this extra thing on us,

00:13:44.971 --> 00:13:46.792

what questions should a

00:13:47.513 --> 00:13:49.134

tech director be asking of

00:13:49.173 --> 00:13:50.894

their ed tech partners when

00:13:50.934 --> 00:13:53.996

it comes to the topic of AI, AI features,

00:13:54.177 --> 00:13:56.258

perhaps can I turn the AI

00:13:56.317 --> 00:13:57.599

features on and off on my

00:13:57.678 --> 00:13:59.019

side or are they now just a

00:13:59.100 --> 00:14:00.039

part of this world?

00:14:00.400 --> 00:14:01.561

What questions should school

00:14:01.581 --> 00:14:02.461

districts be asking?

00:14:03.336 --> 00:14:03.436

Well,

00:14:03.475 --> 00:14:05.017

I think lots of people are asking

00:14:05.216 --> 00:14:08.698

questions around age restrictions.

00:14:09.080 --> 00:14:10.559

So those keep changing.

00:14:10.580 --> 00:14:13.162

I would also,

00:14:14.023 --> 00:14:16.183

speaking of the idea of changing,

00:14:16.724 --> 00:14:19.405

this landscape and these

00:14:19.466 --> 00:14:21.726

technologies are ever changing, right?

00:14:22.128 --> 00:14:24.009

All of the models keep getting updated.

00:14:24.489 --> 00:14:27.671

So I might want to ask if I

00:14:27.691 --> 00:14:28.672

were a tech director,

00:14:29.884 --> 00:14:31.504

How will I be notified of

00:14:31.585 --> 00:14:32.965

changes to the model?

00:14:34.706 --> 00:14:37.147

I think it is a good question to ask.

00:14:37.346 --> 00:14:39.729

Is there a way to limit access,

00:14:40.369 --> 00:14:42.149

turn features on and off?

00:14:44.585 --> 00:14:45.644

The other thing to note

00:14:45.965 --> 00:14:47.566

about the perspective that

00:14:47.785 --> 00:14:50.626

I bring from Edmentum is

00:14:50.706 --> 00:14:53.488

that we're not putting AI

00:14:53.687 --> 00:14:56.068

into our products at this point.

00:14:56.828 --> 00:14:58.528

We've taken a more kind of

00:14:58.609 --> 00:14:59.789

cautious approach.

00:15:00.549 --> 00:15:02.770

We're absolutely looking at

00:15:02.990 --> 00:15:04.230

use cases for our own

00:15:04.291 --> 00:15:06.331

efficiencies and the work

00:15:06.351 --> 00:15:08.033

that we need to do to create things.

00:15:08.572 --> 00:15:11.453

But when it comes to in our products,

00:15:12.173 --> 00:15:12.494

No.

00:15:12.974 --> 00:15:13.414

Rather,

00:15:13.634 --> 00:15:15.075

what we're doing is trying to

00:15:15.154 --> 00:15:16.576

figure out how to help

00:15:16.855 --> 00:15:18.657

teachers who use our

00:15:18.697 --> 00:15:21.538

products think about when

00:15:21.577 --> 00:15:23.558

and if or how students

00:15:23.698 --> 00:15:26.900

might use AI to complete assignments,

00:15:27.380 --> 00:15:28.160

what they should worry

00:15:28.221 --> 00:15:30.322

about or not when it comes to that,

00:15:30.662 --> 00:15:31.462

and how to have the

00:15:31.543 --> 00:15:32.582

teachers find their own

00:15:32.623 --> 00:15:34.384

efficiencies with AI in

00:15:34.443 --> 00:15:36.184

terms of using our products.

00:15:37.072 --> 00:15:37.972

I wanna throw one more

00:15:38.033 --> 00:15:38.793

question at you from the

00:15:38.833 --> 00:15:39.634

tech director chair,

00:15:39.673 --> 00:15:40.375

and this is a biggie.

00:15:40.894 --> 00:15:43.076

And there are spreadsheets

00:15:43.177 --> 00:15:44.278

running around the internet

00:15:44.317 --> 00:15:46.679

right now that have all of

00:15:46.720 --> 00:15:47.660

this information,

00:15:48.501 --> 00:15:49.322

but I think there's a lot

00:15:49.363 --> 00:15:50.403

that's premature right now.

00:15:51.556 --> 00:15:53.136

Do you have any recommendations?

00:15:53.197 --> 00:15:54.778

I know you're not legal, right?

00:15:55.158 --> 00:15:56.100

But do you have any

00:15:56.179 --> 00:15:57.900

recommendations on language

00:15:58.241 --> 00:16:00.864

that should be in or things

00:16:00.923 --> 00:16:02.625

that should be in some kind

00:16:02.664 --> 00:16:05.668

of an official board doc AI policy?

00:16:06.248 --> 00:16:07.830

I know school districts are jumping in,

00:16:07.870 --> 00:16:09.311

but they don't have an AI policy.

00:16:09.350 --> 00:16:10.432

Some school districts are saying,

00:16:10.831 --> 00:16:11.773

why do I need one?

00:16:11.793 --> 00:16:12.572

And then there's some school

00:16:12.592 --> 00:16:14.195

districts that are making

00:16:14.235 --> 00:16:16.256

the document that everything is in there.

00:16:16.876 --> 00:16:17.236

Do you have any

00:16:17.256 --> 00:16:18.437

recommendations or do you

00:16:18.457 --> 00:16:19.898

have a chance to see what

00:16:19.918 --> 00:16:20.798

other school districts and

00:16:20.818 --> 00:16:21.418

stuff are doing?

00:16:22.359 --> 00:16:24.299

So I've had a couple of

00:16:24.379 --> 00:16:25.620

opportunities to see what

00:16:25.660 --> 00:16:27.081

other school districts are doing.

00:16:28.802 --> 00:16:29.662

One of the things I do on

00:16:29.701 --> 00:16:31.163

the side is I still I

00:16:31.202 --> 00:16:32.663

couldn't leave academia altogether.

00:16:32.722 --> 00:16:34.043

So I still teach an

00:16:34.124 --> 00:16:35.244

instructional technology

00:16:35.303 --> 00:16:37.024

course at the University of Pennsylvania.

00:16:37.345 --> 00:16:38.426

And I taught it last fall.

00:16:38.905 --> 00:16:40.046

And the course gets taught

00:16:40.206 --> 00:16:41.346

to school leaders.

00:16:41.586 --> 00:16:43.268

And so I had a cohort of 20

00:16:43.268 --> 00:16:44.469

something school leaders.

00:16:44.808 --> 00:16:45.568

And of course,

00:16:45.708 --> 00:16:48.090

in an instructional technology module,

00:16:48.190 --> 00:16:48.890

we were talking about

00:16:48.931 --> 00:16:50.831

generative AI and policy.

00:16:50.871 --> 00:16:52.493

So I got to see some of the

00:16:52.533 --> 00:16:57.595

policies that those folks as my students,

00:16:57.674 --> 00:16:59.155

but in their day jobs were

00:16:59.235 --> 00:17:00.356

creating in their districts.

00:17:00.736 --> 00:17:02.258

I've also been following

00:17:02.418 --> 00:17:03.818

what New York City schools

00:17:04.019 --> 00:17:05.559

have been doing, right?

00:17:05.680 --> 00:17:07.020

They were one of the first

00:17:07.201 --> 00:17:09.242

school districts to ban

00:17:09.883 --> 00:17:11.763

chat GPT when it first came out,

00:17:11.825 --> 00:17:12.625

when everybody was trying

00:17:12.644 --> 00:17:13.425

to figure things out.

00:17:13.865 --> 00:17:15.366

And now we see that they

00:17:15.406 --> 00:17:16.488

have come around and

00:17:17.048 --> 00:17:18.689

created a more thoughtful approach.

00:17:18.729 --> 00:17:20.650

They've got a group working

00:17:20.769 --> 00:17:22.151

on it and they're trying to

00:17:22.270 --> 00:17:23.330

make things public.

00:17:23.371 --> 00:17:24.571

So I think if I were your

00:17:24.612 --> 00:17:26.873

tech director and I was in that chair,

00:17:27.373 --> 00:17:28.634

what I would do is I would

00:17:28.653 --> 00:17:30.334

probably go look to some of

00:17:30.354 --> 00:17:31.454

the bigger districts who

00:17:31.515 --> 00:17:33.016

have the resources and the

00:17:33.056 --> 00:17:34.276

money and the manpower to

00:17:34.316 --> 00:17:36.297

be thinking about this more deeply.

00:17:36.798 --> 00:17:37.938

And I would look to see what

00:17:37.978 --> 00:17:40.598

their current policy is and

00:17:40.679 --> 00:17:41.940

see what language might

00:17:42.039 --> 00:17:43.740

need to be included in my

00:17:43.820 --> 00:17:45.662

own district's policy.

00:17:47.045 --> 00:17:48.486

So let's take that hat off

00:17:48.526 --> 00:17:49.788

for a second here and let's

00:17:49.827 --> 00:17:51.288

put on the coaching hat or

00:17:51.308 --> 00:17:52.430

the curricular hat, right?

00:17:52.911 --> 00:17:53.991

One of the questions and

00:17:54.672 --> 00:17:55.613

topics that have come up on

00:17:55.653 --> 00:17:57.234

our Ask the Tech Coach show has been,

00:17:57.734 --> 00:17:59.856

how do you introduce this

00:17:59.978 --> 00:18:01.939

concept to teachers, right?

00:18:03.019 --> 00:18:04.721

We think of this as the calculator, right?

00:18:04.801 --> 00:18:06.584

Teachers are saying you can't use it,

00:18:06.644 --> 00:18:07.924

you can't use it, you can't use it,

00:18:07.964 --> 00:18:09.446

but now everybody has a calculator.

00:18:11.993 --> 00:18:13.355

There's so many coaches out

00:18:13.395 --> 00:18:14.938

there right now that are

00:18:15.398 --> 00:18:16.461

jumping in and saying,

00:18:16.500 --> 00:18:17.782

can I have 20 minutes at a

00:18:17.823 --> 00:18:19.664

faculty meeting just to put

00:18:19.704 --> 00:18:20.666

that first toe in,

00:18:21.126 --> 00:18:23.049

just to have that conversation?

00:18:24.031 --> 00:18:25.814

Even myself as a technology teacher,

00:18:25.974 --> 00:18:26.976

I want to try this.

00:18:27.676 --> 00:18:29.739

But I don't want to teach my

00:18:29.838 --> 00:18:30.599

kids something.

00:18:30.640 --> 00:18:31.601

I feel weird saying this.

00:18:31.862 --> 00:18:32.642

I don't want to teach my

00:18:32.682 --> 00:18:34.463

kids something that my

00:18:34.523 --> 00:18:35.826

colleagues are going to be

00:18:35.986 --> 00:18:37.968

uncomfortable with them knowing.

00:18:38.509 --> 00:18:38.689

Right.

00:18:39.169 --> 00:18:39.430

Right?

00:18:39.809 --> 00:18:41.491

So all of that being said,

00:18:41.511 --> 00:18:43.193

if you were somebody who

00:18:43.213 --> 00:18:43.894

was in charge of

00:18:43.934 --> 00:18:45.016

professional development...

00:18:46.635 --> 00:18:46.955

How do,

00:18:47.056 --> 00:18:48.457

and this is gonna be a two-part question.

00:18:48.537 --> 00:18:50.679

How do you start the conversation?

00:18:51.138 --> 00:18:53.140

What's an application that you would use?

00:18:53.460 --> 00:18:54.500

Do you have an example of

00:18:54.540 --> 00:18:56.882

maybe a first group assignment?

00:18:57.643 --> 00:18:59.584

What's that 30 second pitch

00:18:59.624 --> 00:19:01.224

or speech or anything that

00:19:01.265 --> 00:19:02.066

you would do if you were

00:19:02.086 --> 00:19:03.586

that coach and you were

00:19:03.626 --> 00:19:04.987

given a faculty meeting and said,

00:19:05.567 --> 00:19:06.567

introduce the topic,

00:19:06.587 --> 00:19:07.828

but don't go too far in the water.

00:19:08.308 --> 00:19:09.569

Right.

00:19:09.589 --> 00:19:10.510

I love this question.

00:19:10.691 --> 00:19:11.411

So this,

00:19:11.510 --> 00:19:12.771

I have actually done a bunch of

00:19:12.791 --> 00:19:14.933

thinking about is how to get it started.

00:19:16.144 --> 00:19:17.785

I think I would take a

00:19:18.305 --> 00:19:20.625

problem that is broad for

00:19:20.826 --> 00:19:21.905

most of my colleagues.

00:19:22.006 --> 00:19:23.185

And I would venture a guess

00:19:23.226 --> 00:19:24.386

that most of your listeners

00:19:25.126 --> 00:19:25.987

are dealing with the

00:19:26.027 --> 00:19:28.127

challenge of the various

00:19:28.268 --> 00:19:30.229

languages that our students

00:19:30.288 --> 00:19:31.308

come to our classrooms

00:19:31.328 --> 00:19:33.890

speaking and their families, right?

00:19:34.009 --> 00:19:36.290

We have a huge variety of

00:19:36.371 --> 00:19:38.771

multilingual learners who

00:19:39.051 --> 00:19:41.633

are trying to learn our content,

00:19:41.833 --> 00:19:43.272

but they still don't know

00:19:43.333 --> 00:19:44.252

the English language that

00:19:44.272 --> 00:19:45.114

we're speaking to them in.

00:19:45.973 --> 00:19:46.894

So one of the things that I

00:19:46.954 --> 00:19:48.655

might show my colleagues if

00:19:48.756 --> 00:19:50.477

I were a coach is I'd show

00:19:50.517 --> 00:19:52.978

them we found Clawed AI,

00:19:53.038 --> 00:19:54.078

which is one that we have

00:19:54.118 --> 00:19:55.019

not yet mentioned.

00:19:56.160 --> 00:19:59.142

But Clawed AI was the tool

00:19:59.182 --> 00:20:00.563

that we found at the time

00:20:00.603 --> 00:20:01.844

when we ran our experiments

00:20:01.884 --> 00:20:03.444

a few months ago was the

00:20:03.484 --> 00:20:06.686

best at taking a prompt

00:20:07.809 --> 00:20:09.172

and you put it in there and

00:20:09.211 --> 00:20:11.114

you ask it to translate

00:20:11.173 --> 00:20:12.935

that prompt and explain the

00:20:12.976 --> 00:20:16.401

concept to a speaker of say, for example,

00:20:16.480 --> 00:20:17.521

Moroccan Arabic.

00:20:18.470 --> 00:20:20.611

So what Cloud AI does,

00:20:20.631 --> 00:20:22.512

it's better than Google Translate,

00:20:22.752 --> 00:20:24.034

which just gives you a

00:20:24.094 --> 00:20:25.634

one-to-one translation and

00:20:25.855 --> 00:20:27.155

who knows how good it is.

00:20:27.816 --> 00:20:29.777

But what Cloud AI does is it

00:20:29.817 --> 00:20:31.137

will give you the translation,

00:20:31.459 --> 00:20:33.700

explain in both English and

00:20:33.779 --> 00:20:35.540

in the language of choice,

00:20:35.721 --> 00:20:36.541

the target language,

00:20:37.122 --> 00:20:39.003

why it made choices that it did.

00:20:40.625 --> 00:20:42.665

to explain the concept and

00:20:42.705 --> 00:20:43.866

to make it more accessible.

00:20:44.346 --> 00:20:45.807

So if you've got a teacher

00:20:46.127 --> 00:20:47.588

who's got students who are

00:20:47.628 --> 00:20:49.349

speaking maybe a handful of

00:20:49.410 --> 00:20:50.631

languages in their class

00:20:50.671 --> 00:20:51.471

and they're just trying to

00:20:51.511 --> 00:20:53.692

teach them math and you're

00:20:53.732 --> 00:20:55.354

trying to explain to them

00:20:55.394 --> 00:20:56.815

the Pythagorean theorem and

00:20:56.894 --> 00:21:00.198

how that works and you need

00:21:00.837 --> 00:21:02.219

your multilingual learners

00:21:02.239 --> 00:21:02.798

to understand it,

00:21:02.919 --> 00:21:04.200

I might show them how that

00:21:04.259 --> 00:21:06.221

works and how easy that is

00:21:07.162 --> 00:21:09.963

to give them the explanation

00:21:10.044 --> 00:21:11.644

that will allow them to differentiate.

00:21:12.226 --> 00:21:13.646

The next thing I'd do is I'd say,

00:21:14.126 --> 00:21:16.147

so who's interested in learning more?

00:21:17.388 --> 00:21:18.410

And I think professional

00:21:18.450 --> 00:21:19.810

development in this area

00:21:20.310 --> 00:21:21.311

needs to start with a

00:21:21.412 --> 00:21:22.492

coalition of the willing.

00:21:24.061 --> 00:21:25.461

So bring together the

00:21:25.501 --> 00:21:28.303

teachers who aren't fully afraid of it,

00:21:28.442 --> 00:21:30.183

who want to dip their toes in the water.

00:21:30.923 --> 00:21:33.144

And what we advocated when

00:21:33.203 --> 00:21:34.664

we wrote about this last year,

00:21:36.365 --> 00:21:38.145

we advocated for bringing

00:21:38.185 --> 00:21:39.246

this group together and

00:21:39.346 --> 00:21:41.605

creating a culture of experimentation.

00:21:41.987 --> 00:21:43.227

So getting the school to

00:21:43.326 --> 00:21:45.166

give them some space, some time,

00:21:45.227 --> 00:21:46.127

maybe some professional

00:21:46.167 --> 00:21:47.008

development hours.

00:21:47.768 --> 00:21:50.151

to start running their own experiments,

00:21:50.211 --> 00:21:51.352

to start using these

00:21:51.412 --> 00:21:53.194

different tools to see what works.

00:21:53.474 --> 00:21:53.955

There's also,

00:21:54.016 --> 00:21:55.217

we haven't talked yet about

00:21:55.297 --> 00:21:56.598

all of the generative AI

00:21:56.679 --> 00:21:59.001

that are school focused, right?

00:21:59.082 --> 00:22:00.522

That are not these other ones.

00:22:00.563 --> 00:22:02.746

So like school AI, magic school,

00:22:03.626 --> 00:22:05.107

These ones that essentially

00:22:05.188 --> 00:22:07.469

take a chat GPT engine,

00:22:07.888 --> 00:22:08.990

put it in a wrapper and

00:22:09.049 --> 00:22:10.609

start to program it and

00:22:10.650 --> 00:22:11.911

give it a personality and a

00:22:11.951 --> 00:22:13.392

persona that meets

00:22:13.511 --> 00:22:15.132

different grade levels or

00:22:15.211 --> 00:22:17.192

subject areas and starts to

00:22:17.232 --> 00:22:19.534

do some of the design work for teachers.

00:22:19.595 --> 00:22:20.855

So it lessens the load,

00:22:20.894 --> 00:22:22.496

the burden about designing

00:22:22.516 --> 00:22:23.276

your own prompts.

00:22:24.057 --> 00:22:25.616

And just have these folks

00:22:25.676 --> 00:22:27.397

experiment and learn about

00:22:27.458 --> 00:22:28.659

the different cases that it

00:22:28.739 --> 00:22:30.199

might work and then let it

00:22:30.299 --> 00:22:31.059

start to spread.

00:22:32.067 --> 00:22:32.807

When you're looking,

00:22:33.407 --> 00:22:34.307

I'm gonna go back a hat.

00:22:34.887 --> 00:22:36.588

When you're looking at, you know,

00:22:36.729 --> 00:22:38.648

here's Claude, here's Magic School.

00:22:41.009 --> 00:22:42.730

You're suggesting that this

00:22:42.789 --> 00:22:44.069

be at the teacher level,

00:22:44.549 --> 00:22:45.589

which to the best of my

00:22:45.650 --> 00:22:46.549

knowledge means I don't

00:22:46.690 --> 00:22:48.651

need to worry about privacy agreements,

00:22:49.010 --> 00:22:50.451

or is this where you go to

00:22:50.490 --> 00:22:51.570

your tech director as the

00:22:51.590 --> 00:22:52.471

tech coach and say,

00:22:53.291 --> 00:22:54.951

I'd like to try this Claude thing,

00:22:56.152 --> 00:22:56.951

go get an agreement.

00:22:57.451 --> 00:22:59.593

So I can now go do my 30

00:22:59.593 --> 00:23:01.333

minute faculty meeting.

00:23:03.666 --> 00:23:05.769

what's the legalities on that?

00:23:06.068 --> 00:23:06.970

At what point does school

00:23:06.990 --> 00:23:07.631

districts need to be

00:23:07.691 --> 00:23:08.551

reaching out to all these

00:23:08.592 --> 00:23:09.392

different companies?

00:23:10.492 --> 00:23:11.233

There's no students.

00:23:11.674 --> 00:23:14.237

You have not yet said student logs into,

00:23:14.436 --> 00:23:14.636

right?

00:23:14.656 --> 00:23:14.917

Right.

00:23:15.417 --> 00:23:16.278

But you're still asking

00:23:17.038 --> 00:23:18.580

teachers to log into that.

00:23:18.820 --> 00:23:19.701

And that's kind of where I

00:23:19.781 --> 00:23:21.523

am right now is I'd love to

00:23:21.544 --> 00:23:22.744

start trying these things,

00:23:23.224 --> 00:23:24.046

but I don't want to be

00:23:24.125 --> 00:23:25.446

crossing the district line

00:23:25.487 --> 00:23:27.048

that I might not know exists.

00:23:27.790 --> 00:23:29.374

As the director of learning design,

00:23:29.674 --> 00:23:30.215

thankfully,

00:23:30.477 --> 00:23:33.182

the legal aspect of it is not my purview.

00:23:33.321 --> 00:23:34.785

So I am not the best person

00:23:34.825 --> 00:23:35.707

to answer that question.

00:23:37.175 --> 00:23:39.277

fair okay so you're the tech

00:23:39.297 --> 00:23:40.498

director so you're the tech

00:23:40.518 --> 00:23:42.818

coach and I love the idea

00:23:43.200 --> 00:23:44.440

let's have a conversation

00:23:44.480 --> 00:23:45.980

with a problem the problem

00:23:46.040 --> 00:23:47.082

is I've got students that I

00:23:47.142 --> 00:23:47.741

need to be able to

00:23:47.781 --> 00:23:49.063

communicate with here's how

00:23:49.103 --> 00:23:50.864

this works if anybody else

00:23:50.923 --> 00:23:51.924

wants more and we talk a

00:23:51.964 --> 00:23:53.345

lot on this show about the

00:23:53.384 --> 00:23:54.445

innovation curve where once

00:23:54.465 --> 00:23:55.866

you get to that 13 or so

00:23:55.906 --> 00:23:57.607

percent now you got your

00:23:57.667 --> 00:23:59.148

first followers right how

00:23:59.189 --> 00:24:00.429

do you get to that next 23

00:24:00.429 --> 00:24:03.010

to get your we talk about

00:24:03.050 --> 00:24:06.192

that one a lot on here excellent what

00:24:07.873 --> 00:24:11.895

other ideas do you have for

00:24:12.477 --> 00:24:14.057

bringing these topics in by

00:24:14.097 --> 00:24:15.419

the way and I love coming

00:24:15.618 --> 00:24:16.359

coming from a school

00:24:16.380 --> 00:24:17.840

district that supported 75

00:24:17.840 --> 00:24:19.863

languages and being the guy

00:24:19.923 --> 00:24:21.163

that brought in things like

00:24:21.523 --> 00:24:22.545

powerpoint live and

00:24:22.585 --> 00:24:23.746

microsoft translate and

00:24:23.766 --> 00:24:26.807

here's the app I love the idea for mlls

00:24:29.539 --> 00:24:31.060

What's the dog and pony, right?

00:24:31.121 --> 00:24:31.882

Is the dog and pony,

00:24:31.922 --> 00:24:33.403

here's designer.microsoft,

00:24:33.442 --> 00:24:34.364

give me a prompt and it's

00:24:34.384 --> 00:24:35.183

going to make a picture.

00:24:35.243 --> 00:24:37.826

Now go try something.

00:24:38.125 --> 00:24:38.886

What's the next thing

00:24:38.946 --> 00:24:40.268

outside of MLL students?

00:24:42.088 --> 00:24:42.470

I think...

00:24:44.321 --> 00:24:46.483

I think you'll get most

00:24:46.663 --> 00:24:49.244

teachers to buy in and want

00:24:49.306 --> 00:24:51.467

to understand more if we

00:24:51.567 --> 00:24:53.147

help solve problems for them.

00:24:53.607 --> 00:24:55.189

So I don't actually think

00:24:55.470 --> 00:24:57.911

it's the cool whiz bang dog and pony.

00:24:58.811 --> 00:25:01.394

Even I myself don't always

00:25:01.453 --> 00:25:04.214

appreciate what an AI tool

00:25:04.296 --> 00:25:06.596

can do in terms of making a

00:25:06.656 --> 00:25:07.758

presentation look better

00:25:08.057 --> 00:25:09.679

because I've got years of

00:25:09.719 --> 00:25:10.680

doing a presentation.

00:25:10.720 --> 00:25:12.221

That's not a problem I feel

00:25:12.260 --> 00:25:13.362

like I'm trying to solve.

00:25:14.281 --> 00:25:16.343

I think if you go at

00:25:16.982 --> 00:25:17.903

authentic problems that

00:25:17.923 --> 00:25:18.583

teachers are trying to

00:25:18.863 --> 00:25:20.844

solve and then think about it.

00:25:21.243 --> 00:25:22.545

I think there's always the

00:25:22.585 --> 00:25:25.105

problem of differentiation.

00:25:25.305 --> 00:25:26.986

We talked about multilingual learners,

00:25:27.086 --> 00:25:28.385

but another way of thinking

00:25:28.445 --> 00:25:29.467

about differentiation.

00:25:31.606 --> 00:25:33.208

A lesson I have learned when

00:25:33.248 --> 00:25:34.628

it comes to generative AI

00:25:34.749 --> 00:25:36.230

is you always need to keep

00:25:36.330 --> 00:25:37.270

humans in the loop.

00:25:37.951 --> 00:25:40.193

You cannot just totally rely

00:25:40.453 --> 00:25:42.555

on what the generative AI

00:25:43.076 --> 00:25:44.017

produces for you.

00:25:44.477 --> 00:25:46.137

You've got to keep checking it.

00:25:46.638 --> 00:25:48.660

And so that's why I think

00:25:51.294 --> 00:25:55.316

We'll never see AI fully replace teachers.

00:25:55.375 --> 00:25:57.497

We need their humanity and

00:25:57.537 --> 00:25:58.377

their understanding and

00:25:58.397 --> 00:25:59.419

their relationships with

00:25:59.479 --> 00:26:00.239

kids in the loop.

00:26:00.278 --> 00:26:02.460

So one way that we can leverage that is,

00:26:02.980 --> 00:26:04.141

say this is a middle school

00:26:04.181 --> 00:26:05.521

teacher in your context,

00:26:05.942 --> 00:26:08.483

and they have multiple classes of kids,

00:26:09.104 --> 00:26:10.305

and they're still, again,

00:26:10.464 --> 00:26:11.704

trying to teach maybe a

00:26:11.765 --> 00:26:14.027

social studies concept or a math concept.

00:26:14.747 --> 00:26:16.548

But they've got five sections of kids,

00:26:16.968 --> 00:26:18.409

and they all like different things.

00:26:18.609 --> 00:26:20.270

And the teacher can't keep

00:26:20.411 --> 00:26:21.731

coming up with all of these

00:26:21.771 --> 00:26:22.711

different examples

00:26:41.203 --> 00:26:44.165

Now give me an example to teach that.

00:26:44.465 --> 00:26:46.367

x math concept to all of

00:26:46.407 --> 00:26:49.049

these kids and it will

00:26:49.089 --> 00:26:51.411

generate it in seconds and

00:26:51.451 --> 00:26:52.511

I love that you just said

00:26:52.551 --> 00:26:53.913

that because a couple weeks

00:26:53.952 --> 00:26:55.054

ago I was teaching my kids

00:26:55.074 --> 00:26:56.315

how to do autobiographies

00:26:56.775 --> 00:26:57.776

and right in front of them

00:26:57.836 --> 00:26:58.997

I opened up copilot and

00:26:59.017 --> 00:26:59.577

said I need an

00:26:59.637 --> 00:27:00.959

autobiography that has this

00:27:00.999 --> 00:27:02.179

this and I basically

00:27:02.259 --> 00:27:03.079

plugged in what their

00:27:03.119 --> 00:27:06.001

assignment was and the kids

00:27:06.021 --> 00:27:08.763

were just like wait how'd you do that and

00:27:09.809 --> 00:27:10.653

That was kind of fun.

00:27:11.335 --> 00:27:12.519

But let me put on my third

00:27:12.558 --> 00:27:14.425

hat here as the technology teacher,

00:27:14.486 --> 00:27:16.010

as somebody who's in the classrooms.

00:27:17.289 --> 00:27:19.751

I'm still nervous to show

00:27:19.811 --> 00:27:21.073

this stuff to my students,

00:27:21.153 --> 00:27:22.512

even though it's on my accounts,

00:27:22.614 --> 00:27:23.894

even though they're not

00:27:23.913 --> 00:27:25.035

getting their hands on it.

00:27:25.695 --> 00:27:26.836

I still feel like I'm the

00:27:26.875 --> 00:27:27.696

guy that's teaching them

00:27:27.737 --> 00:27:28.917

how to use the calculator

00:27:28.938 --> 00:27:30.939

when the math teacher says no calculator,

00:27:30.979 --> 00:27:31.199

right?

00:27:31.419 --> 00:27:31.598

Right.

00:27:31.679 --> 00:27:33.039

I still feel like if I go in

00:27:33.079 --> 00:27:34.240

there and I show them how

00:27:34.340 --> 00:27:36.162

to use these things,

00:27:36.321 --> 00:27:37.803

eventually they're going to

00:27:37.843 --> 00:27:38.604

find the... And I don't

00:27:38.644 --> 00:27:39.443

want to be blamed as the

00:27:39.845 --> 00:27:40.644

guy who's teaching them all

00:27:40.664 --> 00:27:41.246

the back doors.

00:27:41.726 --> 00:27:41.885

Right.

00:27:42.465 --> 00:27:42.626

So...

00:27:43.487 --> 00:27:45.888

We talked about when you're doing the PD,

00:27:46.650 --> 00:27:48.250

help the teacher solve the problem,

00:27:48.270 --> 00:27:49.112

get them interested,

00:27:49.152 --> 00:27:50.472

and then you start to build from there.

00:27:51.614 --> 00:27:51.815

What...

00:27:52.994 --> 00:27:54.236

advice would you have for

00:27:54.336 --> 00:27:55.916

anybody trying to show off

00:27:56.698 --> 00:27:58.960

artificial intelligence for

00:27:59.099 --> 00:28:01.182

to students but doing in a

00:28:01.221 --> 00:28:02.863

way that's not the oh it's

00:28:02.883 --> 00:28:03.824

going to help me cheat on

00:28:03.844 --> 00:28:05.826

my you know right that

00:28:05.945 --> 00:28:06.945

stuff right how do you

00:28:07.106 --> 00:28:08.207

actually start to bring in

00:28:08.227 --> 00:28:10.128

this as a tool and we can

00:28:10.169 --> 00:28:11.349

discuss the canvas of the

00:28:11.390 --> 00:28:12.371

world and the fireflies

00:28:12.391 --> 00:28:14.392

like but what what's a good

00:28:14.432 --> 00:28:15.894

couple intro lessons for students

00:28:16.874 --> 00:28:19.914

So where I like to go is

00:28:20.634 --> 00:28:21.776

common sense education.

00:28:22.215 --> 00:28:23.455

I don't know if you've looked at,

00:28:23.576 --> 00:28:24.997

they're really well known

00:28:25.096 --> 00:28:27.237

for their digital citizenship curriculum,

00:28:27.777 --> 00:28:29.637

and they've now put out a

00:28:29.718 --> 00:28:32.078

series of lessons for students

00:28:33.038 --> 00:28:35.361

on AI that explains what it

00:28:35.500 --> 00:28:38.223

is and also kind of takes

00:28:38.284 --> 00:28:39.785

this digital citizenship

00:28:39.884 --> 00:28:42.887

approach to teaching and

00:28:42.948 --> 00:28:44.449

learning about AI.

00:28:44.750 --> 00:28:47.071

So if I were in your shoes

00:28:47.271 --> 00:28:48.373

as the tech teacher,

00:28:48.813 --> 00:28:49.953

I'd probably start there

00:28:50.134 --> 00:28:51.476

with their lessons because

00:28:51.496 --> 00:28:52.115

you're building an

00:28:52.135 --> 00:28:53.657

understanding of the tool,

00:28:53.758 --> 00:28:55.058

not just showing the cool

00:28:55.118 --> 00:28:56.619

whiz bang how it would help me.

00:28:57.461 --> 00:28:58.041

kind of a thing.

00:28:58.541 --> 00:29:00.063

So I think it's really

00:29:00.123 --> 00:29:01.805

important when we're

00:29:01.884 --> 00:29:03.886

talking with students that

00:29:03.946 --> 00:29:05.268

we help them understand

00:29:05.347 --> 00:29:07.289

what the tools do and don't do.

00:29:07.529 --> 00:29:08.570

We help them understand the

00:29:08.631 --> 00:29:10.612

biases that are built into them.

00:29:11.853 --> 00:29:15.296

We help them understand what

00:29:15.355 --> 00:29:16.657

they need to look out for

00:29:16.738 --> 00:29:18.058

that they can't just

00:29:20.865 --> 00:29:22.847

put in a prompt and turn in

00:29:22.968 --> 00:29:24.190

whatever it spits out.

00:29:24.510 --> 00:29:25.172

So again,

00:29:25.633 --> 00:29:26.894

translating the humans in the

00:29:26.934 --> 00:29:28.096

loop back to them.

00:29:28.576 --> 00:29:30.079

I would start with that

00:29:30.200 --> 00:29:31.141

resource and that

00:29:31.201 --> 00:29:33.345

collection of lessons as my

00:29:33.384 --> 00:29:34.105

first place to go.

00:29:35.156 --> 00:29:36.439

Then I would probably if

00:29:36.939 --> 00:29:39.821

your school allows you, you know,

00:29:39.883 --> 00:29:40.923

you've asked you've raised

00:29:40.943 --> 00:29:41.865

a bunch of important

00:29:41.924 --> 00:29:45.608

questions about the

00:29:45.648 --> 00:29:47.471

legalities of data sharing

00:29:47.550 --> 00:29:48.732

and having the right agreement.

00:29:48.813 --> 00:29:50.253

So let's say you do have

00:29:50.375 --> 00:29:52.836

permission to show it and

00:29:53.057 --> 00:29:54.098

your school has worked out

00:29:54.159 --> 00:29:55.380

all those legal details.

00:29:56.573 --> 00:29:58.355

I would probably start with

00:29:58.615 --> 00:30:02.877

the brainstorming capacity that AI does.

00:30:03.137 --> 00:30:05.721

So not doing the finished

00:30:05.861 --> 00:30:07.342

product part of it,

00:30:07.521 --> 00:30:09.344

because that's where some

00:30:09.384 --> 00:30:10.223

of your colleagues are

00:30:10.344 --> 00:30:13.067

probably kind of got their

00:30:13.126 --> 00:30:15.028

hackles up about cheating

00:30:15.208 --> 00:30:16.529

and the potential for cheating.

00:30:17.190 --> 00:30:19.310

And until we get all of our

00:30:19.351 --> 00:30:20.932

colleagues to change their

00:30:21.011 --> 00:30:23.792

pedagogy from the kinds of

00:30:23.853 --> 00:30:25.334

assignments that could be

00:30:25.634 --> 00:30:27.134

replicated and spit out by

00:30:27.173 --> 00:30:28.055

a generative AI,

00:30:28.855 --> 00:30:30.855

what I think we're best to

00:30:30.915 --> 00:30:32.957

do with the youth is to

00:30:33.096 --> 00:30:34.778

teach them how the tools

00:30:34.857 --> 00:30:36.218

could be a thought buddy,

00:30:36.238 --> 00:30:37.558

a brainstorming partner,

00:30:37.699 --> 00:30:38.900

an idea generator.

00:30:39.839 --> 00:30:41.980

Tools like ChatGPT are great for that.

00:30:46.894 --> 00:30:47.734

All of these topics that

00:30:47.755 --> 00:30:48.815

we're talking about are

00:30:48.855 --> 00:30:50.636

going to be detailed in our show notes.

00:30:50.717 --> 00:30:51.636

I'm making sure that we have

00:30:51.696 --> 00:30:53.459

links to all the different AI tools.

00:30:54.118 --> 00:30:55.019

I found the link to the

00:30:55.079 --> 00:30:56.201

Common Sense article.

00:30:56.641 --> 00:30:58.402

And speaking of articles, Dr. Lammers,

00:30:58.561 --> 00:31:00.002

you recently at Edmentum

00:31:00.943 --> 00:31:03.346

published an article about generative AI.

00:31:03.786 --> 00:31:05.867

And that article was called

00:31:06.567 --> 00:31:08.690

AI in Education Experiments,

00:31:09.190 --> 00:31:10.290

Lessons Learned.

00:31:11.030 --> 00:31:11.830

Talk to us a little bit

00:31:11.892 --> 00:31:13.551

about this post and specifically,

00:31:13.751 --> 00:31:15.212

what have some of the

00:31:15.252 --> 00:31:16.794

lessons been that you and

00:31:16.814 --> 00:31:18.314

your team have learned about AI?

00:31:19.075 --> 00:31:19.214

Well,

00:31:19.255 --> 00:31:20.535

I've already shared a couple of them.

00:31:20.634 --> 00:31:22.076

So the experiment from

00:31:22.155 --> 00:31:23.636

Claude and translating

00:31:23.696 --> 00:31:24.997

comes directly from that

00:31:25.396 --> 00:31:27.678

article that you'll link to.

00:31:28.199 --> 00:31:29.679

The other thing that we did is...

00:31:31.962 --> 00:31:34.865

The needing to try different

00:31:34.964 --> 00:31:37.007

tools and to try them over

00:31:37.067 --> 00:31:38.788

time to see how they work

00:31:38.828 --> 00:31:39.788

and how they change.

00:31:40.410 --> 00:31:42.290

So to see whether or not

00:31:42.371 --> 00:31:44.373

ChatGPT might be better at

00:31:44.413 --> 00:31:47.576

something versus Copilot

00:31:47.675 --> 00:31:50.178

versus Gemini versus Cloud AI.

00:31:51.898 --> 00:31:54.121

The other thing that when I

00:31:54.161 --> 00:31:55.541

go back to this idea of the

00:31:55.602 --> 00:31:56.782

coalition of the willing

00:31:56.942 --> 00:31:57.884

who are going to run

00:31:57.983 --> 00:31:59.724

experiments and try things,

00:32:00.586 --> 00:32:02.866

that I think this works best

00:32:04.048 --> 00:32:05.607

if they can then have the

00:32:05.647 --> 00:32:06.788

time and space to come

00:32:06.848 --> 00:32:08.289

together and critically

00:32:08.369 --> 00:32:10.030

reflect on what they've learned,

00:32:10.391 --> 00:32:11.570

to share resources,

00:32:11.971 --> 00:32:14.413

that there be created some sort of a hub.

00:32:14.913 --> 00:32:16.693

For us at Edmentum,

00:32:16.713 --> 00:32:18.555

we used a Microsoft Teams channel,

00:32:19.555 --> 00:32:21.915

which we called our AI brainstorming hub.

00:32:22.076 --> 00:32:23.836

And any resource gets shared

00:32:23.876 --> 00:32:25.337

there so that anyone who's

00:32:25.397 --> 00:32:26.898

interested can follow along,

00:32:27.419 --> 00:32:29.259

can dialogue about it.

00:32:30.079 --> 00:32:33.488

So I think that idea of

00:32:33.627 --> 00:32:35.070

creating this space for

00:32:35.152 --> 00:32:37.155

experimentation is really helpful.

00:32:38.048 --> 00:32:40.329

The article also shares the

00:32:40.369 --> 00:32:41.631

lesson we've already talked about,

00:32:41.671 --> 00:32:43.511

about keeping humans in the loop,

00:32:43.811 --> 00:32:45.231

that you need to have

00:32:46.692 --> 00:32:49.815

people look over what the AI creates,

00:32:50.694 --> 00:32:52.276

find hallucinations,

00:32:53.375 --> 00:32:54.576

which that's another key

00:32:54.636 --> 00:32:56.218

term that we haven't touched on,

00:32:56.657 --> 00:33:00.059

but because of the way AI is designed,

00:33:00.720 --> 00:33:01.619

it could generate

00:33:01.780 --> 00:33:04.602

falsehoods that look very believable,

00:33:04.781 --> 00:33:05.362

because again,

00:33:05.461 --> 00:33:06.563

it's just trying to please you.

00:33:07.002 --> 00:33:08.222

It's trying to give you what

00:33:08.262 --> 00:33:09.003

it thinks you want.

00:33:10.644 --> 00:33:12.203

And so if you get to the

00:33:12.284 --> 00:33:14.224

point where you are using

00:33:14.244 --> 00:33:16.105

AI with students,

00:33:16.645 --> 00:33:18.986

that article also has some

00:33:19.046 --> 00:33:20.086

lessons learned that

00:33:20.227 --> 00:33:23.446

specifically speak to work with students.

00:33:24.688 --> 00:33:25.907

And this idea that we need

00:33:25.928 --> 00:33:27.208

to promote critical

00:33:27.368 --> 00:33:29.388

thinking and reflection on

00:33:29.409 --> 00:33:30.848

the student's part as they

00:33:30.990 --> 00:33:32.509

analyze AI's output.

00:33:33.988 --> 00:33:35.631

You mentioned Cloud AI

00:33:35.730 --> 00:33:39.237

earlier about being a good tool for MLL.

00:33:39.257 --> 00:33:43.084

I want to say this the right way.

00:33:43.565 --> 00:33:45.749

Have you focused these AI tools

00:33:46.957 --> 00:33:48.376

for certain subjects.

00:33:48.436 --> 00:33:49.037

For instance,

00:33:49.958 --> 00:33:51.518

have you noticed that Copilot

00:33:51.557 --> 00:33:52.958

might be good at some subjects,

00:33:53.018 --> 00:33:55.818

but Gemini is better at others?

00:33:57.419 --> 00:34:00.819

I find there's people in

00:34:00.839 --> 00:34:01.579

certain circles that

00:34:01.720 --> 00:34:02.381

they're just going to try

00:34:02.381 --> 00:34:03.560

100 different AI tools,

00:34:03.941 --> 00:34:04.760

and they're going to always

00:34:04.820 --> 00:34:06.020

have 100 AI tools because

00:34:06.300 --> 00:34:07.201

they know what's there.

00:34:07.662 --> 00:34:09.242

But the majority of teachers are either,

00:34:09.262 --> 00:34:10.282

I don't want it,

00:34:10.402 --> 00:34:12.682

or show me the one that I need.

00:34:13.163 --> 00:34:14.143

Right, exactly.

00:34:14.362 --> 00:34:15.344

In a school district, look,

00:34:15.623 --> 00:34:17.164

if you're a Google school,

00:34:17.204 --> 00:34:18.105

you're going to do this one.

00:34:18.144 --> 00:34:19.264

If you're a Microsoft school,

00:34:19.304 --> 00:34:20.266

you're going to do this one.

00:34:20.326 --> 00:34:20.865

If you're not,

00:34:21.385 --> 00:34:22.405

here are some other options.

00:34:23.327 --> 00:34:24.067

Have you found some

00:34:24.166 --> 00:34:26.768

favorites yet and for specific reasons?

00:34:28.788 --> 00:34:29.068

Well,

00:34:30.108 --> 00:34:34.371

I know that when we were trying to

00:34:34.490 --> 00:34:37.512

use ChatGPT to do certain calculations,

00:34:37.813 --> 00:34:38.813

it couldn't always be

00:34:38.873 --> 00:34:40.014

trusted with the math.

00:34:40.655 --> 00:34:40.936

Now,

00:34:41.416 --> 00:34:43.777

I say that with a huge caveat that

00:34:44.257 --> 00:34:46.958

when we were doing our experiments,

00:34:47.039 --> 00:34:47.940

that was last year.

00:34:48.280 --> 00:34:49.300

That might as well be a

00:34:49.380 --> 00:34:51.722

decade ago in AI terms, right?

00:34:52.483 --> 00:34:54.184

So it is ever-changing.

00:34:55.126 --> 00:34:57.148

So I don't know that there

00:34:57.188 --> 00:34:58.829

is a great answer to your

00:34:58.889 --> 00:35:01.012

question definitively, Jeff.

00:35:01.974 --> 00:35:03.514

I think that as these models

00:35:03.574 --> 00:35:04.516

continue to change,

00:35:04.576 --> 00:35:05.637

that's why we need a

00:35:06.978 --> 00:35:08.561

culture of experimentation.

00:35:10.161 --> 00:35:10.782

There is another...

00:35:11.800 --> 00:35:13.402

form of ai that we haven't

00:35:13.463 --> 00:35:14.764

talked about yet and I

00:35:14.824 --> 00:35:15.905

really haven't talked about

00:35:15.945 --> 00:35:17.047

it much on this channel

00:35:17.108 --> 00:35:17.887

because I'm still

00:35:18.869 --> 00:35:20.371

fascinated by how it works

00:35:20.592 --> 00:35:21.413

and I'm just gonna I don't

00:35:21.432 --> 00:35:22.012

even know what this is

00:35:22.253 --> 00:35:23.675

specifically called but I i

00:35:23.715 --> 00:35:24.996

like the term second brain

00:35:25.637 --> 00:35:27.119

so I i call it your second

00:35:27.139 --> 00:35:29.302

brain ai and specifically things that

00:35:30.333 --> 00:35:32.173

They will look at all of

00:35:32.273 --> 00:35:34.096

your personal information

00:35:34.335 --> 00:35:35.976

and help you make decisions,

00:35:36.577 --> 00:35:37.757

help you organize.

00:35:38.378 --> 00:35:39.900

I'll give you two examples

00:35:40.159 --> 00:35:41.380

that helped me run my life

00:35:41.420 --> 00:35:42.641

and helped me run TeacherCast.

00:35:43.672 --> 00:35:45.032

I'm a big fan of an

00:35:45.152 --> 00:35:46.974

application called Notion.

00:35:47.735 --> 00:35:49.436

And Notion is a note-taking

00:35:49.516 --> 00:35:51.777

application on one level,

00:35:51.836 --> 00:35:53.398

but it's also a way to

00:35:53.518 --> 00:35:54.818

create databases and

00:35:54.878 --> 00:35:56.840

note-take and you name it.

00:35:56.880 --> 00:35:57.320

Basically,

00:35:57.521 --> 00:35:59.021

everything that you've ever seen

00:35:59.081 --> 00:36:00.182

on TeacherCast for the last

00:36:00.242 --> 00:36:02.304

couple of years is designed in Notion.

00:36:03.105 --> 00:36:03.804

And recently,

00:36:04.224 --> 00:36:06.487

Notion came out with their own AI tool,

00:36:07.067 --> 00:36:09.047

but instead of searching the world,

00:36:09.528 --> 00:36:11.389

it's searching itself, right?

00:36:11.650 --> 00:36:12.911

So when we say things like

00:36:12.951 --> 00:36:14.130

the term second brain,

00:36:14.150 --> 00:36:16.112

it literally is thinking for me.

00:36:16.532 --> 00:36:18.193

And I can actually go into

00:36:18.233 --> 00:36:19.954

the AI tool and I can ask it,

00:36:20.333 --> 00:36:21.894

tell me how many times Dr.

00:36:21.954 --> 00:36:23.195

Lammers was on the show and

00:36:23.255 --> 00:36:24.795

what the episodes were about.

00:36:25.695 --> 00:36:26.916

Maybe because in six months

00:36:26.936 --> 00:36:27.876

you're going to be back on

00:36:27.916 --> 00:36:28.838

and I want to make sure

00:36:28.878 --> 00:36:30.057

that we're having a similar

00:36:30.097 --> 00:36:31.278

yet different conversation.

00:36:31.838 --> 00:36:32.659

Or I can say,

00:36:33.079 --> 00:36:34.500

show me all the podcast

00:36:34.559 --> 00:36:36.081

episodes that we discussed.

00:36:36.661 --> 00:36:37.842

artificial intelligence

00:36:37.902 --> 00:36:38.882

because maybe I'm doing a

00:36:38.902 --> 00:36:40.364

blog post on my top 10

00:36:40.364 --> 00:36:41.585

whatever and I want to

00:36:41.626 --> 00:36:43.668

start to reference other

00:36:43.708 --> 00:36:45.009

shows so notion is a way

00:36:45.048 --> 00:36:46.630

that it'll actually take

00:36:46.690 --> 00:36:48.592

your your again your second

00:36:48.632 --> 00:36:50.313

brain it'll only think

00:36:50.574 --> 00:36:53.637

inside of that co-pilot is

00:36:53.697 --> 00:36:55.719

another option co-pilot

00:36:55.760 --> 00:36:56.880

depending on how you're

00:36:57.061 --> 00:36:58.501

using it and I i again I

00:36:58.541 --> 00:36:59.963

pay for it inside of my

00:37:00.003 --> 00:37:00.965

teacher cast domain

00:37:02.483 --> 00:37:05.023

as a switch that says internal,

00:37:05.043 --> 00:37:06.164

I forget what the exact words are,

00:37:06.184 --> 00:37:07.505

but basically it's internal

00:37:07.545 --> 00:37:09.186

of your domain or the web.

00:37:09.206 --> 00:37:10.925

So if I click on the

00:37:11.086 --> 00:37:12.567

internal switch and I don't

00:37:12.586 --> 00:37:14.947

remember the name of it, but I can say,

00:37:15.387 --> 00:37:17.469

show me all of my podcast

00:37:17.568 --> 00:37:19.710

episodes and it'll find

00:37:20.210 --> 00:37:21.951

only the podcast episodes

00:37:22.170 --> 00:37:23.992

inside of my OneDrive.

00:37:25.012 --> 00:37:26.733

Whereas if I search the web,

00:37:26.833 --> 00:37:28.393

now it's basically doing a Bing search.

00:37:30.347 --> 00:37:32.168

And so I love these

00:37:32.889 --> 00:37:34.032

companies that are coming

00:37:34.112 --> 00:37:35.994

up with ways for us to do

00:37:36.114 --> 00:37:38.458

more using the tools that

00:37:38.498 --> 00:37:40.981

we're currently building, right?

00:37:41.320 --> 00:37:42.682

So I spend a lot of time

00:37:43.815 --> 00:37:45.695

on my Notion, on my dashboards,

00:37:46.217 --> 00:37:46.896

I'm making sure that

00:37:47.036 --> 00:37:48.577

everything is there and named correctly,

00:37:48.597 --> 00:37:50.199

because I know someday soon

00:37:50.659 --> 00:37:51.679

I'm going to need to pull

00:37:51.719 --> 00:37:52.940

that information out.

00:37:53.842 --> 00:37:55.362

And the same thing with Microsoft.

00:37:55.422 --> 00:37:56.663

Microsoft is checking all of

00:37:56.702 --> 00:37:57.704

your PowerPoints and words

00:37:57.744 --> 00:37:58.925

and Excels and it's

00:37:59.005 --> 00:38:00.266

checking the entire

00:38:00.346 --> 00:38:02.987

knowledge graph out there of yourself.

00:38:03.047 --> 00:38:05.009

So that way you can find what you need.

00:38:05.048 --> 00:38:05.628

Now, obviously,

00:38:06.090 --> 00:38:07.811

if I'm searching my own stuff,

00:38:08.451 --> 00:38:09.891

it doesn't know what you as

00:38:09.931 --> 00:38:10.952

my coworkers doing.

00:38:11.833 --> 00:38:13.173

But that's okay because I

00:38:13.193 --> 00:38:14.295

don't always want to know

00:38:14.315 --> 00:38:15.695

what the entire planet's doing.

00:38:16.114 --> 00:38:16.775

I just want to know what's

00:38:16.815 --> 00:38:18.817

in my own bedroom or my own house.

00:38:20.077 --> 00:38:21.518

Do you have any experience

00:38:21.637 --> 00:38:23.438

using any applications like that?

00:38:23.579 --> 00:38:24.500

Or you were shaking your

00:38:24.539 --> 00:38:27.021

head about using the co-pilot stuff.

00:38:28.021 --> 00:38:29.282

Are you one or is your team

00:38:29.882 --> 00:38:32.543

one to be making these second brains,

00:38:32.603 --> 00:38:33.445

second thinking,

00:38:34.505 --> 00:38:35.985

digital versions of yourself?

00:38:36.146 --> 00:38:36.666

And if so,

00:38:36.686 --> 00:38:39.027

do you have any suggestions on those?

00:38:39.047 --> 00:38:39.387

Yeah.

00:38:40.106 --> 00:38:42.507

The only place that I have used this,

00:38:42.588 --> 00:38:45.289

I have not dug into this

00:38:45.329 --> 00:38:47.670

kind of second brain AI for

00:38:47.731 --> 00:38:50.572

myself very much beyond, you know,

00:38:50.672 --> 00:38:51.932

working in a corporation

00:38:51.972 --> 00:38:53.313

that uses Microsoft

00:38:53.353 --> 00:38:56.076

products and also Atlassian products,

00:38:56.215 --> 00:38:57.617

Confluence, right?

00:38:58.137 --> 00:39:01.179

We use SharePoint, there's stuff on Teams,

00:39:01.219 --> 00:39:03.179

there's files that get emailed to you,

00:39:03.400 --> 00:39:03.780

all of this.

00:39:04.340 --> 00:39:07.083

So I often use the tool

00:39:07.322 --> 00:39:10.865

Delve in Microsoft to find, okay,

00:39:11.005 --> 00:39:13.307

I know this person sent me a file.

00:39:14.469 --> 00:39:15.190

Where is it?

00:39:16.250 --> 00:39:17.030

Help me find it.

00:39:17.231 --> 00:39:18.411

And so I don't have to

00:39:18.452 --> 00:39:19.773

search email and then

00:39:19.853 --> 00:39:21.954

search Teams and then search, you know,

00:39:22.414 --> 00:39:22.976

Confluence.

00:39:23.275 --> 00:39:25.358

That's probably the best one that I use.

00:39:25.398 --> 00:39:26.539

And I use it regularly

00:39:26.559 --> 00:39:28.139

because I know I saw that

00:39:28.199 --> 00:39:29.221

file from somebody.

00:39:29.240 --> 00:39:29.840

Yes.

00:39:30.382 --> 00:39:30.621

Mm-hmm.

00:39:32.449 --> 00:39:33.710

There's a lot, right?

00:39:33.731 --> 00:39:34.271

There's a lot.

00:39:34.951 --> 00:39:37.072

And I think where we are

00:39:37.172 --> 00:39:38.373

right now is we're at that

00:39:38.492 --> 00:39:39.793

point in the curve where

00:39:40.213 --> 00:39:41.594

people are jumping on board

00:39:41.614 --> 00:39:44.115

or some of them are even saying,

00:39:45.356 --> 00:39:45.996

I don't have the time.

00:39:47.257 --> 00:39:50.438

So much stuff, grades, curriculum, parents,

00:39:51.159 --> 00:39:53.380

post-pandemic, behavior.

00:39:54.021 --> 00:39:56.141

I don't have time for one more thing.

00:39:56.641 --> 00:39:57.742

And you've got this wave of

00:39:57.862 --> 00:40:00.344

educators coming in going, no, no, no,

00:40:00.583 --> 00:40:01.744

this is the thing.

00:40:02.085 --> 00:40:02.425

Right?

00:40:02.505 --> 00:40:04.168

And even a couple of shows ago, we did the,

00:40:04.208 --> 00:40:04.449

you know,

00:40:05.329 --> 00:40:07.353

how would you relate AI to other

00:40:07.452 --> 00:40:08.295

recent things?

00:40:08.355 --> 00:40:08.835

And we're like, no,

00:40:08.856 --> 00:40:10.617

this isn't Google Cardboard

00:40:10.677 --> 00:40:12.481

where many people try it

00:40:12.521 --> 00:40:13.483

and now it's in the corner.

00:40:13.523 --> 00:40:13.663

Like,

00:40:14.485 --> 00:40:16.045

This is the thing, right?

00:40:16.065 --> 00:40:17.246

Like this is the thing that

00:40:17.266 --> 00:40:18.286

we're going to look at and go,

00:40:18.306 --> 00:40:20.268

this isn't going anywhere.

00:40:21.168 --> 00:40:22.588

This is the calculator that

00:40:22.628 --> 00:40:23.469

suddenly you turn around

00:40:23.489 --> 00:40:24.670

and everyone's got one in their pocket.

00:40:24.690 --> 00:40:27.451

Like everything is going into here.

00:40:27.590 --> 00:40:28.630

So how do we learn?

00:40:28.650 --> 00:40:32.333

And let's take one final lap around here.

00:40:32.932 --> 00:40:34.034

If you were listening to

00:40:34.074 --> 00:40:35.614

this show and you wanted to

00:40:35.653 --> 00:40:37.474

take that first step to try things,

00:40:37.514 --> 00:40:39.516

as you said, button push, test things out,

00:40:39.976 --> 00:40:40.757

play with things.

00:40:41.577 --> 00:40:42.318

what would be one of the

00:40:42.418 --> 00:40:43.498

first things that you would

00:40:43.538 --> 00:40:44.778

do or the first

00:40:45.380 --> 00:40:46.420

applications that you would

00:40:46.460 --> 00:40:48.822

look towards just to sit in

00:40:48.862 --> 00:40:49.742

your office one day and

00:40:49.782 --> 00:40:50.443

push some buttons?

00:40:51.224 --> 00:40:51.364

Well,

00:40:51.403 --> 00:40:53.945

if I'm at a school that uses Microsoft,

00:40:54.065 --> 00:40:55.025

I would use Microsoft

00:40:55.045 --> 00:40:56.067

Copilot because it's

00:40:56.106 --> 00:40:57.588

probably the easiest one to

00:40:57.628 --> 00:40:59.188

know that the data is protected,

00:40:59.248 --> 00:41:00.230

so I won't get in trouble.

00:41:01.166 --> 00:41:02.688

If I'm not at a school that uses that,

00:41:02.788 --> 00:41:04.088

I just go to chat GPT

00:41:04.148 --> 00:41:05.068

because there's a lot

00:41:05.469 --> 00:41:07.590

talked about jet chat GPT.

00:41:07.789 --> 00:41:10.090

And so I could find resources easily.

00:41:10.451 --> 00:41:11.731

So I'd pick one of those two,

00:41:11.871 --> 00:41:13.552

whichever one is the most accessible.

00:41:14.251 --> 00:41:15.393

And then I would sit down

00:41:15.432 --> 00:41:16.472

and think about what are

00:41:16.612 --> 00:41:18.373

all of the repetitive tasks

00:41:18.614 --> 00:41:20.175

that take me lots of time

00:41:21.235 --> 00:41:25.896

and how might I find or try myself

00:41:28.108 --> 00:41:29.949

a prompt that helps me save

00:41:30.048 --> 00:41:32.028

time with any one of those tasks,

00:41:32.170 --> 00:41:34.070

whether it's parent communications,

00:41:34.550 --> 00:41:36.570

whether it's designing

00:41:37.070 --> 00:41:39.152

student samples for essays

00:41:39.251 --> 00:41:40.331

as I'm trying to teach

00:41:40.411 --> 00:41:42.052

something in my English class,

00:41:42.413 --> 00:41:43.413

whatever it may be,

00:41:43.472 --> 00:41:46.054

if it's a differentiation task,

00:41:46.114 --> 00:41:47.755

and I'm trying to make sure that

00:41:48.894 --> 00:41:50.476

All of the kids have

00:41:51.416 --> 00:41:53.378

examples that relate to

00:41:53.438 --> 00:41:54.719

their particular interest.

00:41:55.019 --> 00:41:55.920

Whatever it may be,

00:41:55.940 --> 00:41:58.083

I would use one of those

00:41:58.143 --> 00:41:59.463

tools to try to create

00:41:59.543 --> 00:42:00.644

things that save me time.

00:42:02.481 --> 00:42:03.483

And I would add in there,

00:42:04.063 --> 00:42:05.804

try prompts that are serious.

00:42:06.565 --> 00:42:07.947

Try prompts that are silly.

00:42:07.987 --> 00:42:09.007

There's nothing wrong with

00:42:09.088 --> 00:42:10.668

opening up Copilot or any

00:42:10.688 --> 00:42:11.369

of these and saying,

00:42:11.409 --> 00:42:12.490

tell me a knock-knock joke.

00:42:13.271 --> 00:42:13.472

Right.

00:42:13.771 --> 00:42:14.932

Just try something.

00:42:14.972 --> 00:42:15.193

You know,

00:42:15.253 --> 00:42:16.634

today was the last day of our

00:42:16.653 --> 00:42:17.414

marking period.

00:42:17.474 --> 00:42:19.396

I had to write those emails to parents.

00:42:19.777 --> 00:42:20.838

There's nothing wrong with

00:42:20.878 --> 00:42:21.639

going in and saying,

00:42:22.039 --> 00:42:23.199

write me a letter to this

00:42:23.280 --> 00:42:24.561

parent about their student

00:42:24.601 --> 00:42:26.041

who is not doing so well.

00:42:26.402 --> 00:42:27.744

And you don't have to send it,

00:42:28.264 --> 00:42:30.246

but just see what it comes back with.

00:42:30.952 --> 00:42:32.153

And what I like to do in the

00:42:32.233 --> 00:42:35.094

write me a letter kind of

00:42:35.255 --> 00:42:37.996

use case is write me a letter about,

00:42:38.097 --> 00:42:38.896

you know, student.

00:42:39.077 --> 00:42:40.418

You put the student's name in.

00:42:40.797 --> 00:42:43.039

You're still protecting privacy because,

00:42:43.099 --> 00:42:44.320

you know, they don't know that student.

00:42:44.340 --> 00:42:45.219

You use the first name.

00:42:45.621 --> 00:42:46.221

And you say,

00:42:46.501 --> 00:42:47.661

and I want to make sure that I

00:42:47.681 --> 00:42:49.222

tell the parent three things.

00:42:49.262 --> 00:42:51.344

And you just put it in bullet point form.

00:42:51.563 --> 00:42:52.945

And I need it to be clear.

00:42:54.025 --> 00:42:56.206

Two paragraphs long, I need it to come,

00:42:56.565 --> 00:42:56.985

whatever.

00:42:57.186 --> 00:42:58.626

However much you want to give it,

00:42:59.007 --> 00:43:00.126

and you'll see that it

00:43:00.186 --> 00:43:01.206

creates something for you.

00:43:01.588 --> 00:43:02.407

And then the other thing

00:43:02.427 --> 00:43:03.088

that I would tell the

00:43:03.128 --> 00:43:04.889

teachers who are just trying this out,

00:43:05.389 --> 00:43:06.708

remember that this is a chat.

00:43:06.869 --> 00:43:08.489

So if you don't like what it gave you,

00:43:09.309 --> 00:43:11.030

tell it to change something, right?

00:43:11.411 --> 00:43:12.670

So you don't have to take

00:43:12.731 --> 00:43:15.192

the initial output and then use it or say,

00:43:15.692 --> 00:43:16.353

this doesn't work.

00:43:16.652 --> 00:43:18.632

Because where the real power

00:43:18.733 --> 00:43:21.474

comes is in its ability to

00:43:21.574 --> 00:43:23.275

iterate based on feedback from you.

00:43:23.942 --> 00:43:26.623

I first got into chat GPT

00:43:27.103 --> 00:43:28.465

when I was redesigning my

00:43:28.525 --> 00:43:30.025

resume and I popped it in

00:43:30.226 --> 00:43:32.387

and I popped the entire resume and I said,

00:43:32.467 --> 00:43:33.327

make it better, right?

00:43:33.367 --> 00:43:34.728

Because that's basic.

00:43:34.748 --> 00:43:35.949

You're learning how to do stuff.

00:43:36.710 --> 00:43:37.811

And it was okay,

00:43:37.871 --> 00:43:40.092

but still on that overall horrible side.

00:43:40.552 --> 00:43:41.612

And so then I just ended up,

00:43:41.693 --> 00:43:43.173

I went bullet point by bullet point.

00:43:43.653 --> 00:43:45.335

Here's a thing that's on my resume.

00:43:45.815 --> 00:43:47.757

Please make this sound more professional.

00:43:49.382 --> 00:43:50.983

And little by little,

00:43:51.045 --> 00:43:52.885

I just started carving out my documents.

00:43:52.905 --> 00:43:54.208

And then I went into my bio.

00:43:54.708 --> 00:43:55.728

Here's what I have.

00:43:56.128 --> 00:43:59.132

Please add these three or four new things.

00:43:59.713 --> 00:44:01.394

And then here it is.

00:44:01.434 --> 00:44:02.476

And then you put down,

00:44:02.976 --> 00:44:07.159

please give me this for a job interview.

00:44:07.199 --> 00:44:08.802

Please give me this for my website.

00:44:08.842 --> 00:44:10.563

Please give me this for a presentation.

00:44:10.623 --> 00:44:11.744

Please give me this in 150 words or less.

00:44:13.346 --> 00:44:14.226

And again,

00:44:14.445 --> 00:44:15.626

whether you use them or not is a

00:44:15.646 --> 00:44:16.086

different kind,

00:44:16.126 --> 00:44:17.407

but you're just trying

00:44:17.527 --> 00:44:18.387

things and you're putting

00:44:18.447 --> 00:44:19.146

stuff out there.

00:44:19.186 --> 00:44:20.168

You're putting your toe in

00:44:20.188 --> 00:44:22.288

the water and seeing where it is.

00:44:23.407 --> 00:44:23.708

Obviously,

00:44:23.728 --> 00:44:25.528

you mentioned that your team

00:44:25.548 --> 00:44:27.289

started doing this research last year.

00:44:27.768 --> 00:44:28.849

Where are you today?

00:44:29.269 --> 00:44:31.510

Where do you plan on being tomorrow?

00:44:31.829 --> 00:44:33.130

What's in the future for

00:44:33.170 --> 00:44:34.951

your team in studying and

00:44:34.990 --> 00:44:36.411

in using and in sharing the

00:44:36.451 --> 00:44:37.492

knowledge about artificial

00:44:37.532 --> 00:44:38.351

intelligence with the world?

00:44:39.277 --> 00:44:39.498

Well,

00:44:40.137 --> 00:44:42.760

we continue to run experiments and do

00:44:42.840 --> 00:44:44.541

projects to help figure out

00:44:44.621 --> 00:44:45.860

how to save us time.

00:44:45.960 --> 00:44:47.802

So as you mentioned at the beginning,

00:44:48.822 --> 00:44:50.664

we have our products in

00:44:51.264 --> 00:44:52.545

districts around the country.

00:44:52.605 --> 00:44:53.824

So we're always looking to

00:44:53.864 --> 00:44:55.266

make sure that our products

00:44:55.365 --> 00:44:57.067

meet the standards for all

00:44:57.106 --> 00:44:58.288

of these different states.

00:44:58.947 --> 00:45:00.268

And since we don't have a

00:45:00.329 --> 00:45:01.469

centralized curriculum in

00:45:01.489 --> 00:45:02.230

the United States,

00:45:02.650 --> 00:45:04.251

you can imagine that a

00:45:04.831 --> 00:45:06.152

large language model and

00:45:06.574 --> 00:45:07.634

different machine learning

00:45:07.693 --> 00:45:09.034

could help us look across

00:45:09.275 --> 00:45:10.456

all of the state standards

00:45:11.556 --> 00:45:12.818

to make sure that we have

00:45:12.838 --> 00:45:14.338

the alignment that we say we do.

00:45:14.900 --> 00:45:16.681

So that's one very popular

00:45:16.721 --> 00:45:18.402

project and one use that

00:45:18.442 --> 00:45:19.702

we're using AI for.

00:45:21.143 --> 00:45:23.166

But what we're continuing to

00:45:23.206 --> 00:45:24.447

do is to try to have

00:45:24.547 --> 00:45:25.849

conversations with our

00:45:25.929 --> 00:45:27.090

education partners and the

00:45:27.170 --> 00:45:28.293

folks in the schools who

00:45:28.452 --> 00:45:29.815

use our products and who

00:45:29.875 --> 00:45:31.436

are worried about our AI.

00:45:31.737 --> 00:45:33.298

And we're continuing to have

00:45:33.318 --> 00:45:35.101

this kind of internal

00:45:35.161 --> 00:45:37.364

experimentation so that we

00:45:37.483 --> 00:45:39.927

know how to advise our

00:45:39.987 --> 00:45:40.708

education partners.

00:45:40.768 --> 00:45:41.487

One of the things that I

00:45:41.527 --> 00:45:43.088

really enjoy about working

00:45:43.188 --> 00:45:46.070

for a company that really

00:45:46.150 --> 00:45:48.110

values educators first,

00:45:48.250 --> 00:45:49.271

like Edmentum does,

00:45:49.710 --> 00:45:50.871

is that we're not just

00:45:50.951 --> 00:45:52.391

trying to sell our products.

00:45:52.492 --> 00:45:54.271

We're really trying to be in

00:45:54.331 --> 00:45:55.632

relationship with those

00:45:55.693 --> 00:45:56.693

folks who use it and to

00:45:56.813 --> 00:45:57.974

understand their daily

00:45:58.594 --> 00:46:00.393

realities and to help them

00:46:00.494 --> 00:46:02.635

figure out how to make

00:46:02.735 --> 00:46:04.056

things work best for those

00:46:04.096 --> 00:46:04.936

daily realities.

00:46:05.922 --> 00:46:07.204

Talking today to Dr. Jane

00:46:07.284 --> 00:46:08.565

Lambers from Edmentum.

00:46:08.746 --> 00:46:08.967

Jane,

00:46:09.027 --> 00:46:10.088

where can we learn more about the

00:46:10.128 --> 00:46:11.190

great work you're doing and

00:46:11.269 --> 00:46:12.010

how do we get in touch with

00:46:12.030 --> 00:46:13.413

you if you have any other questions?

00:46:13.893 --> 00:46:15.295

I think LinkedIn is the best

00:46:15.376 --> 00:46:17.278

way to reach me and I'll

00:46:17.318 --> 00:46:18.400

make sure you have that to

00:46:18.440 --> 00:46:19.242

put in your show notes.

00:46:19.914 --> 00:46:20.295

And of course,

00:46:20.315 --> 00:46:20.916

you can find out more

00:46:20.936 --> 00:46:22.876

information over at edmentum.com.

00:46:22.996 --> 00:46:23.737

All of our show notes are

00:46:23.757 --> 00:46:24.398

going to be over there.

00:46:24.438 --> 00:46:25.898

This is Digital Learning Today.

00:46:25.938 --> 00:46:26.460

You can, of course,

00:46:26.500 --> 00:46:27.740

check out everything we

00:46:27.780 --> 00:46:28.840

have going on over at the

00:46:28.860 --> 00:46:30.322

TeacherCast Educational Network.

00:46:30.722 --> 00:46:31.802

Find out more information,

00:46:31.822 --> 00:46:32.643

like and subscribe,

00:46:32.684 --> 00:46:33.644

all that great stuff over

00:46:33.684 --> 00:46:34.864

at teachercast.net.

00:46:35.085 --> 00:46:35.686

Dr. Lammers,

00:46:35.865 --> 00:46:37.827

thank you so much for joining us today.

00:46:38.347 --> 00:46:38.867

Thank you, Jeff.

00:46:38.987 --> 00:46:39.568

It was a pleasure.

00:46:40.048 --> 00:46:40.809

And that wraps up this

00:46:40.849 --> 00:46:42.190

episode of Digital Learning today.

00:46:42.271 --> 00:46:43.550

I hope you guys had a good

00:46:43.590 --> 00:46:44.692

time and I hope you learned

00:46:45.012 --> 00:46:45.833

something that you can

00:46:45.893 --> 00:46:46.934

share with your faculty.

00:46:47.233 --> 00:46:48.054

There's one thing that we

00:46:48.094 --> 00:46:49.715

know here about artificial intelligence.

00:46:49.914 --> 00:46:51.737

It ain't going away.

00:46:52.197 --> 00:46:53.757

So have a good time with it.

00:46:53.838 --> 00:46:54.938

Let us know what you're thinking.

00:46:54.958 --> 00:46:56.340

And if you're interested,

00:46:57.059 --> 00:46:57.619

reach out to me.

00:46:57.900 --> 00:46:58.920

Would love to have you be a

00:46:58.981 --> 00:47:00.121

guest on this show as we

00:47:00.141 --> 00:47:01.123

get into the summertime.

00:47:01.443 --> 00:47:02.143

And that wraps up this

00:47:02.182 --> 00:47:03.123

episode of TeacherCast.

00:47:03.143 --> 00:47:04.304

On behalf of Dr. Lammers and

00:47:04.405 --> 00:47:05.585

everybody here on TeacherCast,

00:47:06.146 --> 00:47:07.226

my name is Jeff Bradbury,

00:47:07.407 --> 00:47:08.186

reminding you guys to keep

00:47:08.226 --> 00:47:08.967

up the great work in your

00:47:08.987 --> 00:47:10.148

classrooms and continue

00:47:10.168 --> 00:47:11.929

sharing your passions with your students.