WEBVTT
00:00:01.120 --> 00:00:04.160
Hello, and welcome to the Designing with Love podcast.
00:00:04.160 --> 00:00:11.759
I am your host, Jackie Pelegrin, where my goal is to bring you information, tips, and tricks as an instructional designer.
00:00:11.759 --> 00:00:17.039
Hello, instructional designers and educators.
00:00:17.039 --> 00:00:21.359
Welcome to episode 94 of the Designing with Love Podcast.
00:00:21.359 --> 00:00:24.480
I'm thrilled to have Sairam Sundaresan with me today.
00:00:24.480 --> 00:00:29.519
Cyram is an AI engineering leader and the author of AI for the rest of us.
00:00:29.519 --> 00:00:38.719
With his passion for making AI accessible and practical, he brings a valuable perspective on how this technology can be understood and applied by everyone.
00:00:38.719 --> 00:00:40.719
Welcome to the show, Cyram.
00:00:41.920 --> 00:00:43.200
Thanks for having me, Jackie.
00:00:43.200 --> 00:00:44.159
It's a pleasure to be here.
00:00:44.640 --> 00:00:45.520
Yes, thank you.
00:00:45.520 --> 00:00:52.240
And I usually like to give a shout out to Podmatch when I uh when I have guests that come on that have been matched through Podmatch.
00:00:52.240 --> 00:00:59.039
So I'm glad we got connected and had a chance to have this wonderful and engaging conversation today.
00:00:59.439 --> 00:01:00.719
Yeah, thank you, Podmatch.
00:01:00.719 --> 00:01:03.520
Yeah, excited to talk about some cool topics.
00:01:03.920 --> 00:01:05.040
Great, yes.
00:01:05.040 --> 00:01:11.359
So to start, can you tell us a little bit about yourself and share what inspired you to focus on AI engineering?
00:01:11.920 --> 00:01:12.319
Sure.
00:01:12.319 --> 00:01:16.000
Um so I am an AI engineering leader.
00:01:16.000 --> 00:01:25.280
I currently lead a team of engineers who are trying to help cars um drive safely and uh park safely by themselves.
00:01:25.280 --> 00:01:38.879
Um and prior to this, I was working um in both the data center and uh mobile phone um industries, and I was working on AI-related topics there as well.
00:01:38.879 --> 00:01:46.000
So I've sort of spanned uh 15 years in AI, which makes me sound like a dinosaur considering how fast AI is moving.
00:01:46.000 --> 00:01:57.359
But uh that being said, it's given me the opportunity to see several inflection points in this amazing technology and also work with some of the brightest minds in the world on this.
00:01:57.359 --> 00:02:02.959
And uh what's got me into the field was actually my passion for photography.
00:02:02.959 --> 00:02:08.879
Um when I was a kid, I was uh taking a ton of pictures with my uh parents' camera.
00:02:08.879 --> 00:02:15.360
It was a film camera at the time, and I used to love manipulating the images um in the lab.
00:02:15.360 --> 00:02:22.560
But what I realized along the way of that journey is that I could use those images to teach computers to understand the world.
00:02:22.560 --> 00:02:27.360
So I found that there was this entire field that focused on this particular topic, and I was hooked.
00:02:27.360 --> 00:02:32.000
And since then, I've spent every moment uh trying to learn as much as I can.
00:02:32.000 --> 00:02:37.199
And that's how I've gotten into AI, and that's how I continue to learn.
00:02:37.840 --> 00:02:38.319
That's great.
00:02:38.319 --> 00:02:48.879
So you didn't expect to actually go into this field, and it it was a love of another another subject in another area of photography that led you to do this.
00:02:48.879 --> 00:02:50.000
So that's really great.
00:02:50.000 --> 00:02:50.639
Wow.
00:02:50.639 --> 00:03:03.680
I I love that when you be kind of you go into it accidentally or it falls into your lap and you don't realize, wow, this is something I really love and I want to continue doing it and uh and advocate for the for the field, right?
00:03:03.680 --> 00:03:14.879
So it sounds like that's what you're doing too, because you work with a a team of engineers and you're able to help them hone their craft and their skills and help move that this field forward.
00:03:14.879 --> 00:03:16.240
So that's that's really great.
00:03:16.240 --> 00:03:20.560
Sounds like you're advocating for the profession and for the industry as well.
00:03:20.560 --> 00:03:21.599
So that I love that.
00:03:21.840 --> 00:03:23.680
Um I I really appreciate that.
00:03:23.680 --> 00:03:31.599
And and what I like to uh tell people is that by day I'm teaching machines to understand the world, and by night I'm teaching people to understand machines.
00:03:31.599 --> 00:03:32.960
So Oh, wow.
00:03:33.199 --> 00:03:35.039
So it's a cylindric process, right?
00:03:35.039 --> 00:03:35.280
Yeah.
00:03:36.080 --> 00:03:37.759
It is a cyclical process, yeah.
00:03:38.159 --> 00:03:39.759
Wow, that's great.
00:03:39.759 --> 00:03:51.919
So um in talking about your book, as I mentioned earlier in the introduction, you've written AI for the rest of us with the goal of making this technology approachable because it can be a little bit uh scary at times, right?
00:03:51.919 --> 00:03:58.879
So for those who might feel overwhelmed by AI, what are some of the biggest misconceptions and how do you help people move past them?
00:03:59.599 --> 00:04:03.680
So there are a ton of misconceptions, and I'll probably try to stick to a few.
00:04:03.680 --> 00:04:06.800
Um the first thing is that AI is going to solve everything.
00:04:06.800 --> 00:04:08.400
Um, it's not.
00:04:08.400 --> 00:04:16.079
And at the moment, um the the if you think about artificial intelligence, there's like two broad classifications.
00:04:16.079 --> 00:04:21.120
There's artificial narrow intelligence uh and artificial general intelligence.
00:04:21.120 --> 00:04:36.560
The latter is what people are concerned about, and that's something where uh an AI can do um all kinds of tasks at a level that is uh you know at least human level or significantly higher.
00:04:36.560 --> 00:04:39.199
And m my honest take is we are not there yet.
00:04:39.199 --> 00:04:43.920
And I uh and in full follow-up of people might be asking, oh, when is AGI coming?
00:04:43.920 --> 00:04:45.279
Should we be worried?
00:04:45.279 --> 00:04:46.319
Um, I don't know.
00:04:46.319 --> 00:04:55.839
And the honest answer is there's a lot of smart people looking into um how AGI might come about, and I don't have a timeline for it.
00:04:55.839 --> 00:05:06.720
Now coming back to artificial narrow intelligence, most of if not all of the things that we see in modern AI is you know is going to fall into this category.
00:05:06.720 --> 00:05:15.120
It this means that these are AI um models that are good at one specific task and are insanely good at it.
00:05:15.120 --> 00:05:22.319
And uh, for example, your um chatbot is really good at giving you responses, it understands your question.
00:05:22.319 --> 00:05:31.120
Quote aircotes understands your question and responds and responds to you in a way that you know feels like uh it's getting what you're saying.
00:05:31.120 --> 00:05:46.079
Um your um smart speaker at home listens to your voice, is able to decipher what you say, and then reply to you with either an action, like you know, turn on the lights or you know, play my favorite song, something like that.
00:05:46.079 --> 00:05:49.519
So these are things that are good at one specific task.
00:05:49.519 --> 00:05:59.279
And with AI, the idea right now is if the task is very well defined, is narrow and well scoped out, chances are AI will be very good at it.
00:05:59.279 --> 00:06:05.040
If the task is very broad, generic, and not well scoped out, AI won't be as good.
00:06:05.040 --> 00:06:13.600
And the misconception I see a lot of people having is AI is going to, you know, I I'll throw AI at any problem and it will solve it.
00:06:13.600 --> 00:06:21.040
It might, but chances are unless you are very particular with the constraints, it's not going to do a good job.
00:06:21.040 --> 00:06:23.839
So that's probably the biggest misconception.
00:06:23.839 --> 00:06:29.839
And happy to pause here and maybe talk about uh other things if you have a follow-up.
00:06:30.480 --> 00:06:32.399
Yeah, wow, that's amazing.
00:06:32.399 --> 00:06:35.600
Because I think some people just have this misconception I agree.
00:06:35.600 --> 00:06:42.959
People have a misconception that they can just give it anything, give it any prompt, and it'll just give them what they want, right?
00:06:42.959 --> 00:06:44.480
And it'll give them the answer.
00:06:44.480 --> 00:06:46.000
But that's not always the case.
00:06:46.000 --> 00:06:47.519
So yeah, that's interesting.
00:06:47.519 --> 00:06:52.000
Because we have a uh where I work, we have a closed system model.
00:06:52.000 --> 00:06:59.680
It's it's our own LMM that they created because we have proprietary information that's has curriculum and policies and procedures.
00:06:59.680 --> 00:07:10.319
And I've been actually working with this model and it has chatbots within it, but I've been working for it, Cyram, for about three or four days trying to get something specific.
00:07:10.319 --> 00:07:13.199
And I I was very specific in my prompt.
00:07:13.199 --> 00:07:15.279
Like I said, you're an instructional designer.
00:07:15.279 --> 00:07:16.399
This is what I'm trying to do.
00:07:16.399 --> 00:07:19.439
And what I'm trying to do is get a curriculum map.
00:07:19.439 --> 00:07:21.360
And I'm trying to get a downloadable.
00:07:21.360 --> 00:07:26.720
I tried an Excel document before, and I tried uploading all the syllabi documents.
00:07:26.720 --> 00:07:28.720
And there's just so many in this program.
00:07:28.720 --> 00:07:32.879
I think it it I don't think the system is able to handle it right now.
00:07:32.879 --> 00:07:38.800
So I tried copying and pasting it from the syllabus in there and just putting the text in there.
00:07:38.800 --> 00:07:43.519
And it seemed to work better, but uh still after three days, I'm still not getting an output.
00:07:43.519 --> 00:07:46.319
And I keep asking it, are you how's it coming along?
00:07:46.319 --> 00:07:48.079
And it's like, oh, I'm doing this.
00:07:48.079 --> 00:07:51.519
And I'm and I'm like, okay, is it am I really gonna get an output?
00:07:51.519 --> 00:07:59.120
And so even today, before I got off work, before I logged off, I've checked with uh with it and I said, How's it coming along?
00:07:59.120 --> 00:08:00.720
Oh, I'll get it to you by the end of the day.
00:08:00.720 --> 00:08:03.120
And I'm like, still don't have it.
00:08:03.120 --> 00:08:12.160
So I'm so I've tried different tried tried different formats and I I've tried different strategies and I'm still I'm get stuck at a certain point.
00:08:12.160 --> 00:08:13.759
So it's very interesting.
00:08:13.759 --> 00:08:18.160
I'm like, well, maybe the technology is just not caught up to that yet.
00:08:18.160 --> 00:08:25.839
And maybe they, you know, they need to our tech team needs to finesse this LLM model a little bit more.
00:08:26.480 --> 00:08:37.120
So that's that's where the the interesting bit is because um I feel like with AI, especially the use cases, we are kind of writing the user manual as we go.
00:08:37.120 --> 00:08:40.080
Um so there isn't a right way to do things or a wrong way.
00:08:40.080 --> 00:08:55.360
That I mean, there are just you know, there are best practices around, but um there isn't like a a recipe that if you copy paste it from one model to another or um one interface to another, it just works out of the box.
00:08:55.360 --> 00:09:01.279
And that's kind of where spending time with these tools and seeing how they respond to smaller tasks.
00:09:01.279 --> 00:09:11.759
Uh when when you get past the initial excitement where you you give it a toy task and it does spectacularly well, you're you're you feel like, oh wow, this is going to completely change the way I work.
00:09:11.759 --> 00:09:23.120
Uh then you hit this uh wall of frustration because you try to give it a little more uh uh you know tasks that are uh you know complicated and then it doesn't give you the response you want.
00:09:23.120 --> 00:09:31.440
And then you try uh finessing your prompt, you try to go back and forth, and there's a lot of trial and error that comes into the picture.
00:09:31.440 --> 00:09:38.480
So this is where spending time with these tools and then figuring out what type of response, like how do you break up the response?
00:09:38.480 --> 00:09:45.519
And generally, in especially with chatbots, uh the I have a rule of thumb which is never take the first response.
00:09:45.519 --> 00:10:01.360
Um and the way that I think about it is if you had a new employee join your team or an intern join your team, they're joining your team without any context about the role, about the policies, about the where they can find things and how they need to do their job.
00:10:01.360 --> 00:10:15.519
And if you are asking them to do a task for you, uh their first attempt will not be anywhere close to what you expect uh or what is usually the the standard of the results, usually, right?
00:10:15.519 --> 00:10:21.360
Unless this is an exceptional employee who's uh yeah you know knocking everything out of the park.
00:10:21.360 --> 00:10:27.440
Uh so then how do you get this employee uh to ramp up and then support you?
00:10:27.440 --> 00:10:44.720
You actually give them examples, you show them how it's done, like this is how I do it, here's where you can find these documents, this is what's inside these documents, here's how you connect the dots, here's the task I want, here's the input, and this is what I expect as a result.
00:10:44.720 --> 00:10:50.080
And when you do this, you notice that the the results they produce are closer to what you expect.
00:10:50.080 --> 00:10:52.720
And the same is true for any of these AI models.
00:10:52.720 --> 00:10:57.440
Think of them like an intern or a new employee that you are onboarding onto your team.
00:10:57.440 --> 00:11:16.879
And every chat, if you treat it in that way, it becomes easier because yes, these models have this um notion of memory where you know it it it kind of knows your preferences or like based on your past interactions, it can say, Oh, Jackie preferred this type of output or the did this type of task.
00:11:16.879 --> 00:11:21.200
So I'm guessing she would um you know expect it in this format.
00:11:21.200 --> 00:11:36.639
But uh my take is always you know, think of it like you're onboarding a new colleague and be easy on them, give them as many instructions as you can, as specific instructions as you can, and then go back and forth.
00:11:36.639 --> 00:11:41.039
So that's that's when you'll get more uh useful outputs.
00:11:41.440 --> 00:11:42.000
I like that.
00:11:42.000 --> 00:11:42.639
That's great.
00:11:42.639 --> 00:11:45.600
And that's something I've been practicing too using that.
00:11:45.600 --> 00:11:57.039
And I actually just upgraded with ChatGPT, I was using the free version of it, and I was using it so much that I would, of course, uh hit my limit every day.
00:11:57.039 --> 00:12:01.200
So I was like, oh no, I'm hitting my limit, and then it would go to the lower model, right?
00:12:01.200 --> 00:12:02.399
After it hits that limit.
00:12:02.399 --> 00:12:07.039
So I decided to take the plunge and and get the $20 a month paid one.
00:12:07.039 --> 00:12:10.639
And I would notice such a huge difference between the free and that.
00:12:10.639 --> 00:12:12.799
And uh it's amazing.
00:12:12.799 --> 00:12:30.000
So there was, I think it was the other two days ago, I spent about two hours, two and a half hours on with the tool with Chat GPT, and I was refining some things and working on some materials, and I just kept going back and forth with the model and just saying, okay, yeah, you kind of got that there.
00:12:30.000 --> 00:12:35.840
But, you know, maybe we could try this or, you know, repurpose this and and just work through it.
00:12:35.840 --> 00:12:39.519
And so it really was nice because, oh yeah, you got a good point there.
00:12:39.519 --> 00:12:49.679
And so, you know, I just kind of like you said, just worked with it as if it was my assistant or, you know, or a new employee, as if it doesn't know some of that content.
00:12:49.679 --> 00:12:51.679
And so yeah, it was nice.
00:12:51.679 --> 00:12:57.840
And it kind of encourages you on along the way as long as you're kind of giving it that encouragement too.
00:12:57.840 --> 00:13:00.480
It will it will respond in that way, right?
00:13:00.480 --> 00:13:07.440
So it's it's neat to kind of have that experience and go, okay, I I know this is a machine, but this is pretty cool.
00:13:07.440 --> 00:13:08.399
This is neat.
00:13:08.399 --> 00:13:16.960
It's it's it's it's kind of a fun experience too, to kind of experiment with it and see how where you can go and what you can get from it.
00:13:16.960 --> 00:13:18.559
So that's a lot of fun.
00:13:18.559 --> 00:13:20.000
Yeah, I love that.
00:13:20.000 --> 00:13:20.879
That's great.
00:13:20.879 --> 00:13:27.440
So many of my listeners are educators and instructional designers who are curious about how AI can support their work.
00:13:27.440 --> 00:13:34.480
From your perspective, what are some practical ways AI can enhance teaching learning or the design of learning experiences?
00:14:12.129 --> 00:14:17.250
Um, this is a you know a gold mine of a topic, and there's uh a lot to dive into this.
00:14:17.250 --> 00:14:23.330
Uh so I I I'm I'm I I don't come from an educational background, although I'm a passionate educator myself.
00:14:23.330 --> 00:14:26.450
Uh so here's let's let's let's see.
00:14:26.450 --> 00:14:37.090
Um, one of the things that you immediately recognize is that with learners and students in general, no two uh students or learners are the same.
00:14:37.090 --> 00:14:46.210
They have different needs, they maybe learn better with different media, uh, they may be at different points in their learning experience and journey.
00:14:46.210 --> 00:14:57.410
And uh therefore, if you design a single curriculum that is going to cater to each of these learners, it may not land as effectively.
00:14:57.410 --> 00:15:12.530
And uh the other thing is there are as teachers, you probably have uh you know finite time interacting with these students during the course of a term or um if it's an online course for the duration of the course.
00:15:12.530 --> 00:15:24.129
And the other challenge is for the teacher, it's a lot of cognitive load because one learner might get things really quickly and they may ask you for advanced material.
00:15:24.129 --> 00:15:31.410
They they probably might ask you if uh you have what like what's next, and how do I take this to the next level?
00:15:31.410 --> 00:15:36.290
And then you have another learner who's probably um not there yet and trying to catch up.
00:15:36.290 --> 00:15:44.610
And for them, you may have to invest a ton of time and energy trying to explain how something works or how they can um understand something better.
00:15:44.610 --> 00:16:11.170
And uh if you now add AI, which is um a 24-7 private tutor that's available that can tailor the learning experiences based on the current level of a learner and based on the media that they learn best from, you suddenly have this tool that uh scales education for these teachers, right?
00:16:11.170 --> 00:16:22.210
Because all of a sudden they can ensure that each student gets the care that they need without themselves, with the without these teachers sort of stretching themselves too thin.
00:16:22.210 --> 00:16:33.649
You um get courses that you could have custom quizzes designed by these AI models based on the material that are tailored to the learner's weak spots.
00:16:33.649 --> 00:16:40.129
Because you can it's not something where you have a test and then you know have a test a few weeks later.
00:16:40.129 --> 00:16:43.170
If if you want, you could have quizzes every day.
00:16:43.170 --> 00:16:52.610
Um not saying learners should have quizzes every day, but my point is that if they have uh certain areas where they're having a bit of trouble with, you can spend more time on that.
00:16:52.610 --> 00:16:56.050
And this can be done outside of the classroom environment.
00:16:56.050 --> 00:17:15.410
So all of a sudden, learners now have access to a infinitely patient 24-7, 365-day private tutor who is able to probe and ask and teach at a level that is um sort of helping them learn where they are.
00:17:15.410 --> 00:17:19.970
And for the teacher, it allows them to then design experiences around that.
00:17:19.970 --> 00:17:24.529
So they have now a TA that's available to help.
00:17:24.529 --> 00:17:36.369
And and for designers, uh all of a sudden, you can get instant feedback on like if you have slides for a particular course, you can get feedback on where the students in general didn't get the material.
00:17:36.369 --> 00:17:39.329
So maybe some slide needs to be redesigned, for example.
00:17:39.329 --> 00:17:45.250
Or you can think about um like is there a particular topic that was not covered in depth?
00:17:45.250 --> 00:17:46.929
And was there a demand?
00:17:46.929 --> 00:17:51.490
And you can have a turnaround of a week as opposed to a semester.
00:17:51.490 --> 00:17:55.569
And I feel like that's opening up entire new opportunities in education.
00:17:55.569 --> 00:17:56.769
But I'll stop teaching.
00:17:57.409 --> 00:17:57.809
Yeah.
00:17:57.809 --> 00:17:59.809
Well, that's wonderful, Saram.
00:17:59.809 --> 00:18:06.209
I I love that because I'm a I I work in curriculum design and development, uh, my daytime job.
00:18:06.209 --> 00:18:10.529
And then in the evening, I teach online classes in instructional design.
00:18:10.529 --> 00:18:18.369
So hearing that perspective is great and it reinforces what I'm already doing as an educator and as an instructional designer.
00:18:18.369 --> 00:18:19.409
So it's fabulous.
00:18:19.409 --> 00:18:29.089
And I I really hope that um those that I that work with me listen to this and also my students, because a lot of them are educators, a lot of them are teachers.
00:18:29.089 --> 00:18:37.169
So being able to tap into this technology, I think, like you said, it personalizes the learning and it allows us to make changes on the fly.
00:18:37.169 --> 00:18:41.329
So, like you said, we don't have to wait a whole semester or anything like that.
00:18:41.329 --> 00:18:52.209
Um, because sometimes when we wait too long for this, the curriculum, it's that you have you're trying to play catch up and then you feel like your students are already behind the curveball.
00:18:52.209 --> 00:19:02.289
So it's great when you can have that opportunity to make sure that that curriculum is up to date and it's uh it's applicable to what they're going to learn on the job.
00:19:02.289 --> 00:19:03.490
So that's great.
00:19:03.490 --> 00:19:10.929
Especially in fields like I would think healthcare, like nursing and um, you know, like what you do, engineering.
00:19:10.929 --> 00:19:25.889
We have Grand Canyon University has engineering programs and the science science and engineering are really big areas at the university, but also education is is really big right now and and so uh counseling in those areas.
00:19:25.889 --> 00:19:38.929
And so trying to help them to understand that AI can come alongside them and be their, like you said, their tutor or their assistant and not look at it as uh as something that uh they should be scared of.
00:19:38.929 --> 00:19:39.409
So yeah.
00:19:39.409 --> 00:19:40.449
Right, exactly.
00:19:40.449 --> 00:19:40.689
Right.
00:19:40.769 --> 00:19:56.689
So if you if you think about what teachers can do best, um they they can look at students and see how they're responding to uh material or how they're responding to certain topics in conversation, and you can see it in their body language, their eyes, their face.
00:19:56.689 --> 00:20:06.449
Um, and you can see when somebody is motivated or is getting what is being said versus somebody who is a little uh, you know, maybe they they don't get the material at first go.
00:20:06.449 --> 00:20:10.849
And that is a qualitative signal that only teachers can get.
00:20:10.849 --> 00:20:23.889
And on the other hand, the quantitative side comes from these AI tutors because they know how many times a student had to retake a quiz, exactly um which questions they tripped up in and where they need extra support.
00:20:23.889 --> 00:20:31.250
And you all of a sudden you have both sides of the coin, which allows you to design and develop a holistic learning experience.
00:20:31.250 --> 00:20:41.809
So for me, I'm just so excited about what is available for education in this day and age versus what was available um back in the day.
00:20:42.129 --> 00:20:43.250
Right, exactly.
00:20:43.250 --> 00:20:47.009
You know, one thing you mentioned, Saran, was uh cognitive load.
00:20:47.009 --> 00:20:54.609
And that's something in our in my field in education, but especially in instructional design, that we're always trying to keep at the forefront.
00:20:54.609 --> 00:21:03.089
And I always remind my students about that because it's it's based on cognitive psychology that our brain, our working memory can only handle so much information.
00:21:03.089 --> 00:21:16.449
And so if we put too much on the students, if we info dump on them, uh, you know, and the technology is too much for them, you know, if if that's a barrier to them, then they're gonna they're gonna struggle, right?
00:21:16.529 --> 00:21:18.689
And they're not gonna they're not gonna get that.
00:21:18.849 --> 00:21:21.250
So we don't want to put any barriers in front of them.
00:21:21.250 --> 00:21:31.329
We want to make sure that it's accessible for them and that they feel like they're uh they're not having to work around things or anything like that to try to get to what they need to.
00:21:31.329 --> 00:21:32.209
So yeah.
00:21:32.209 --> 00:21:35.329
So it's nice when these tools um can help them with that.
00:21:35.329 --> 00:21:47.409
And like you said, then the the teachers and the educators can focus on uh what they need to do to to help that learning move forward and to help them have those moments where they go, oh, I understand that now.
00:21:47.409 --> 00:21:55.970
And now I can take this into the workplace and I can apply it because the program I teach for, they they do project, it's project-based learning, which is really great.
00:21:55.970 --> 00:22:03.649
I love that because they're able to they don't there's only maybe one or two papers in the whole entire program of the courses I teach.
00:22:03.649 --> 00:22:09.169
All of the rest of the assignments are all project-based where they have to literally do this.
00:22:09.169 --> 00:22:15.169
And I have a course I'm teaching right now, you'd love this, where they they have to use an AI tool twice.
00:22:15.169 --> 00:22:23.409
So they have one one assignment that's about situational leadership and they have to and they have these prompts for these, uh, for the four readiness styles.
00:22:23.409 --> 00:22:30.289
And they have to take the prompt, put it into an AI tool and converse with that tool and get that individual.
00:22:30.289 --> 00:22:36.209
So the AI tool like Chat GPT plays the employee and then they play the instructional designer.
00:22:36.209 --> 00:22:46.929
And so they have this conversation with them, working with them as the leader and trying to get them to move from one leadership style to the to the next or one situational model to the next.
00:22:46.929 --> 00:22:49.409
So it's very interesting to do that.
00:22:49.409 --> 00:23:03.089
And then another one they did that's that I'm grading this weekend is where they had to build an e-learning module on um uh it's it was for like a fake tech company uh about their policies and procedures.
00:23:03.089 --> 00:23:08.929
And so they have to use the tool to help kind of build that content and re refine it.
00:23:08.929 --> 00:23:23.009
So um, but it's funny because some of my students um they they feel like, and I think even educators do too, but some of my students actually told me in messages and in their reflection afterwards that they felt like they were cheating.
00:23:23.009 --> 00:23:24.369
And I was like, really?
00:23:24.369 --> 00:23:25.089
That's surprising.
00:23:25.089 --> 00:23:28.689
Yeah, exactly.
00:23:28.689 --> 00:23:33.889
And it's like, hmm, you know, but yet educators, they're like, they think in the back of their mind.
00:23:33.889 --> 00:23:42.209
So many of them, I've been to webinars where they think in the back of their mind and they'll type it in, they'll be like, how do I know if my students' DQ responses are genuine or not?
00:23:42.209 --> 00:23:44.449
How do I know if they're not just utilizing AI?
00:23:44.449 --> 00:23:48.209
And I'm like, well, yeah, yeah, exactly.
00:23:48.209 --> 00:23:55.009
And I was like, but before, and I thought to myself, to and I put some responses in there, and I said, Well, what happened before AI?
00:23:55.009 --> 00:23:56.769
They found ways to cheat, you know.
00:23:56.769 --> 00:24:01.329
So if they're if students are gonna cheat, they're gonna cheat no matter what, whether AI is there or not.
00:24:01.329 --> 00:24:04.609
You know, if you remove the tool, they'll they'll find ways to do it.
00:24:04.609 --> 00:24:13.329
I mean, we I could search quiz questions for the classes that I help design, and you can find them online, or you go to Course Euro and you can find them.
00:24:13.329 --> 00:24:16.769
So absent of AI, I think they can they can still do it.
00:24:16.769 --> 00:24:18.129
Exactly.
00:24:18.129 --> 00:24:26.369
I mean, they they've had Turnitin, Turnitin's been around for a long time, and institutions have used those to check for plagiarism, right?
00:24:26.369 --> 00:24:29.889
So yeah, so before AI was around, Turnitin was around.
00:24:29.889 --> 00:24:31.329
So it's funny.
00:24:31.329 --> 00:24:38.049
Um, but yeah, you know, my institution, they don't want to police the students, they want to make sure that they're using it ethically.
00:24:38.049 --> 00:24:40.289
So that's the the most important aspect.