March 1, 2026

AI Made Practical for Teachers and Designers with Sairam Sundaresan

AI Made Practical for Teachers and Designers with Sairam Sundaresan
Apple Podcasts podcast player iconSpotify podcast player iconYoutube Music podcast player iconAmazon Music podcast player iconiHeartRadio podcast player iconPodcast Addict podcast player iconPodchaser podcast player iconPocketCasts podcast player iconDeezer podcast player iconPlayerFM podcast player iconCastro podcast player iconCastbox podcast player iconGoodpods podcast player icon
Apple Podcasts podcast player iconSpotify podcast player iconYoutube Music podcast player iconAmazon Music podcast player iconiHeartRadio podcast player iconPodcast Addict podcast player iconPodchaser podcast player iconPocketCasts podcast player iconDeezer podcast player iconPlayerFM podcast player iconCastro podcast player iconCastbox podcast player iconGoodpods podcast player icon

Think AI can do everything? We put that assumption under the microscope with AI engineering leader and author Sairam Sundaresan, and walk away with a playbook that’s practical, ethical, and built for real classrooms and design teams. We break down why narrow, well-scoped tasks are where AI shines, how to turn prompting into a repeatable workflow, and what it looks like to treat a model like a new hire you’re onboarding—clear roles, examples, and tight feedback loops.

We dig into the big wins for educators and instructional designers: personalized learning at scale, faster feedback cycles, and smarter revisions between sessions. Imagine a 24/7 teaching assistant that adapts to your students’ levels, flags weak spots, and helps you adjust the curriculum without waiting for the next term. Pair that with your human superpowers—reading the room, motivating learners, and connecting dots—and you get a learning ecosystem that’s both efficient and deeply human. Along the way, we share hands-on tips with tools like Notebook LM, Canva, Gamma, and Genially to prototype content, translate assets, and build interactive experiences without months of overhead.

Enjoyed the episode? Follow the show, leave a quick review, and share it with a colleague who’s experimenting with AI in education. Your support helps more educators discover practical, ethical ways to use these tools.

🔗 Website and Social Links:

Please visit Sairam Sundaresan’s website to subscribe to his newsletter.

Sairam Sundaresan’s Website

📢 Call-to-Action: Want to explore AI in a way that feels clear and approachable? Connect with Sairam Sundaresan and check out his book AI For the Rest of Us. You’ll find practical insights, real-world examples, and guidance on how to use AI responsibly in work, learning, and life. Visit Sairam’s website to learn more and access resources designed to help you confidently navigate the AI era. 

Send Jackie a Text

Join PodMatch!
Use the link to join PodMatch, a place for hosts and guests to connect.

Disclaimer: This post contains affiliate links. If you make a purchase, I may receive a commission at no extra cost to you.

Support the show

💟 Designing with Love + allows you to support the show by keeping the mic on and the ideas flowing. Click on the link above to provide your support.

Buy Me a Coffee is another way you can support the show, either as a one-time gift or through a monthly subscription.

🗣️ Want to be a guest on Designing with Love? Send Jackie Pelegrin a message on PodMatch, here: Be a guest on the show

🌐 Check out the show's website here: Designing with Love

📱 Send a text to the show by clicking the Send Jackie a Text link above.

👍🏼 Please make sure to like and share this episode with others. Here's to great learning!


00:00 - Welcome & Guest Introduction

01:12 - Cyram’s Path from Photography to AI

03:52 - Narrow vs General AI: Clearing Myths

08:00 - Prompting Limits & Workflow Friction

12:30 - Treat AI Like an Intern

17:04 - Practical Uses in Teaching & Design

21:44 - Personalization, Feedback Loops, and Speed

26:24 - Ethics: Copyright, Hallucinations, Energy

30:49 - Misuse, Impersonation, and Guardrails

34:04 - Future of AI: Fast Prototyping & Tools

37:54 - Hands‑On Tips: Play, Iterate, Build Literacy

41:44 - Assessment Rethinks: Open‑Book with AI

45:44 - Final Advice & Closing Gratitude

46:14 - Support the Show & Ways to Help

WEBVTT

00:00:01.120 --> 00:00:04.160
Hello, and welcome to the Designing with Love podcast.

00:00:04.160 --> 00:00:11.759
I am your host, Jackie Pelegrin, where my goal is to bring you information, tips, and tricks as an instructional designer.

00:00:11.759 --> 00:00:17.039
Hello, instructional designers and educators.

00:00:17.039 --> 00:00:21.359
Welcome to episode 94 of the Designing with Love Podcast.

00:00:21.359 --> 00:00:24.480
I'm thrilled to have Sairam Sundaresan with me today.

00:00:24.480 --> 00:00:29.519
Cyram is an AI engineering leader and the author of AI for the rest of us.

00:00:29.519 --> 00:00:38.719
With his passion for making AI accessible and practical, he brings a valuable perspective on how this technology can be understood and applied by everyone.

00:00:38.719 --> 00:00:40.719
Welcome to the show, Cyram.

00:00:41.920 --> 00:00:43.200
Thanks for having me, Jackie.

00:00:43.200 --> 00:00:44.159
It's a pleasure to be here.

00:00:44.640 --> 00:00:45.520
Yes, thank you.

00:00:45.520 --> 00:00:52.240
And I usually like to give a shout out to Podmatch when I uh when I have guests that come on that have been matched through Podmatch.

00:00:52.240 --> 00:00:59.039
So I'm glad we got connected and had a chance to have this wonderful and engaging conversation today.

00:00:59.439 --> 00:01:00.719
Yeah, thank you, Podmatch.

00:01:00.719 --> 00:01:03.520
Yeah, excited to talk about some cool topics.

00:01:03.920 --> 00:01:05.040
Great, yes.

00:01:05.040 --> 00:01:11.359
So to start, can you tell us a little bit about yourself and share what inspired you to focus on AI engineering?

00:01:11.920 --> 00:01:12.319
Sure.

00:01:12.319 --> 00:01:16.000
Um so I am an AI engineering leader.

00:01:16.000 --> 00:01:25.280
I currently lead a team of engineers who are trying to help cars um drive safely and uh park safely by themselves.

00:01:25.280 --> 00:01:38.879
Um and prior to this, I was working um in both the data center and uh mobile phone um industries, and I was working on AI-related topics there as well.

00:01:38.879 --> 00:01:46.000
So I've sort of spanned uh 15 years in AI, which makes me sound like a dinosaur considering how fast AI is moving.

00:01:46.000 --> 00:01:57.359
But uh that being said, it's given me the opportunity to see several inflection points in this amazing technology and also work with some of the brightest minds in the world on this.

00:01:57.359 --> 00:02:02.959
And uh what's got me into the field was actually my passion for photography.

00:02:02.959 --> 00:02:08.879
Um when I was a kid, I was uh taking a ton of pictures with my uh parents' camera.

00:02:08.879 --> 00:02:15.360
It was a film camera at the time, and I used to love manipulating the images um in the lab.

00:02:15.360 --> 00:02:22.560
But what I realized along the way of that journey is that I could use those images to teach computers to understand the world.

00:02:22.560 --> 00:02:27.360
So I found that there was this entire field that focused on this particular topic, and I was hooked.

00:02:27.360 --> 00:02:32.000
And since then, I've spent every moment uh trying to learn as much as I can.

00:02:32.000 --> 00:02:37.199
And that's how I've gotten into AI, and that's how I continue to learn.

00:02:37.840 --> 00:02:38.319
That's great.

00:02:38.319 --> 00:02:48.879
So you didn't expect to actually go into this field, and it it was a love of another another subject in another area of photography that led you to do this.

00:02:48.879 --> 00:02:50.000
So that's really great.

00:02:50.000 --> 00:02:50.639
Wow.

00:02:50.639 --> 00:03:03.680
I I love that when you be kind of you go into it accidentally or it falls into your lap and you don't realize, wow, this is something I really love and I want to continue doing it and uh and advocate for the for the field, right?

00:03:03.680 --> 00:03:14.879
So it sounds like that's what you're doing too, because you work with a a team of engineers and you're able to help them hone their craft and their skills and help move that this field forward.

00:03:14.879 --> 00:03:16.240
So that's that's really great.

00:03:16.240 --> 00:03:20.560
Sounds like you're advocating for the profession and for the industry as well.

00:03:20.560 --> 00:03:21.599
So that I love that.

00:03:21.840 --> 00:03:23.680
Um I I really appreciate that.

00:03:23.680 --> 00:03:31.599
And and what I like to uh tell people is that by day I'm teaching machines to understand the world, and by night I'm teaching people to understand machines.

00:03:31.599 --> 00:03:32.960
So Oh, wow.

00:03:33.199 --> 00:03:35.039
So it's a cylindric process, right?

00:03:35.039 --> 00:03:35.280
Yeah.

00:03:36.080 --> 00:03:37.759
It is a cyclical process, yeah.

00:03:38.159 --> 00:03:39.759
Wow, that's great.

00:03:39.759 --> 00:03:51.919
So um in talking about your book, as I mentioned earlier in the introduction, you've written AI for the rest of us with the goal of making this technology approachable because it can be a little bit uh scary at times, right?

00:03:51.919 --> 00:03:58.879
So for those who might feel overwhelmed by AI, what are some of the biggest misconceptions and how do you help people move past them?

00:03:59.599 --> 00:04:03.680
So there are a ton of misconceptions, and I'll probably try to stick to a few.

00:04:03.680 --> 00:04:06.800
Um the first thing is that AI is going to solve everything.

00:04:06.800 --> 00:04:08.400
Um, it's not.

00:04:08.400 --> 00:04:16.079
And at the moment, um the the if you think about artificial intelligence, there's like two broad classifications.

00:04:16.079 --> 00:04:21.120
There's artificial narrow intelligence uh and artificial general intelligence.

00:04:21.120 --> 00:04:36.560
The latter is what people are concerned about, and that's something where uh an AI can do um all kinds of tasks at a level that is uh you know at least human level or significantly higher.

00:04:36.560 --> 00:04:39.199
And m my honest take is we are not there yet.

00:04:39.199 --> 00:04:43.920
And I uh and in full follow-up of people might be asking, oh, when is AGI coming?

00:04:43.920 --> 00:04:45.279
Should we be worried?

00:04:45.279 --> 00:04:46.319
Um, I don't know.

00:04:46.319 --> 00:04:55.839
And the honest answer is there's a lot of smart people looking into um how AGI might come about, and I don't have a timeline for it.

00:04:55.839 --> 00:05:06.720
Now coming back to artificial narrow intelligence, most of if not all of the things that we see in modern AI is you know is going to fall into this category.

00:05:06.720 --> 00:05:15.120
It this means that these are AI um models that are good at one specific task and are insanely good at it.

00:05:15.120 --> 00:05:22.319
And uh, for example, your um chatbot is really good at giving you responses, it understands your question.

00:05:22.319 --> 00:05:31.120
Quote aircotes understands your question and responds and responds to you in a way that you know feels like uh it's getting what you're saying.

00:05:31.120 --> 00:05:46.079
Um your um smart speaker at home listens to your voice, is able to decipher what you say, and then reply to you with either an action, like you know, turn on the lights or you know, play my favorite song, something like that.

00:05:46.079 --> 00:05:49.519
So these are things that are good at one specific task.

00:05:49.519 --> 00:05:59.279
And with AI, the idea right now is if the task is very well defined, is narrow and well scoped out, chances are AI will be very good at it.

00:05:59.279 --> 00:06:05.040
If the task is very broad, generic, and not well scoped out, AI won't be as good.

00:06:05.040 --> 00:06:13.600
And the misconception I see a lot of people having is AI is going to, you know, I I'll throw AI at any problem and it will solve it.

00:06:13.600 --> 00:06:21.040
It might, but chances are unless you are very particular with the constraints, it's not going to do a good job.

00:06:21.040 --> 00:06:23.839
So that's probably the biggest misconception.

00:06:23.839 --> 00:06:29.839
And happy to pause here and maybe talk about uh other things if you have a follow-up.

00:06:30.480 --> 00:06:32.399
Yeah, wow, that's amazing.

00:06:32.399 --> 00:06:35.600
Because I think some people just have this misconception I agree.

00:06:35.600 --> 00:06:42.959
People have a misconception that they can just give it anything, give it any prompt, and it'll just give them what they want, right?

00:06:42.959 --> 00:06:44.480
And it'll give them the answer.

00:06:44.480 --> 00:06:46.000
But that's not always the case.

00:06:46.000 --> 00:06:47.519
So yeah, that's interesting.

00:06:47.519 --> 00:06:52.000
Because we have a uh where I work, we have a closed system model.

00:06:52.000 --> 00:06:59.680
It's it's our own LMM that they created because we have proprietary information that's has curriculum and policies and procedures.

00:06:59.680 --> 00:07:10.319
And I've been actually working with this model and it has chatbots within it, but I've been working for it, Cyram, for about three or four days trying to get something specific.

00:07:10.319 --> 00:07:13.199
And I I was very specific in my prompt.

00:07:13.199 --> 00:07:15.279
Like I said, you're an instructional designer.

00:07:15.279 --> 00:07:16.399
This is what I'm trying to do.

00:07:16.399 --> 00:07:19.439
And what I'm trying to do is get a curriculum map.

00:07:19.439 --> 00:07:21.360
And I'm trying to get a downloadable.

00:07:21.360 --> 00:07:26.720
I tried an Excel document before, and I tried uploading all the syllabi documents.

00:07:26.720 --> 00:07:28.720
And there's just so many in this program.

00:07:28.720 --> 00:07:32.879
I think it it I don't think the system is able to handle it right now.

00:07:32.879 --> 00:07:38.800
So I tried copying and pasting it from the syllabus in there and just putting the text in there.

00:07:38.800 --> 00:07:43.519
And it seemed to work better, but uh still after three days, I'm still not getting an output.

00:07:43.519 --> 00:07:46.319
And I keep asking it, are you how's it coming along?

00:07:46.319 --> 00:07:48.079
And it's like, oh, I'm doing this.

00:07:48.079 --> 00:07:51.519
And I'm and I'm like, okay, is it am I really gonna get an output?

00:07:51.519 --> 00:07:59.120
And so even today, before I got off work, before I logged off, I've checked with uh with it and I said, How's it coming along?

00:07:59.120 --> 00:08:00.720
Oh, I'll get it to you by the end of the day.

00:08:00.720 --> 00:08:03.120
And I'm like, still don't have it.

00:08:03.120 --> 00:08:12.160
So I'm so I've tried different tried tried different formats and I I've tried different strategies and I'm still I'm get stuck at a certain point.

00:08:12.160 --> 00:08:13.759
So it's very interesting.

00:08:13.759 --> 00:08:18.160
I'm like, well, maybe the technology is just not caught up to that yet.

00:08:18.160 --> 00:08:25.839
And maybe they, you know, they need to our tech team needs to finesse this LLM model a little bit more.

00:08:26.480 --> 00:08:37.120
So that's that's where the the interesting bit is because um I feel like with AI, especially the use cases, we are kind of writing the user manual as we go.

00:08:37.120 --> 00:08:40.080
Um so there isn't a right way to do things or a wrong way.

00:08:40.080 --> 00:08:55.360
That I mean, there are just you know, there are best practices around, but um there isn't like a a recipe that if you copy paste it from one model to another or um one interface to another, it just works out of the box.

00:08:55.360 --> 00:09:01.279
And that's kind of where spending time with these tools and seeing how they respond to smaller tasks.

00:09:01.279 --> 00:09:11.759
Uh when when you get past the initial excitement where you you give it a toy task and it does spectacularly well, you're you're you feel like, oh wow, this is going to completely change the way I work.

00:09:11.759 --> 00:09:23.120
Uh then you hit this uh wall of frustration because you try to give it a little more uh uh you know tasks that are uh you know complicated and then it doesn't give you the response you want.

00:09:23.120 --> 00:09:31.440
And then you try uh finessing your prompt, you try to go back and forth, and there's a lot of trial and error that comes into the picture.

00:09:31.440 --> 00:09:38.480
So this is where spending time with these tools and then figuring out what type of response, like how do you break up the response?

00:09:38.480 --> 00:09:45.519
And generally, in especially with chatbots, uh the I have a rule of thumb which is never take the first response.

00:09:45.519 --> 00:10:01.360
Um and the way that I think about it is if you had a new employee join your team or an intern join your team, they're joining your team without any context about the role, about the policies, about the where they can find things and how they need to do their job.

00:10:01.360 --> 00:10:15.519
And if you are asking them to do a task for you, uh their first attempt will not be anywhere close to what you expect uh or what is usually the the standard of the results, usually, right?

00:10:15.519 --> 00:10:21.360
Unless this is an exceptional employee who's uh yeah you know knocking everything out of the park.

00:10:21.360 --> 00:10:27.440
Uh so then how do you get this employee uh to ramp up and then support you?

00:10:27.440 --> 00:10:44.720
You actually give them examples, you show them how it's done, like this is how I do it, here's where you can find these documents, this is what's inside these documents, here's how you connect the dots, here's the task I want, here's the input, and this is what I expect as a result.

00:10:44.720 --> 00:10:50.080
And when you do this, you notice that the the results they produce are closer to what you expect.

00:10:50.080 --> 00:10:52.720
And the same is true for any of these AI models.

00:10:52.720 --> 00:10:57.440
Think of them like an intern or a new employee that you are onboarding onto your team.

00:10:57.440 --> 00:11:16.879
And every chat, if you treat it in that way, it becomes easier because yes, these models have this um notion of memory where you know it it it kind of knows your preferences or like based on your past interactions, it can say, Oh, Jackie preferred this type of output or the did this type of task.

00:11:16.879 --> 00:11:21.200
So I'm guessing she would um you know expect it in this format.

00:11:21.200 --> 00:11:36.639
But uh my take is always you know, think of it like you're onboarding a new colleague and be easy on them, give them as many instructions as you can, as specific instructions as you can, and then go back and forth.

00:11:36.639 --> 00:11:41.039
So that's that's when you'll get more uh useful outputs.

00:11:41.440 --> 00:11:42.000
I like that.

00:11:42.000 --> 00:11:42.639
That's great.

00:11:42.639 --> 00:11:45.600
And that's something I've been practicing too using that.

00:11:45.600 --> 00:11:57.039
And I actually just upgraded with ChatGPT, I was using the free version of it, and I was using it so much that I would, of course, uh hit my limit every day.

00:11:57.039 --> 00:12:01.200
So I was like, oh no, I'm hitting my limit, and then it would go to the lower model, right?

00:12:01.200 --> 00:12:02.399
After it hits that limit.

00:12:02.399 --> 00:12:07.039
So I decided to take the plunge and and get the $20 a month paid one.

00:12:07.039 --> 00:12:10.639
And I would notice such a huge difference between the free and that.

00:12:10.639 --> 00:12:12.799
And uh it's amazing.

00:12:12.799 --> 00:12:30.000
So there was, I think it was the other two days ago, I spent about two hours, two and a half hours on with the tool with Chat GPT, and I was refining some things and working on some materials, and I just kept going back and forth with the model and just saying, okay, yeah, you kind of got that there.

00:12:30.000 --> 00:12:35.840
But, you know, maybe we could try this or, you know, repurpose this and and just work through it.

00:12:35.840 --> 00:12:39.519
And so it really was nice because, oh yeah, you got a good point there.

00:12:39.519 --> 00:12:49.679
And so, you know, I just kind of like you said, just worked with it as if it was my assistant or, you know, or a new employee, as if it doesn't know some of that content.

00:12:49.679 --> 00:12:51.679
And so yeah, it was nice.

00:12:51.679 --> 00:12:57.840
And it kind of encourages you on along the way as long as you're kind of giving it that encouragement too.

00:12:57.840 --> 00:13:00.480
It will it will respond in that way, right?

00:13:00.480 --> 00:13:07.440
So it's it's neat to kind of have that experience and go, okay, I I know this is a machine, but this is pretty cool.

00:13:07.440 --> 00:13:08.399
This is neat.

00:13:08.399 --> 00:13:16.960
It's it's it's it's kind of a fun experience too, to kind of experiment with it and see how where you can go and what you can get from it.

00:13:16.960 --> 00:13:18.559
So that's a lot of fun.

00:13:18.559 --> 00:13:20.000
Yeah, I love that.

00:13:20.000 --> 00:13:20.879
That's great.

00:13:20.879 --> 00:13:27.440
So many of my listeners are educators and instructional designers who are curious about how AI can support their work.

00:13:27.440 --> 00:13:34.480
From your perspective, what are some practical ways AI can enhance teaching learning or the design of learning experiences?

00:14:12.129 --> 00:14:17.250
Um, this is a you know a gold mine of a topic, and there's uh a lot to dive into this.

00:14:17.250 --> 00:14:23.330
Uh so I I I'm I'm I I don't come from an educational background, although I'm a passionate educator myself.

00:14:23.330 --> 00:14:26.450
Uh so here's let's let's let's see.

00:14:26.450 --> 00:14:37.090
Um, one of the things that you immediately recognize is that with learners and students in general, no two uh students or learners are the same.

00:14:37.090 --> 00:14:46.210
They have different needs, they maybe learn better with different media, uh, they may be at different points in their learning experience and journey.

00:14:46.210 --> 00:14:57.410
And uh therefore, if you design a single curriculum that is going to cater to each of these learners, it may not land as effectively.

00:14:57.410 --> 00:15:12.530
And uh the other thing is there are as teachers, you probably have uh you know finite time interacting with these students during the course of a term or um if it's an online course for the duration of the course.

00:15:12.530 --> 00:15:24.129
And the other challenge is for the teacher, it's a lot of cognitive load because one learner might get things really quickly and they may ask you for advanced material.

00:15:24.129 --> 00:15:31.410
They they probably might ask you if uh you have what like what's next, and how do I take this to the next level?

00:15:31.410 --> 00:15:36.290
And then you have another learner who's probably um not there yet and trying to catch up.

00:15:36.290 --> 00:15:44.610
And for them, you may have to invest a ton of time and energy trying to explain how something works or how they can um understand something better.

00:15:44.610 --> 00:16:11.170
And uh if you now add AI, which is um a 24-7 private tutor that's available that can tailor the learning experiences based on the current level of a learner and based on the media that they learn best from, you suddenly have this tool that uh scales education for these teachers, right?

00:16:11.170 --> 00:16:22.210
Because all of a sudden they can ensure that each student gets the care that they need without themselves, with the without these teachers sort of stretching themselves too thin.

00:16:22.210 --> 00:16:33.649
You um get courses that you could have custom quizzes designed by these AI models based on the material that are tailored to the learner's weak spots.

00:16:33.649 --> 00:16:40.129
Because you can it's not something where you have a test and then you know have a test a few weeks later.

00:16:40.129 --> 00:16:43.170
If if you want, you could have quizzes every day.

00:16:43.170 --> 00:16:52.610
Um not saying learners should have quizzes every day, but my point is that if they have uh certain areas where they're having a bit of trouble with, you can spend more time on that.

00:16:52.610 --> 00:16:56.050
And this can be done outside of the classroom environment.

00:16:56.050 --> 00:17:15.410
So all of a sudden, learners now have access to a infinitely patient 24-7, 365-day private tutor who is able to probe and ask and teach at a level that is um sort of helping them learn where they are.

00:17:15.410 --> 00:17:19.970
And for the teacher, it allows them to then design experiences around that.

00:17:19.970 --> 00:17:24.529
So they have now a TA that's available to help.

00:17:24.529 --> 00:17:36.369
And and for designers, uh all of a sudden, you can get instant feedback on like if you have slides for a particular course, you can get feedback on where the students in general didn't get the material.

00:17:36.369 --> 00:17:39.329
So maybe some slide needs to be redesigned, for example.

00:17:39.329 --> 00:17:45.250
Or you can think about um like is there a particular topic that was not covered in depth?

00:17:45.250 --> 00:17:46.929
And was there a demand?

00:17:46.929 --> 00:17:51.490
And you can have a turnaround of a week as opposed to a semester.

00:17:51.490 --> 00:17:55.569
And I feel like that's opening up entire new opportunities in education.

00:17:55.569 --> 00:17:56.769
But I'll stop teaching.

00:17:57.409 --> 00:17:57.809
Yeah.

00:17:57.809 --> 00:17:59.809
Well, that's wonderful, Saram.

00:17:59.809 --> 00:18:06.209
I I love that because I'm a I I work in curriculum design and development, uh, my daytime job.

00:18:06.209 --> 00:18:10.529
And then in the evening, I teach online classes in instructional design.

00:18:10.529 --> 00:18:18.369
So hearing that perspective is great and it reinforces what I'm already doing as an educator and as an instructional designer.

00:18:18.369 --> 00:18:19.409
So it's fabulous.

00:18:19.409 --> 00:18:29.089
And I I really hope that um those that I that work with me listen to this and also my students, because a lot of them are educators, a lot of them are teachers.

00:18:29.089 --> 00:18:37.169
So being able to tap into this technology, I think, like you said, it personalizes the learning and it allows us to make changes on the fly.

00:18:37.169 --> 00:18:41.329
So, like you said, we don't have to wait a whole semester or anything like that.

00:18:41.329 --> 00:18:52.209
Um, because sometimes when we wait too long for this, the curriculum, it's that you have you're trying to play catch up and then you feel like your students are already behind the curveball.

00:18:52.209 --> 00:19:02.289
So it's great when you can have that opportunity to make sure that that curriculum is up to date and it's uh it's applicable to what they're going to learn on the job.

00:19:02.289 --> 00:19:03.490
So that's great.

00:19:03.490 --> 00:19:10.929
Especially in fields like I would think healthcare, like nursing and um, you know, like what you do, engineering.

00:19:10.929 --> 00:19:25.889
We have Grand Canyon University has engineering programs and the science science and engineering are really big areas at the university, but also education is is really big right now and and so uh counseling in those areas.

00:19:25.889 --> 00:19:38.929
And so trying to help them to understand that AI can come alongside them and be their, like you said, their tutor or their assistant and not look at it as uh as something that uh they should be scared of.

00:19:38.929 --> 00:19:39.409
So yeah.

00:19:39.409 --> 00:19:40.449
Right, exactly.

00:19:40.449 --> 00:19:40.689
Right.

00:19:40.769 --> 00:19:56.689
So if you if you think about what teachers can do best, um they they can look at students and see how they're responding to uh material or how they're responding to certain topics in conversation, and you can see it in their body language, their eyes, their face.

00:19:56.689 --> 00:20:06.449
Um, and you can see when somebody is motivated or is getting what is being said versus somebody who is a little uh, you know, maybe they they don't get the material at first go.

00:20:06.449 --> 00:20:10.849
And that is a qualitative signal that only teachers can get.

00:20:10.849 --> 00:20:23.889
And on the other hand, the quantitative side comes from these AI tutors because they know how many times a student had to retake a quiz, exactly um which questions they tripped up in and where they need extra support.

00:20:23.889 --> 00:20:31.250
And you all of a sudden you have both sides of the coin, which allows you to design and develop a holistic learning experience.

00:20:31.250 --> 00:20:41.809
So for me, I'm just so excited about what is available for education in this day and age versus what was available um back in the day.

00:20:42.129 --> 00:20:43.250
Right, exactly.

00:20:43.250 --> 00:20:47.009
You know, one thing you mentioned, Saran, was uh cognitive load.

00:20:47.009 --> 00:20:54.609
And that's something in our in my field in education, but especially in instructional design, that we're always trying to keep at the forefront.

00:20:54.609 --> 00:21:03.089
And I always remind my students about that because it's it's based on cognitive psychology that our brain, our working memory can only handle so much information.

00:21:03.089 --> 00:21:16.449
And so if we put too much on the students, if we info dump on them, uh, you know, and the technology is too much for them, you know, if if that's a barrier to them, then they're gonna they're gonna struggle, right?

00:21:16.529 --> 00:21:18.689
And they're not gonna they're not gonna get that.

00:21:18.849 --> 00:21:21.250
So we don't want to put any barriers in front of them.

00:21:21.250 --> 00:21:31.329
We want to make sure that it's accessible for them and that they feel like they're uh they're not having to work around things or anything like that to try to get to what they need to.

00:21:31.329 --> 00:21:32.209
So yeah.

00:21:32.209 --> 00:21:35.329
So it's nice when these tools um can help them with that.

00:21:35.329 --> 00:21:47.409
And like you said, then the the teachers and the educators can focus on uh what they need to do to to help that learning move forward and to help them have those moments where they go, oh, I understand that now.

00:21:47.409 --> 00:21:55.970
And now I can take this into the workplace and I can apply it because the program I teach for, they they do project, it's project-based learning, which is really great.

00:21:55.970 --> 00:22:03.649
I love that because they're able to they don't there's only maybe one or two papers in the whole entire program of the courses I teach.

00:22:03.649 --> 00:22:09.169
All of the rest of the assignments are all project-based where they have to literally do this.

00:22:09.169 --> 00:22:15.169
And I have a course I'm teaching right now, you'd love this, where they they have to use an AI tool twice.

00:22:15.169 --> 00:22:23.409
So they have one one assignment that's about situational leadership and they have to and they have these prompts for these, uh, for the four readiness styles.

00:22:23.409 --> 00:22:30.289
And they have to take the prompt, put it into an AI tool and converse with that tool and get that individual.

00:22:30.289 --> 00:22:36.209
So the AI tool like Chat GPT plays the employee and then they play the instructional designer.

00:22:36.209 --> 00:22:46.929
And so they have this conversation with them, working with them as the leader and trying to get them to move from one leadership style to the to the next or one situational model to the next.

00:22:46.929 --> 00:22:49.409
So it's very interesting to do that.

00:22:49.409 --> 00:23:03.089
And then another one they did that's that I'm grading this weekend is where they had to build an e-learning module on um uh it's it was for like a fake tech company uh about their policies and procedures.

00:23:03.089 --> 00:23:08.929
And so they have to use the tool to help kind of build that content and re refine it.

00:23:08.929 --> 00:23:23.009
So um, but it's funny because some of my students um they they feel like, and I think even educators do too, but some of my students actually told me in messages and in their reflection afterwards that they felt like they were cheating.

00:23:23.009 --> 00:23:24.369
And I was like, really?

00:23:24.369 --> 00:23:25.089
That's surprising.

00:23:25.089 --> 00:23:28.689
Yeah, exactly.

00:23:28.689 --> 00:23:33.889
And it's like, hmm, you know, but yet educators, they're like, they think in the back of their mind.

00:23:33.889 --> 00:23:42.209
So many of them, I've been to webinars where they think in the back of their mind and they'll type it in, they'll be like, how do I know if my students' DQ responses are genuine or not?

00:23:42.209 --> 00:23:44.449
How do I know if they're not just utilizing AI?

00:23:44.449 --> 00:23:48.209
And I'm like, well, yeah, yeah, exactly.

00:23:48.209 --> 00:23:55.009
And I was like, but before, and I thought to myself, to and I put some responses in there, and I said, Well, what happened before AI?

00:23:55.009 --> 00:23:56.769
They found ways to cheat, you know.

00:23:56.769 --> 00:24:01.329
So if they're if students are gonna cheat, they're gonna cheat no matter what, whether AI is there or not.

00:24:01.329 --> 00:24:04.609
You know, if you remove the tool, they'll they'll find ways to do it.

00:24:04.609 --> 00:24:13.329
I mean, we I could search quiz questions for the classes that I help design, and you can find them online, or you go to Course Euro and you can find them.

00:24:13.329 --> 00:24:16.769
So absent of AI, I think they can they can still do it.

00:24:16.769 --> 00:24:18.129
Exactly.

00:24:18.129 --> 00:24:26.369
I mean, they they've had Turnitin, Turnitin's been around for a long time, and institutions have used those to check for plagiarism, right?

00:24:26.369 --> 00:24:29.889
So yeah, so before AI was around, Turnitin was around.

00:24:29.889 --> 00:24:31.329
So it's funny.

00:24:31.329 --> 00:24:38.049
Um, but yeah, you know, my institution, they don't want to police the students, they want to make sure that they're using it ethically.

00:24:38.049 --> 00:24:40.289
So that's the the most important aspect.

00:24:40.289 --> 00:24:41.970
So yeah, that's great.

00:24:41.970 --> 00:24:47.009
So that kind of goes into my next question about you know, this being such a powerful technology.

00:24:47.009 --> 00:24:49.089
There's always those ethical considerations.

00:24:49.089 --> 00:25:09.329
I think just a society as a whole and not just in education, but what are uh what do you think are some of the most important things that educators and designers, or maybe even anyone, you can so you can open this up to anything, that they should keep in mind when using AI responsibly in their classrooms or their training programs, or maybe just in general in life.

00:25:09.329 --> 00:25:12.769
Um feel free to open that up to other things too.

00:25:13.730 --> 00:25:21.809
So I I I feel like um there's a lot of um there's there's a lot of ground to cover here.

00:25:21.809 --> 00:25:27.569
And uh from an ethical perspective, the very first thing that comes to mind is like the material itself.

00:25:27.569 --> 00:25:48.369
Make sure that um if you know respect copyright and ownership of the material, the lineage of the material, so that way um don't um like use material from you know, without consent on training these models to create better educational experiences.

00:25:48.369 --> 00:26:08.369
Because at the end of the day, um this is somebody else's work that is, you know, that they've probably put months or years into to develop, and uh you need consent or appropriate compensation before you consider using it to train or you know, build an AI learning experience.

00:26:08.369 --> 00:26:11.089
So that that's one that comes off the top of my head.

00:26:11.089 --> 00:26:28.049
The second one is is more related to sort of um from like a learner's side, I feel the ethical side we you just touched on it, which is don't coast or you know, mail it in um by just using AI to respond to everything, because it will respond in a very plausible way.

00:26:28.049 --> 00:26:29.490
But but here's the kicker.

00:26:29.490 --> 00:26:33.569
Um chances are it's hallucinating part of the response.

00:26:33.569 --> 00:26:45.169
Like um, there's been cases when the AI model has hallucinated textbooks that don't exist um as part of a citation or uh research papers that don't exist, for example.

00:26:45.169 --> 00:26:53.490
So if you are just you know throwing um the every question into an AI model and expecting an answer, in the long run that's actually going to hurt you.

00:26:53.490 --> 00:27:04.609
And and don't do that to uh you know teachers or educators who've spent uh hours trying to figure out how best to teach um and make sure that you know you you understand the material.

00:27:04.609 --> 00:27:06.369
That's the second one I can think of.

00:27:06.369 --> 00:27:29.889
The third one is um in terms of usage itself, because there is like AI consumes a ton of energy, and so using it responsibly is is sort of important for us to ensure that we aren't uh um you know from the from the perspective of electricity and carbon emissions and all of those things, it's it's it's crucial that we use it.

00:27:29.889 --> 00:27:33.009
Um you know, don't use it just for dad jokes.

00:27:33.009 --> 00:27:37.809
Um use it uh responsibly for something where you need that extra assistance.

00:27:37.809 --> 00:27:53.889
So uh the the the fear that I always have is uh like there might be an you know like if you have a 24-7 TA next to you to help you with everything, what happens when that TA is suddenly not available?

00:27:53.889 --> 00:27:58.689
So don't make it a crutch and you know, don't be overly dependent on it.

00:27:58.689 --> 00:28:06.289
Use your uh thinking first and then see how it lines up with what um the model has to say.

00:28:06.289 --> 00:28:19.089
And the other ethical thing um I would just generally say is um this this sort of applies to um things beyond education like interviewing and all of those things.

00:28:19.089 --> 00:28:33.250
I see that there's a sudden rise in people um either impersonating um others like by using AI to change their appearance or by um using AI to answer the questions that are asked in an interview.

00:28:33.250 --> 00:28:41.809
And those are just areas that you know it it leaves a very bad taste in the mouth for for the companies doing the interviews.

00:28:41.809 --> 00:28:45.809
So that's just uh a few that that came to mind.

00:28:46.209 --> 00:28:46.609
Wow.

00:28:46.609 --> 00:28:48.369
Oh my goodness, that is scary.

00:28:48.369 --> 00:28:52.129
I didn't I didn't think that was that that's something I haven't heard of before.

00:28:52.129 --> 00:28:55.009
So you brought wow, you brought that to my attention.

00:28:55.009 --> 00:28:58.449
So they're impersonating other people in the interview and answering them.

00:28:58.449 --> 00:28:59.009
Wow.

00:28:59.009 --> 00:28:59.970
Oh my goodness.

00:29:00.049 --> 00:29:02.769
So they're basically the creating fake personas.

00:29:02.769 --> 00:29:31.569
Um so the resume says they have all of these skills and everything, and then when you look at their uh when they come on a video interview, they use AI to change their appearance so it looks like some person and uh that way, you know, they're yeah, it's it's it's just it it's it's uh a very murky area that that has been um coming into the news of late, and I just find it, you know, extremely distasteful.

00:29:31.970 --> 00:29:41.970
Wow, that is, and it's uh yeah, it's so unethical because um they're not being genuine about who they are, and so that's really sad and what they know.

00:29:41.970 --> 00:29:42.609
Wow.

00:29:42.609 --> 00:29:43.809
That's really sad.

00:29:43.809 --> 00:29:48.609
So just like with any technology, we have to be careful and have those guardrails in place.

00:29:48.609 --> 00:29:49.329
Wow.

00:29:49.329 --> 00:29:58.369
And it's I'm glad that you brought up the part about um being mindful of how much we use it because of the the energy it consumes, right?

00:29:58.369 --> 00:29:59.889
The carbon emissions and everything.

00:29:59.889 --> 00:30:07.009
Because here in the United States, you know, we have data centers that they're building and they're they're huge data centers.

00:30:07.009 --> 00:30:16.289
I think there's one in Texas they're gonna build, and there's another one, I forget where the other one, there's uh some other one that they're gonna build, I think somewhere south, but I can't remember where.

00:30:16.289 --> 00:30:25.009
And uh so the administration's trying to do that, get ahead of it, because they as they say, we know China is trying to win the AI race.

00:30:25.009 --> 00:30:27.889
So it's like, oh, how do we get ahead of that?

00:30:27.889 --> 00:30:31.649
But yeah, it's like, well, we still have to be ethical about how we go about it.

00:30:31.649 --> 00:30:40.529
So we don't want to get ahead of it, but then and try to beat China, but then you know, hop, skip, and jump over things and do all of that.

00:30:40.529 --> 00:30:42.609
So there's always that ethical component.

00:30:42.609 --> 00:30:44.689
Anytime there's technology in place.

00:30:45.250 --> 00:30:46.769
There's always two sides to the coin, right?

00:30:46.769 --> 00:30:52.289
So um just being cognizant of that and being responsible, I think that's that's crucial.

00:30:52.929 --> 00:30:53.649
Absolutely.

00:30:53.649 --> 00:30:54.209
Yeah.

00:30:54.209 --> 00:30:55.009
Love that.

00:30:55.009 --> 00:30:56.049
Thank you so much for that.

00:30:56.049 --> 00:30:57.250
That's that's amazing.

00:30:57.250 --> 00:31:01.649
Those ethical considerations are always something that we have to keep at the forefront of our minds.

00:31:01.649 --> 00:31:02.689
Absolutely.

00:31:02.689 --> 00:31:09.970
So uh before we move into the final question, I'd love to slip one more in as a bonus question that I know will resonate with my listeners.

00:31:09.970 --> 00:31:19.809
So looking ahead, what are a couple of things that excite you the most about the future of AI and what should educators and instructional designers be preparing for now to stay ahead?

00:31:20.929 --> 00:31:31.009
So in terms of what excites me, it's just the number of things that we considered impossible just a couple of years ago are now suddenly within grasp.

00:31:31.009 --> 00:31:41.329
And um a lot of the um ideas that seem like fiction are some suddenly like possible if not plausible, right?

00:31:41.329 --> 00:31:55.649
It's uh it's it's important for us to the thing I'm most excited about is just being um empowered to do and try different things and actually see the outcome of those things without a long turnaround time.

00:31:55.649 --> 00:32:14.449
But because previously um for anything in tech if you needed to build something and see how it works, it would take like weeks or months and you'd need a large team and it would you know it's like you you couldn't try out a whole bunch of ideas all at once and see how they turned out because it would just be infeasible.

00:32:14.449 --> 00:32:19.569
Now all of a sudden with AI assistance you can quickly prototype ideas.

00:32:19.569 --> 00:32:38.849
And therefore you can um either you know veto a whole bunch that that aren't um you know go good to go forward yet or and you can find the the ones that have a lot of potential very very quickly and with a fraction of the time and cost that it it would take uh otherwise.

00:32:38.849 --> 00:32:42.369
The other thing I think I already spoke about is like education.

00:32:42.369 --> 00:32:46.449
I'm super excited about how AI is going to change the field.

00:32:46.449 --> 00:32:50.609
And um for me just being able to even even for me, right?

00:32:50.609 --> 00:33:05.809
I mean like there's this tool called Notebook LM that I uh use all the time and it's completely changed the way that I learn new topics or concepts because AI is evolving so fast and I try to keep up with a whole bunch of research papers that come out.

00:33:05.809 --> 00:33:21.730
And all of a sudden notebook LM allows me to do that more efficiently and I can create mind maps, I can create uh quizzes, I can even have like if I'm short on time, it creates a podcast for me to listen to based on the material which is absolutely mind blowing.

00:33:21.730 --> 00:33:36.609
And um just for me to have all of these tools at uh you know at my fingertips is is is a blessing and just uh I'm excited to see where this goes from here in terms of um education and for me to help uh teach others.

00:33:36.609 --> 00:33:38.849
So those are the two things I'm excited about.

00:33:38.849 --> 00:33:47.490
Now coming to what learners need to be um and instructional designers need to be preparing for I would just say the easiest way to start is spend time with these tools.

00:33:47.490 --> 00:34:04.529
Just pick a tool that you want to use it can be anything your chat GPT, Claude, something else if you're in design maybe a tool like Canva or gamma and just just play around with the AI component of those tools and just see how to use it, interact with it.

00:34:04.529 --> 00:34:07.970
Like I said there is no user manual we're kind of writing it as we go.

00:34:07.970 --> 00:34:16.369
So just as as you spend more and more time with it you figure out the patterns that um allow you to get to the desired results faster.

00:34:16.369 --> 00:34:29.329
And they also allow you uh you know a safe space to experiment, try new ideas out and then see in and maybe unlock new use cases that were previously uh something you hadn't considered or thought about.

00:34:29.329 --> 00:34:42.449
So from that perspective it's um just display of the tools and then uh the other aspect is maybe have a basic sense of AI literacy um in just understand how these models work behind the scenes.

00:34:42.449 --> 00:34:48.130
I'm not saying you need to uh know how to code these up or do any of those fancy things.

00:34:48.130 --> 00:35:00.530
If you can fantastic but um just understand how they work and how they're built because that gives you uh context into their strengths and limitations and then that in turn allows you to work more effectively with them.

00:35:00.530 --> 00:35:08.930
You know where they're not good, where you need to step in and conversely where they are good so that you can step back and then do something else that you are good at.

00:35:08.930 --> 00:35:11.730
So um that's what I suggested.

00:35:12.370 --> 00:35:13.410
Those are great tips.

00:35:13.410 --> 00:35:14.930
I love that that's great.

00:35:14.930 --> 00:35:44.530
And I I'd love that you mentioned for education Canva and Gamma because I use those tools quite a bit myself and I even did it for I used it for work recently with Gamma and I took some of my old PowerPoint slides because I've been teaching OneNote for for quite a while to help my coworkers to get more organized and kind of take maybe some of their ideas and just kind of organize them a little bit better because we have OneNote as part of Office 365.

00:35:44.530 --> 00:35:56.050
So I had them you know PowerPoint's great and PowerPoint serves great purposes but I thought can I kind of refresh these and give them a little bit of a boost and and everything like that.

00:35:56.050 --> 00:36:01.650
So I went to Gamma because one of my students had her presentation in Gamma and I was like oh what's this?

00:36:01.650 --> 00:36:05.329
Okay, I'm learning all kinds of tools here as an instructor it's amazing.

00:36:05.329 --> 00:36:14.690
And then my students are like hey yeah go look go use 11 labs and go and I'm like oh my goodness it's amazing how much I've learned in five years and then just how the tools just keep getting better.

00:36:14.690 --> 00:36:21.170
But yeah I was able to take that those presentations and really uh take them to the next level.

00:36:21.170 --> 00:36:25.170
So it was fun and it didn't take a lot of time because I already had the content.

00:36:25.170 --> 00:36:31.730
So I was just putting it in a different format and and kind of just reimagining it and what it could look like.

00:36:31.730 --> 00:36:33.730
So yeah it was it's fantastic.

00:36:33.730 --> 00:36:34.930
Yeah I love that.

00:36:34.930 --> 00:36:37.650
And Canva, oh my goodness, I love Canva.

00:36:37.650 --> 00:36:45.170
Wow they the Canva code, I've used that a few times and it's really neat to actually see that's it's coding and it's creating it for you.

00:36:45.170 --> 00:36:50.610
And then like you said, just like with AI, you have to kind of tweak it a little bit work with it.

00:36:50.610 --> 00:36:58.690
I've I've had to do that sometimes where I asked it to do a matching game and it didn't turn out that great the first time.

00:36:58.690 --> 00:37:10.690
It was okay but not all not everything worked and I was like okay well can you do that and then I asked it to actually do like a a little uh celebration thing at the end where it gave it and actually works now really well.

00:37:10.690 --> 00:37:17.010
But at the end it'll when you match everything it'll this little trophy will come up and it'll do the little confetti at the end.

00:37:17.010 --> 00:37:17.730
So it's really neat.

00:37:17.730 --> 00:37:21.570
Nice so I asked it can you please do this and then it I saw it as it was coding.

00:37:21.570 --> 00:37:29.809
It was really fun to actually see it in action and then just work with it a little bit and say okay you know not quite right can you can you do this?

00:37:29.809 --> 00:37:31.970
And so yeah it was a lot of fun.

00:37:31.970 --> 00:37:39.730
Love that yeah there's another tool out there have you heard of genially before it's spelled G-E-N-I-A-L-O-I.

00:37:39.730 --> 00:37:46.769
It's an interactive tool where you can do like interactive infographs and quizzes and escape rooms and stuff like that.

00:37:46.769 --> 00:38:23.809
I need to take this out yeah you'll have to check out Genially one of my students introduced it to me a few years ago and now they've got AI technology in there where you can actually take your interactive piece and you can if you have you have to I have the EDU pro version which is $60 a year in uh here in the United States I don't know how much it would be where you are but it's really nice because you can get premium templates and you get some of the AI technology but if you go up I think I can't remember if it's elite I can't remember which one it is but you can actually have it translate your genially into different languages.

00:38:23.809 --> 00:38:33.570
So for example if I was teaching a Spanish class I could you know have it in Spanish too or if I have learners from all over I can I can translate it to other languages.

00:38:33.570 --> 00:38:35.329
So I was like wow that's pretty neat.

00:38:35.329 --> 00:38:37.970
And so it uses that AI technology to translate it.

00:38:38.050 --> 00:39:19.090
But if I did that I would want someone that knows that language really well to double check my the work and not just assume that it translated it correctly but I love that you said that I was just going to say that's that's uh and and that applies to any educational experience right I mean all of a sudden you you you as a teacher are no longer limited by language you can teach uh students worldwide who come from different cultures but again trust but verify um so have somebody absolutely I love that great so as we wrap up based on your journey as an AI leader and author what's one piece of encouragement or advice you would give to listeners who are looking to enhance their skills in AI technology?

00:39:19.090 --> 00:39:46.289
I would say that pick a topic that you are really excited about or pick a problem that is bothering you a lot like scratch your own edge and if it's a a problem that is uh you know something that you do on day try to solve on a daily basis just write it out and I mean write out the steps basically and then try to explore how AI can replace each step one by one.

00:39:46.289 --> 00:39:54.130
It may be able to replace a step it may not but just see how you can bring in AI into that workflow.

00:39:54.130 --> 00:40:07.250
And chances are a lot of that might be automated away and as you do this you then start learning to use AI for different types of subproblems and then over time those become different types of problems.

00:40:07.250 --> 00:40:15.250
So you're building the muscles that are necessary to use AI and think in a way where you try to solve a problem first by yourself.

00:40:15.250 --> 00:40:32.769
You use your own critical thinking and and wherever you feel that there are inefficiencies or where you need that extra support you figure out okay here's where AI can step in and this allows me time to do something else or this allows me to completely automate away this this problem so I don't need to think about it.

00:40:32.769 --> 00:40:34.930
That's how I would approach it.

00:40:34.930 --> 00:40:48.930
And then the second one is like just um trying to keep up with what's going on I know that there's a lot of hype around the area there's like every day there are announcements on this tools come out and it's going to completely change the world and all of that.

00:40:48.930 --> 00:40:51.570
But uh try to keep a pulse on what's going on.

00:40:51.570 --> 00:40:59.250
Generally if a particular topic or tool is in the news for more than four to six weeks chances are there's something there.

00:40:59.250 --> 00:41:08.850
If it's there for only one or two days then I would exercise caution um so those are the two suggestions in addition to what I said about AI literacy.

00:41:09.250 --> 00:41:09.570
Right.

00:41:09.570 --> 00:41:28.289
Great I love that keeping keeping an eye on those things and not taking it for face value right and really keeping that critical thinking lens and I that's what some of the instructors I've talked to over time the last few months that they said they they're fearful that their students are going to lose that critical thinking.

00:41:28.289 --> 00:41:38.610
And I said well then our job as educators is to ensure that they don't and create authentic assignments assignments where they don't cheat right automatically.

00:41:38.610 --> 00:41:41.970
They don't want to do it because they want to actually do that assignment.

00:41:41.970 --> 00:41:50.450
And uh so we're trying to as an institution you know that we're trying to get away from so many papers and not have a lot of papers.

00:41:50.450 --> 00:42:00.530
I mean they still need to know how to write but we always try to think of what do they need to know in the field to be able to get into whatever particular industry they're trying to go into.

00:42:00.530 --> 00:42:07.250
So I think that's so important to have that um in place and so that they know that they can do that.

00:42:07.250 --> 00:42:08.930
So yeah that's that's great.

00:42:08.930 --> 00:42:09.490
I love that.

00:42:09.809 --> 00:42:25.329
Lots of great we've we've had open book exams um in the past and one of the things that affords the the the teacher is that you can set really um interesting questions that you don't have a direct answer to.

00:42:25.329 --> 00:42:34.289
And you'd have to do a bit of work you'd have to look at different sources and then come up with an answer that is sort of uh helping you assimilate the material.

00:42:34.289 --> 00:42:46.050
Now with AI that open book concept becomes you know it it goes to a completely different level because you instantly get the answer but again you need to fact check there's all those kinds of things.

00:42:46.050 --> 00:43:12.370
But I wonder if there is a way to create an experience that simulates open book but you know using AI in a way where you are practically testing things as opposed to like you said just having people done in reports and papers because those are the kinds of assignments where I feel students might learn the best because they are forced to use AI and at the same time they're forced to use their own thinking.

00:43:12.769 --> 00:43:13.090
Right.

00:43:13.090 --> 00:43:23.250
I love that yeah kind of like the assignments that I I'm braiding and it it forces them like you said it's got doesn't just have one part to it has multiple layers to it.

00:43:23.250 --> 00:43:35.329
So you you can see did the students really understand it and I can tell you know they're not just putting that input in there and just getting the output and doing a screenshot, but they're actually learning from it.

00:43:35.329 --> 00:43:43.970
And and then the reflection that they did afterwards when they did that first assignment was, you know, uh what went well what would you do differently next time?

00:43:43.970 --> 00:43:53.410
How would you converse with the uh with those different individuals you know in the AI tool next time to make it better and what did you learn through that experience?

00:43:53.410 --> 00:43:58.130
So that's something AI can't repu replicate is uh is those experiences.

00:43:58.130 --> 00:43:59.730
So it's really great to see that.

00:43:59.730 --> 00:44:01.010
Yeah I love that.

00:44:01.010 --> 00:44:05.090
Yeah just keep growing and keep learning right that's the that's the key.

00:44:05.329 --> 00:44:06.530
Great that's the key.

00:44:06.769 --> 00:44:11.809
Yes thank you so much Siram for sharing your insights and experiences with us today here on the podcast.

00:44:11.809 --> 00:44:27.329
The wisdom that you've offered for making AI more accessible to your guidance and applying it responsibly will no doubt inspire and encourage my listeners in their own paths as they continue to explore this tool these tools and learn to integrate them into their lives.

00:44:27.329 --> 00:44:28.130
I love it.

00:44:28.690 --> 00:44:39.890
Thank you so much it was a pleasure to learn uh from you as well I think I learned a ton about how um the thought that that goes behind building courses and learning experiences.

00:44:39.890 --> 00:44:42.050
So this was super interesting.

00:44:42.450 --> 00:44:50.450
Great yes I look forward to having you back on the show because once I have someone on my show once then we always think about there's something else we could talk about.

00:44:50.450 --> 00:45:01.490
Like you said earlier we could dig in and we could really uh you know hone in on some other areas too so if there's anything you think of along the way and you want to come back on the show feel free to do that.

00:45:01.490 --> 00:45:08.289
I'd love to have you back on and we can dig a little deeper and um and maybe even see where things take us.

00:45:08.289 --> 00:45:09.250
So I'd love that.

00:45:09.250 --> 00:45:15.090
Definitely great well thanks again Sarah I appreciate it and look forward to having you back.

00:45:15.730 --> 00:45:17.570
Thank you thank you Becky this is fun.

00:45:18.210 --> 00:45:25.170
Yes absolutely thank you for taking some time to listen to this podcast episode today.

00:45:25.170 --> 00:45:27.410
Your support means the world to me.

00:45:27.410 --> 00:45:36.210
If you'd like to help keep the podcast going you can share it with a friend or colleague leave a heartfelt review or offer a monetary contribution.

00:45:36.210 --> 00:45:41.490
Every act of support big or small makes a difference and I'm truly thankful for you.