AI & Children's Wellbeing
As AI continues to advance and become more prevalent in various aspects of our lives, it is crucial to understand its potential impact on the mental, emotional, and physical health of students. Join us as our expert panel discusses new research findings from the School Social Work + AI study and explores the role of school-based health staff in guiding the responsible adoption of AI in schools.
Attendees gained valuable insights into:
- The latest research on AI and children's wellbeing 
- Best practices for the responsible adoption of AI in schools 
- Strategies for health staff to contribute to AI policy development 
- Opportunities for collaboration to create AI-enhanced learning environments that prioritize student wellbeing 
- 
      
        
          
        
      
      National Survey of School Social Workers on Generative Artificial Intelligence Ofcom Online Nation 2023 Report Pew Research Center survey of U.S. teens AI Interventions to Address Cyberbullying Alongside App Clinican-Powered AI for Youth mental Health Chatbot AI-Assisted Fakes Mean for CyberBullying in Schools & Policies 
- 
      
        
      
      Amanda Bickerstaff Amanda is the Founder and CEO of AI for Education. A former high school science teacher and EdTech executive with over 20 years of experience in the education sector, she has a deep understanding of the challenges and opportunities that AI can offer. She is a frequent consultant, speaker, and writer on the topic of AI in education, leading workshops and professional learning across both K12 and Higher Ed. Amanda is committed to helping schools and teachers maximize their potential through the ethical and equitable adoption of AI. Rebecca K. Oliver Ms. Oliver is the Executive Director of the School Social Work Association of America. Prior to becoming the Executive Director, Mrs. Oliver served on the SSWAA Board of Directors and has over 20 years experience working as a school social worker. In her current role with SSWAA, Rebecca is able to support school social workers across the nation and advocate for the profession about which she is so passionate. When not working, Rebecca enjoys traveling with her husband Jon, singing, running, reading, doing home-improvements, and outdoor activities including walks with her two dogs, Abby & Buddy. Andrew Buher Mr. Buher is the Founder and Managing Director of Opportunity Labs, a national non-profit research, policy, and consulting lab. Andrew is a visiting lecturer at Princeton University's School of Public and International Affairs. Andrew was appointed by President Obama to serve as a White House Fellow at the U.S. Department of Housing and Urban Development. Prior to his work at HUD, Andrew was the Chief Operating Officer of the New York City Department of Education and formerly Chief of Staff to Chancellor Dennis Walcott. Andrew is a White House Fellow, Forbes 30 under 30 honoree, French American Foundation Young Leader, Atlantic Council Millennium Fellow, and an Education Pioneer. He received an MPA from Columbia University and a BA from Rider University. His writing has been featured in The Hill, CNN, Politico, and the NY Daily News. Dr. Marina Badillo-Diaz Dr. Badillo-Diaz is an experienced school administrator and counseling director with a demonstrated history of working in community mental health and in education as a social worker. Currently, Dr. Badillo-Diaz is a consultant focusing on the training of educators and social workers. She is also an adjunct professor and board member of the National School Social Work Association of America. She is also the author of the blog, “The AI Social Worker”, a 21st Century Skills and AI practice guide for Social Workers. Her areas of interest include 21st-century skills, social-emotional learning programming, school social work practice, education, youth mental health, clinical supervision for social workers, ethical AI, data management, and career development. 
- 
      
        
      
      Amanda Bickerstaff: Hello! Hi! Everyone welcome our first webinar of what feels like summer here in New York City really excited to have you all here with us today takes us a little bit for us all to get in with zooms, letting everyone in one at a time. But we're really excited to have you here. We're gonna get started just in about a minute. And so if you've been with us before. We know we love that you say hello. So please. 00:00:34.700 --> 00:00:59.519 Amanda Bickerstaff: put in the chat like who you are, what you do and where you're coming from. I know we have people from all over the world. And so it's really great. We're so excited to have this conversation today. And you know, I think that this is a a really important conversation that we don't believe as a collective. So our wonderful panelists today is being talked about enough 00:00:59.520 --> 00:01:25.389 Amanda Bickerstaff: and so really excited to have you all here as we really gonna dive into, like what AI means for children, for child wellbeing student wellbeing. Because this is something that we know. There has been an enormous impact of technology on student well, being already with the advent of devices, Internet. And of course, social media. And so now that we have this new arrival technology where this technology is something that we are, it's it's not just 00:01:25.390 --> 00:01:49.090 Amanda Bickerstaff: impacting us as educators and leaders and and industry people, but also really students. And so there's a recent piece of research that we're talking about today. And Andrew's gonna come on screen in a moment that talks about this. But also we know from research that students have been interacting with things like Snapchat AI for a full year all the way down to 7, 8 years old. And so we're really gonna start focusing on what I think will be the first of many conversations 00:01:49.090 --> 00:02:12.919 Amanda Bickerstaff: on the impact of AI on student. Well, being but I'm actually gonna just remind everyone, if you've been here before or not, this is your first webinar with us it is meant to be a community of practice. So I really appreciate everyone that's already said hello in the chat. What we're gonna do is you use the chat to like kind of share resources, chat with everybody. But there's a specific question for the panel. If you can put that into the QA. 00:02:12.920 --> 00:02:36.701 Amanda Bickerstaff: That would be great. So that makes it a little bit easier because we do have a really active chat. So we wanna make sure that we don't miss any of those great questions. But if you do have a resource, let's say that you have a piece that you really love or strategy? Around signal being an AI, please put that into the chat, and we'll be dropping some of those resources in as we go as well. So what I'm gonna do is I'm gonna come off share. And I'm actually gonna invite Andrew Burr. And I have to say, Andrew, 00:02:37.110 --> 00:03:03.769 Amanda Bickerstaff: it's a kindred spirit in a lot of ways. But when I first talked to Andrew about generative AI, the first thing that he had this we had to talk about was this idea of like, what will the impact be on wellbeing and students themselves? And so there's a lot of great research has come out of opportunity labs if you want to come on screen, Andrew, that would be great. And what we're gonna do is I'm gonna hand it over to Andrew to introduce himself, but also to run us through a quick kind of 7 min overview of the research. 00:03:03.770 --> 00:03:19.839 Amanda Bickerstaff: Evidence is the most important thing that we have today, especially this new crazy time. So to be able to situate that together is really important. And then what we're gonna do is we have a couple of really awesome panelists that are gonna come on and we're gonna have a bit of a discussion. So I'm gonna turn over to you, Andrew, and take it away. 00:03:20.820 --> 00:03:34.381 Andrew Buher: Great thanks. So much, Amanda. For the warm welcome to the AI for education team, for coordinating and thanks to everyone for joining. I guess I should say good evening. Good afternoon. Good morning, since it seems like there's folks from 00:03:34.730 --> 00:03:58.959 Andrew Buher: all over the world. On this call. Thanks for being here taking time out of your busy schedules. And more importantly, thanks for the work that you do. On behalf of kids every day. So, as Amanda mentioned, I am Andrew Beer. I am from opportunity labs. We are a national nonprofit organization. We've been around for about 5 years. And I also teach at Princeton University. 00:03:59.343 --> 00:04:10.459 Andrew Buher: Here in New Jersey. So what I am hoping to do in the next 5 min is to build a foundation for what we hope is a very 00:04:10.590 --> 00:04:13.460 Andrew Buher: practical and pragmatic conversation. 00:04:13.600 --> 00:04:20.969 Andrew Buher: I want to spend those 5 min basically talking about 3 things. Why we undertook 00:04:20.990 --> 00:04:23.299 Andrew Buher: this research that we're going to discuss 00:04:23.360 --> 00:04:29.899 Andrew Buher: what we learned and then some policy considerations. So I'm gonna go ahead and share my screen 00:04:31.240 --> 00:04:33.750 Andrew Buher: hopefully, everybody can see this 00:04:35.040 --> 00:04:36.520 Andrew Buher: alright. So 00:04:37.015 --> 00:04:50.130 Andrew Buher: we put a national survey in the field in partnership with Rebecca Oliver and her team, who you will meet momentarily and had a really great response. Almost 600 respondents. 00:04:50.762 --> 00:04:58.809 Andrew Buher: and so I wanna start fundamentally by talking about why we undertook this research. 00:04:59.630 --> 00:05:04.669 Andrew Buher: There's there's basically 2 reasons. First, I think we can 00:05:04.730 --> 00:05:08.850 Andrew Buher: disagree about the scale of 00:05:08.910 --> 00:05:14.770 Andrew Buher: harm to kids caused by increased screen time and social media use 00:05:14.780 --> 00:05:29.390 Andrew Buher: over the last 14 ish years. But to argue that those concerns are baseless or too minimal to matter. I think just isn't supported by the evidence anymore at this point. So 00:05:29.500 --> 00:05:32.450 Andrew Buher: and I'm sure so many of you on this call 00:05:33.012 --> 00:05:44.329 Andrew Buher: have experienced the suffering that children and families and schools and communities. Have had to navigate since the early 20 tens, because of 00:05:44.430 --> 00:05:46.530 Andrew Buher: kids, online lives. 00:05:46.830 --> 00:05:52.650 Andrew Buher: And of course, within schools, school-based health staff, including school social workers. 00:05:53.034 --> 00:06:03.400 Andrew Buher: Have acutely felt this impact. You know, when we talk to district superintendents and school leaders and school based health staff, we hear again and again. The stories about 00:06:04.000 --> 00:06:10.569 Andrew Buher: these really troubling and instances of of impact that they have had to to navigate 00:06:11.315 --> 00:06:23.659 Andrew Buher: and so, if generative AI has the potential to exacerbate these challenges even a little bit. We thought it was really important to understand the perspectives of 00:06:24.147 --> 00:06:35.260 Andrew Buher: folks that are closest to kits. And one of those roles, obviously within schools is school social workers. And so we really tried to 00:06:35.681 --> 00:06:46.910 Andrew Buher: get a sense of their perspective across of a range of issues related to generative AI. That's the first reason we undertook this research. The second reason 00:06:47.070 --> 00:07:09.009 Andrew Buher: is that school social workers are super unique. Right? They have a unique set of skills. Their role is really unique. They are incredibly proximate to students, staff and families. And we think that they should be at the table. When policy and programmatic decisions are being made. 00:07:09.510 --> 00:07:13.720 Andrew Buher: That have the potential to influence their work. And obviously. 00:07:14.165 --> 00:07:19.269 Andrew Buher: the adoption of new technologies has the potential to do exactly that. 00:07:19.800 --> 00:07:22.769 Andrew Buher: Okay, so what did we learn? 00:07:23.159 --> 00:07:40.310 Andrew Buher: Everybody? Hopefully, has seen this research. If you haven't, you can check it out, I think. The AI for Ed team shared it with with everyone. But I wanna highlight, maybe 3 non obvious takeaways that we have been thinking about as a result of of this research. 00:07:40.530 --> 00:07:41.860 Andrew Buher: So first. 00:07:42.202 --> 00:07:45.529 Andrew Buher: we call this the the don't knows, and the neutrals 00:07:46.109 --> 00:07:52.220 Andrew Buher: there were lots of don't know. And neutral answers across questions. 00:07:53.450 --> 00:07:55.749 Andrew Buher: we think that's actually pretty meaningful 00:07:56.216 --> 00:08:12.030 Andrew Buher: because the don't knows. And neutrals represent knowledge gaps in what school social workers know about generative AI to no fault of their own. Of course. We actually think this is directly attributable to the fact that 00:08:13.610 --> 00:08:21.240 Andrew Buher: access to AI literacy programming for school-based health staff, including school social workers, is 00:08:21.360 --> 00:08:25.930 Andrew Buher: uneven at best and non existent in a lot of cases. 00:08:26.650 --> 00:08:33.149 Andrew Buher: So that's the the first non obvious takeaway. Second, we call this the what the how 00:08:34.819 --> 00:08:35.980 Andrew Buher: respondents 00:08:36.059 --> 00:09:03.609 Andrew Buher: overwhelmingly agreed about potential risks associated with the adoption of generative AI. And I wanna emphasize the word potential right? There are risks related to safety and data, privacy and cognitive development, and and certainly mental and and social wellbeing. And I think we can call these the what right? 00:09:04.440 --> 00:09:07.999 Andrew Buher: But respondents were also deeply unsure about how. 00:09:08.651 --> 00:09:11.139 Andrew Buher: These risks would manifest 00:09:11.737 --> 00:09:25.309 Andrew Buher: and you know, I think if we're being honest intellectually with with ourselves, no one really knows the answer to how just yet. But there are starting to be some really smart projections about 00:09:25.410 --> 00:09:30.530 Andrew Buher: how these risks might and might is bold and capitalized here. 00:09:30.944 --> 00:09:34.100 Andrew Buher: Manifest, which we think is is really meaningful. 00:09:35.700 --> 00:09:49.499 Andrew Buher: third, important, non obvious takeaway. We're thinking about as out of the loop. And maybe this isn't as non obvious as the the first 2 00:09:49.640 --> 00:09:55.490 Andrew Buher: takeaways that that I discussed. But I think it might be the most eye opening. 00:09:56.040 --> 00:10:04.389 Andrew Buher: School social workers are just out of the loop too much. When it comes to the adoption of new technologies in schools. 00:10:04.550 --> 00:10:10.710 Andrew Buher: They aren't consulted as frequently and consistently as they should be. 00:10:10.760 --> 00:10:16.129 Andrew Buher: And as a result they aren't talking to important stakeholders, including families 00:10:16.210 --> 00:10:20.640 Andrew Buher: as often as they should be about these issues. 00:10:21.240 --> 00:10:45.639 Andrew Buher: So the the 3 big, non obvious takeaways, the don't know's and the neutrals, the what's but not the house. And then the fact that school social workers and and honestly, most school based health staff, including school nurses and school counselors, are often out of the loop when it comes to conversations about the adoption of emerging technologies in schools and in the classroom in particular. 00:10:45.980 --> 00:10:53.979 Andrew Buher: Okay, again, you have the research. This is the full list of policy considerations that we that we 00:10:54.490 --> 00:10:57.789 Andrew Buher: codified after after reading the research. 00:10:58.030 --> 00:11:03.099 Andrew Buher: But if I could summarize, I would effectively say 3 things 00:11:03.580 --> 00:11:19.009 Andrew Buher: leaders at the district level at the school level should be asking as many questions and seeking as much input as possible from folks that are closest to kids, particularly folks that are dealing with. 00:11:19.621 --> 00:11:23.389 Andrew Buher: The manifestation of the youth, mental health crisis. 00:11:23.630 --> 00:11:25.400 Andrew Buher: and of course, that 00:11:25.460 --> 00:11:30.899 Andrew Buher: includes school social workers. But it really includes all school based health staff, and of course, educators as well. 00:11:31.367 --> 00:11:37.639 Andrew Buher: They should be working to democratize the flow of information, and that's inclusive of 00:11:37.690 --> 00:11:50.040 Andrew Buher: access to knowledge, building opportunities like the AI literacy work AI, for education provides so that all stakeholders are informed enough to make thoughtful decisions. 00:11:50.360 --> 00:11:51.829 Andrew Buher: And then, finally, I think. 71 00:11:51.860 --> 00:11:58.500 Andrew Buher: putting in place a system of systems that help all stakeholders prepare for both 72 00:11:58.650 --> 00:12:10.609 Andrew Buher: potential opportunities which are many and potential risks associated with this this new technology. Is gonna be really, really important. So 73 00:12:11.246 --> 00:12:15.220 Andrew Buher: hopefully, this gives folks a good sense of 74 00:12:15.270 --> 00:12:28.920 Andrew Buher: why we undertook this research. What we learned and some ideas for action. But most importantly, hopefully, this is a foundation for a really interactive conversation that is both 75 00:12:28.980 --> 00:12:35.049 Andrew Buher: pragmatic and practical. And so I'm going to pause there and pass it back to Amanda. 76 00:12:35.270 --> 00:12:41.400 Amanda Bickerstaff: Thank you so much, Andrew, and I just want to kind of put a PIN on a couple of things. First of all, we see a lot of research where people 77 00:12:41.420 --> 00:13:06.359 Amanda Bickerstaff: are kind of put to what like like make a decision. But I think it actually is really important to have that neutral. We don't know yet as actually something that social workers are identifying. And I think that when you talk about a gap in terms of like training and also in terms of being a part of those bigger conversation with technology is something that's just really important. And there's Carolyn in the in the chat talks about being a librarian and not being 78 00:13:06.638 --> 00:13:31.159 Amanda Bickerstaff: brought in. So like, what we see a lot is that when schools and systems actually bring in school psychologists, counselors and media specialists and other people that are in and around students. For these generative AI trainings you see a much richer discussion that goes way, way way deeper, because we're looking at the whole child now, and the whole child experience in schools. And so I hope that this is a you know a galvanizing moment to have that focus which I think is really important. 79 00:13:31.160 --> 00:13:47.699 Amanda Bickerstaff: But there's also, before you bring up the panel. There was a couple of questions about this is a first step, right? So this is focus on teachers. But do you have a a plan or know of a a plan to bring students into the fold in terms of following up with similar questions to them about the potential harms and benefits. 80 00:13:49.560 --> 00:14:08.413 Andrew Buher: Yeah, we've seen quite a bit of research already. That has surveyed students. So we've seen some from pew we've seen some from the Walton family foundation. At least that was commissioned by the Walton family foundation. And you know, I would say most of that research is focused on 81 00:14:08.750 --> 00:14:24.359 Andrew Buher: usage practices, right? How young folks are interacting with the technology and what ways they are utilizing it. And the extent to which they're utilizing it. I haven't seen anything that has gotten really deep on potential impacts to 82 00:14:25.145 --> 00:14:47.054 Andrew Buher: well, being. And so there is certainly an opportunity, I think, to broaden the conversation and bring youth voice into a deeper dialogue about the potential impacts on on their wellbeing. Which is gonna be really really critical, I think, in the United Kingdom. You are starting to see some 83 00:14:47.998 --> 00:14:53.480 Andrew Buher: youth voice coalesce around concerns around screen time and social media 84 00:14:53.929 --> 00:15:05.220 Andrew Buher: which is really really interesting. But I will try to find and drop the links for the research that I've seen from queue and and Walton. About 85 00:15:05.650 --> 00:15:09.380 Andrew Buher: youth interacting with these technologies. As this conversation goes on. 86 00:15:09.650 --> 00:15:33.710 Amanda Bickerstaff: That's great. And if my team could also find the Ofcom online survey would be great as well. It's a Uk study, but I think, as a little bit of a a teaser, Andrew. And for people here, we're actually going to be starting a student in the AI education community. That's gonna be a partner to our women in AI education community for that exactly that reason of having more opportunity for students that are concerned about the impact or just wanted 87 00:15:33.710 --> 00:15:58.659 Amanda Bickerstaff: to make sure that student voice is part of the conversation cause I think that that's gonna be so important, moving forward, so really excited to like. Now bring up our panel. And we actually have, 2 amazing panelists, Rebecca, who is supportive of this techno of this research, but also Marina, who is still doing the work as well. And so I'm gonna call out both. And Marina can come on video as well. And what we're gonna do is we're gonna have a bit of a just a discussion. 88 00:15:58.660 --> 00:16:26.900 Amanda Bickerstaff: And I know there was a question in the QA. About potential risks and harms going deeper into that. So we're gonna take a little. We're gonna take a pragmatic approach, a balanced approach which I know that all of us here love to have. So we are gonna talk about concerns, but also potential benefits, because we always wanna have that balanced approach. But we start with every single webinar starts the same way with our panelists. Which is, do you mind introducing yourself and then talking about that first time that you interacted with generative AI. So we're gonna start with Marina. If you want to go first. That would be amazing. 89 00:16:27.660 --> 00:16:38.586 Dr. Marina A. Badillo-Diaz: So Hi, everybody! Thank you again for putting together this webinar and panel. So my name is Dr. Marina Badio Diaz. I am a a school social worker working in education for the past 10 years. 90 00:16:38.890 --> 00:17:03.759 Dr. Marina A. Badillo-Diaz: I started in K through 8 education as a school social worker, moved into a leadership role and position as a youth development director and an alternative transfer high school setting here in New York City. But most recently I've been teaching a school social work at the the masters level. So teaching future social workers and also doing consulting in K through 12 schools. And so a bit about my entry into using Jenner of any 91 00:17:03.760 --> 00:17:24.150 Dr. Marina A. Badillo-Diaz: AI really came about a year ago, as I am scrolling on my consultant Instagram account of these reels on like generate content with with AI. Put these props to the chat. Gbt, and I'm a social worker. I didn't go to school for marketing, and this is a really big challenge of mine to put together content for my social media pages. 92 00:17:24.150 --> 00:17:37.509 Dr. Marina A. Badillo-Diaz: And I just found it so useful and practical to try chat to Vt, and it really just grew from there, as I'm doing multiple projects at the same time. And I just found it so useful, not just in my business, but also as a practitioner, as well. 93 00:17:38.020 --> 00:17:58.369 Amanda Bickerstaff: That's great. And I love like you're like, what is this? What is this thing that everyone's talking about? But I love that you actually use it for productivity as well cause that as someone that's going to be as entrepreneurial as you are, and also still doing the work in in schools. I think it's a great way to to find some balance which is amazing. So, Rebecca, do you same question to you? Would love to know more about you. 94 00:17:58.790 --> 00:18:21.160 Rebecca Oliver: Sure. Hi, everybody! My name is Rebecca Oliver. I again also wanna thank Amanda and her team for pulling this together, Andrew for looping school social workers in on this really important research. I started as a school social worker worked in school settings for over 20 years, and then I've been with the School Social Work Association of America for almost 10 95 00:18:21.170 --> 00:18:25.910 Rebecca Oliver: and so we're just really excited to be able to 96 00:18:26.030 --> 00:18:32.960 Rebecca Oliver: really get in on these conversations as we just see. At least, I have seen AI just 97 00:18:33.030 --> 00:18:45.840 Rebecca Oliver: kind of explode over the last couple of years. So just seeing its impact. So I'd say, maybe my first experience with it was some chat bots that I realized really weren't humans and 98 00:18:46.090 --> 00:19:00.479 Rebecca Oliver: or just regurgitating some information to me, and then did become familiar with Chat Gpt last year through my husband and his work and some things. Some ways he was stepping into that area. 99 00:19:01.090 --> 00:19:23.310 Amanda Bickerstaff: Yes, and I I think that chat bots were pretty bad for a really long time, and they suddenly got pretty early. Good which was, although some of them are still really really bad. I think we can all agree to that. And then, Andrew, you've already introduced yourself, but also like what was your first? I know you kind of went all in a bit that we met at a for those that don't know. We met at a think tank around this. I know you've been thinking about it a lot. 100 00:19:24.310 --> 00:19:36.958 Andrew Buher: Yeah, I have. But it's funny. Amanda. So my wife and I were talking at dinner maybe a week ago, about sort of like the inverse relationship between age and first experiences. 101 00:19:37.758 --> 00:19:57.079 Andrew Buher: You really have to like, try hard to create new experiences as you get older, I don't remember interacting with generative AI for the first time. So I certainly missed that first experience. But you know, shortly after the this sort of boom, a couple of winters ago. 102 00:19:57.791 --> 00:20:03.159 Andrew Buher: There was someone that effectively rewrote 103 00:20:04.406 --> 00:20:11.190 Andrew Buher: a national education law using chat gpt, and so I remember that sticking out very prominently. 104 00:20:12.540 --> 00:20:35.055 Amanda Bickerstaff: Oh, yeah, I think there was a lot of that ve very much in in the beginning days where we started to see some content that probably shouldn't have been used chat fee should have been used, or at least a clo disclosed. So all very interesting. I I we use this kind of as our workshop test, because we all come to it in different ways, and we all are. I think this has become part of our lives in such a a unique and intentional and unintentional way, right? Because it's something that's 105 00:20:35.290 --> 00:20:47.079 Amanda Bickerstaff: been what we talk about. If you're on Linkedin, you hear this all the time right? It feels like it's ever present. But it is really interesting, because I think just to ground us back into the research, there is still an enormous knowledge gap 106 00:20:47.080 --> 00:21:12.050 Amanda Bickerstaff: in schools around these technologies and the potential impact. And I think that that's a big part about what we need to think about going forward. But I think one of the the most interesting things is that AI is especially chat bots are designed to feel like they are human. There's a human or human like intelligence on the other end. And I think that that's what I really want us to focus on this first question about like student. Well, being, we're gonna talk specifically around 107 00:21:12.050 --> 00:21:29.180 Amanda Bickerstaff: AI, not just AI as a larger field, but a lot of these use cases of like character, AI or Snapchat, AI, or other tools that students are using. I would really like to understand what you think the actual impact on student well being could be with with these generative tools. Starting with Marina. 108 00:21:31.730 --> 00:21:55.199 Dr. Marina A. Badillo-Diaz: So, yeah, I mean, I definitely think there's gonna be a a wide sort of range of the social, emotional impact of the of the use of these generate AI tools? And again, I I wanna also acknowledge our youth are our most vulnerable. They are in their periods of life, where their brains are still developing. They're still developing socially. And so as they're using these tools as a young person, I think there's a lot of 109 00:21:55.630 --> 00:22:20.489 Dr. Marina A. Badillo-Diaz: unintentional like impact this could have, I think, both. And we're gonna talk about this in the webinar, both for positive use, but also for just some ethical and just and again, social, emotional concerns. So I think for me, what I think sort of the impact can be on sort of the the negative end is increasing isolation, increasing anxiety increasing. You know, depression, I think, about the potential, the impact of 110 00:22:20.550 --> 00:22:27.850 Dr. Marina A. Badillo-Diaz: of bullying, and how that could really be upped and amt in in cyber bullying. Through these uses and tools of generative AI. 111 00:22:28.820 --> 00:22:53.589 Amanda Bickerstaff: Yeah. And I mean, I think we we see that right, like deep fakes non consensual fake nudes. You know the ability to, not just bully like kids, but like a adults as well, is like a really big, like meaty, potential impact that we already have seen right? Like, we don't have to wait for that to happen, and it seems to have just fallen into patterns that already existed in our schools before. But now they're being hyper charged because that 112 00:22:53.590 --> 00:23:10.299 Amanda Bickerstaff: that ability to create things that feel real is is so much at our hands now. So I think, thank you for calling that out, and we'll actually, I think Amanda, from our team can drop in the deep fax piece that we did recently to to that point as well. Rebecca, do you like? What do you think the impacts gonna be and suit well, being, or has already been. 113 00:23:11.160 --> 00:23:24.116 Rebecca Oliver: Yeah. You know, we know. Andrew touched on this in his intro, that student, mental health issues have been increasing. And and you know, at alarming levels to where we're calling it a youth mental health crisis. 114 00:23:24.440 --> 00:23:45.399 Rebecca Oliver: And so I think, as Andrew said, if there's going to be any impact, even if it might be slight. We need to look at it very carefully. So, and and a lot of the things Marina mentioned as well, you know that we're working with our most vulnerable. We, as adults, have to be the ones that step into that space and get the information and help youth. 115 00:23:45.895 --> 00:24:09.200 Rebecca Oliver: You know I know I've seen some things that I thought were real, and my husband was like that's not real. That was made up with AI, you know. And if, as an adult, I'm not able to discriminate and and use other knowledge, I have to determine what's true, what's not true, how can we expect young people to do that? So I think we have to step into that space to make sure our children are safe 116 00:24:09.260 --> 00:24:22.630 Rebecca Oliver: and that we need to ask the questions and be really critical in thinking about the impact or the potential impact. And I think, as we've already said, a lot of it's unknown. We, you know, we really don't know. 117 00:24:23.380 --> 00:24:27.049 Amanda Bickerstaff: Yeah. And I I I love that idea, though, about like 118 00:24:27.110 --> 00:24:50.540 Amanda Bickerstaff: starting to think through that now. Like, if it like, it's like one of those pieces where it doesn't have to be negative. But like should we be looking at if it is like? And actually, instead of like social media? What if we had actually taken a moment? Said new technology in kids hands? What would that mean to their mental health and well being? How much different the world would be right now potentially for these young people. So it's not even like a deficit mindset like, if this is going to be in the hands of students that are using it for 119 00:24:50.970 --> 00:25:08.810 Amanda Bickerstaff: like AI girlfriends or chat bots like pie that were brought up by someone in the audience around that are, they're designed to be emotional support like these are going to have impacts because it's never happened before. We've never had synthetic relationships. Or, you know, artificial intimacy before. Right? So like. 120 00:25:08.810 --> 00:25:25.370 Amanda Bickerstaff: let's not not put pay attention to it from the perspective of student mental health and wellbeing. And it and I think, adult as well, so to you, Andrew, like, what do you think about in terms of that impact like, based on both the the research? But also, I know you have some pretty strong statements about this moment. In time we need to stop. 121 00:25:25.460 --> 00:25:28.990 Amanda Bickerstaff: reflect, and be intentional. So you talk a little bit. That would be great. 122 00:25:29.270 --> 00:25:40.729 Andrew Buher: Yeah, sure. So a a couple of of points and and some of these are exclamation points on what Rebecca and and Marina just said. But you know, I think obviously, there is 123 00:25:40.750 --> 00:26:09.969 Andrew Buher: significant potential upside from a teaching and learning perspective from a school operations perspective and potentially even from a social connection perspective. If these tools are used. And when I say these tools, I'm specifically talking about tools that are powered by generative AI are used within a set of really thoughtful guardrails that are optimizing for kids wellbeing. 124 00:26:10.331 --> 00:26:22.258 Andrew Buher: But I think that the the knotty thing that we have to untangle is that there's also risks right associated with these technologies, and they're hard to qualify and quantify right now. 125 00:26:23.063 --> 00:26:29.929 Andrew Buher: But I think we are increasingly convinced that if we don't attempt to mitigate those risks upfront. 126 00:26:30.040 --> 00:26:52.449 Andrew Buher: we never actually get to the full benefits of these technologies, because folks will begin to lose enthusiasm around them. Right? And they'll be harder to adopt at scale. In schools and and outside of schools. Frankly, so, I sort of think about like 2 questions. Here. One is extremely 127 00:26:52.560 --> 00:27:01.749 Andrew Buher: practical or practical ish, and the other is pretty theoretical, and maybe bordering on the verge of philosophical. So the the first 128 00:27:01.910 --> 00:27:11.379 Andrew Buher: question is, really, how can we leverage transitive AI tools to enhance social connection? Not erode it? 129 00:27:12.175 --> 00:27:28.729 Andrew Buher: And I think there's some early indications that, like there are some use cases where where these tools can be used to to have students working, not independently, but in in small groups, right? And reflecting on each other's work. 130 00:27:29.269 --> 00:27:42.060 Andrew Buher: and so on. The philosophical question, one that I've been thinking about, and I wouldn't say keeps me up at night yet, but it's sort of trending in that direction is 131 00:27:42.648 --> 00:27:44.989 Andrew Buher: what does it mean for 132 00:27:45.030 --> 00:27:52.489 Andrew Buher: children? To live in a world where it is very hard to distinguish what is true and what is not 133 00:27:52.680 --> 00:27:57.579 Andrew Buher: right. What does that do to kids experience? And how does that inform 134 00:27:58.100 --> 00:28:18.120 Andrew Buher: their mental and social and behavioral wellbeing. And so that is a question that we are trying to parse. But I think, Amanda, it goes right back to your point about deep fakes and misinformation as the most prominent negative use case that's popped up so far. 135 00:28:19.660 --> 00:28:43.630 Amanda Bickerstaff: Yeah. And this is a really interesting 1, one of like character. AI, so and we're there are lots of student facing tools that are that are that are either intentional or not intentional. So you've got your magic schools, your school ais, or missus that are chat bots that are supposed to be pedagogically focused. And then you've got the ones that are more like students find on their own right the consumer technology. So the Snapchat ais etc. 136 00:28:43.660 --> 00:29:07.590 Amanda Bickerstaff: But I think the most interesting one is Character AI. So character AI is A is a tool that has 18 million avatars that you can talk to, and you can actually create your own. So you could talk to Marie Curie or Naruto, or you know someone from, you know, you could make it up. And what's really interesting is that they they're very popular with 16 to 24 year olds, and even somewhat younger students with the average use of around 2 h. 137 00:29:07.710 --> 00:29:25.919 Amanda Bickerstaff: and then what they see is that you know the they have this new feature called a group chat feature, which is a combination of real people and bots. But the thing that is so fascinating to me is that the only way that you know what is real or an avatar run by a generative tool is if you go in the group settings. 138 00:29:26.110 --> 00:29:39.589 Amanda Bickerstaff: So they did like it is indistinguishable. And it is actually something in which the group itself is starting to blur the lines between real and not real. And so I think that these and that is what's popular. That's 20 of the site traffic of Chat. Gbt. 139 00:29:39.590 --> 00:30:04.570 Amanda Bickerstaff: so it actually is something that I think is going to be really interesting. But I do think, as we start thinking about these tools like we might think of them as generative tools that we bring into schools. But I think, really, our focus right now is what kids are bringing with them that we may not know that they use. And so our next question is, gonna be like a a little bit of a combo. So we're gonna look at this idea of what do you actually think is your biggest fear or concern? But then, what can we do as social work. 140 00:30:04.570 --> 00:30:13.729 Amanda Bickerstaff: or those that are in and around students to start mitigating that. And the answer may be, we don't know exactly yet, but I would love to get your perspective on what you think we can do. 141 00:30:14.270 --> 00:30:15.500 Amanda Bickerstaff: Sorry with Marina. 142 00:30:16.020 --> 00:30:39.678 Dr. Marina A. Badillo-Diaz: Yeah. So I'm again, thank you for raising this really important question, because we do have to acknowledge the fear and have these conversations more deeply and so definitely. My biggest concern and concern is that we know meant to help us on the rise. Adolescents, suicide, suicidality rates are on the rise. And now I think about generative AI and bullying, and how this could continue to exacerbate that. So that's definitely my biggest fear is the cyber bullying and the impact on 143 00:30:39.960 --> 00:30:48.449 Dr. Marina A. Badillo-Diaz: depression and and suicidality. So I think a real response that we have to talk deeply now in this point in time, with 144 00:30:48.791 --> 00:31:13.389 Dr. Marina A. Badillo-Diaz: with our school administrators, with school psychologists, school counselors, school social workers and our educators is thinking about how we're gonna be addressing and providing crisis intervention supports for those students who are impacted by these deep fakes. You know, this is this is, gonna be a a real source of of concern. And so I think about again the intervention aspect as a school social work, of really 145 00:31:13.390 --> 00:31:36.140 Dr. Marina A. Badillo-Diaz: supporting their mental health and connecting them to resources, but also addressing this as a school community so that we can create safe learning communities for all. And so I think about that more preventative stance as well, that Us. School counselor, school social workers. We need to have these conversations with our students about responsible and safe use and also addressing cyber bullying and the impacts that has on, on Mental health. 146 00:31:37.060 --> 00:31:52.770 Amanda Bickerstaff: Oh, man, I love that. I mean, I think it's it's like, sometimes we wait till something happens, and then we create the plan. It's very reactive and not proactive. But if we already know again, it's already happening. This is how kids interact. Already. This is something that could exacerbate this 147 00:31:52.770 --> 00:32:17.720 Amanda Bickerstaff: like, let's be really, really intentional about doing that first, and it's being anticipatory. So like for the next like, say, school year is about to end. But like, do you have a plan for this before the next school year? Do you have a training for students? Is there a decision tree of next steps? Are there examples like, I think, that that is such a unique and positive conversation to have. And it kind of goes next to guidance. But it's even different than that right? And I think that that could be. It's super super 148 00:32:17.720 --> 00:32:29.289 Amanda Bickerstaff: tactical, which I love, Marina. But I do think that that could be something that would really really be helpful for next year and going forward. So that's really awesome. So, Rebecca, what do you think? What is your biggest concern? And then how can we mitigate that. 149 00:32:30.000 --> 00:32:59.420 Rebecca Oliver: Yeah. So it's hard for me to come up with just one you know, I wanna elevate a little bit of what school social workers that were included in this data. What some of those issues and fears and concerns that came up, things that Andrews touched on. You know, concern about the impact on mental health, as Marina mentioned, and also how it interacts with the cognitive development of these young people, potential exposure, inappropriate content to inaccurate content to biased content. 150 00:32:59.862 --> 00:33:17.569 Rebecca Oliver: You know those concerns potential cyber bullying and confidentiality risks. So there are a lot of them. And I think that, coupled with again indications from the data that many school staff are concerned, that families don't have the information they need to help. 151 00:33:17.963 --> 00:33:45.929 Rebecca Oliver: Put those guardrails around their their young people. And also again, that school social workers and other school professionals aren't being included as much as needed in those conversations. I think those are my biggest takeaways. Just that. There is such an immense amount of information in this area, and many parents and families are just trying to get food on the table and work and get their children to appointments. 152 00:33:45.930 --> 00:33:58.900 Rebecca Oliver: and having the time to wade through. All this information is challenging. And so we just need to bring, as many people like Andrew said, that are close to children. Into these conversations. 153 00:33:59.250 --> 00:34:28.960 Amanda Bickerstaff: That's great. And I just wanna call out Patty in the the chat he's also talking about like doing an AI literacy event. Not just. For like teachers and students with families like that could be a wonderful way to start to build those relationships, too, cause one of the things that is really unique. Part of this is that we can help te like parents actually help their kids more with these tools, the same that there's productivity gains to be had for teachers using these tools effectively could be the same for students. And I think equipping parents with that, too, could help with some of those 154 00:34:28.960 --> 00:34:45.610 Amanda Bickerstaff: is those deeper connections that we sometimes miss as social workers. Right? That, don't we? We have trouble getting that that cycle going and really getting that engagement with parents for a long period of time. So I think there are some other opportunities that could be done as well, which is super super great. So, Andrew, yeah, go ahead. Go ahead. 155 00:34:45.610 --> 00:35:14.799 Rebecca Oliver: Sorry I mean just to, because I didn't really sweep around to your last part of that question of how school social workers can assist with this, and I think school social workers very much are a link between home school and community. And so school social workers are uniquely positioned to help draw parents into those conversations, to put together resources, to maybe offer trainings, or find those trainings that can be offered. And so I think they can be a really impactful part of the conversation. 156 00:35:15.000 --> 00:35:16.020 Amanda Bickerstaff: Absolutely. 157 00:35:16.290 --> 00:35:18.599 Amanda Bickerstaff: And so, Andrew, to you. Same question. 158 00:35:19.130 --> 00:35:25.769 Andrew Buher: Yeah, sure. I might filibuster a little bit here, Amanda. First, just on the question in itself. 159 00:35:25.820 --> 00:35:34.930 Andrew Buher: I think we should avoid using the term fear, right? Probably a disservice for kids, just like treating generative AI 160 00:35:35.477 --> 00:35:42.982 Andrew Buher: powered products and tools. As a silver bullet which we know just doesn't exist in education. So 161 00:35:43.790 --> 00:35:53.099 Andrew Buher: I am unclear about the impacts. So I'm gonna talk about what I'm concerned about more broadly, which is speed 162 00:35:54.540 --> 00:35:55.470 Andrew Buher: and 163 00:35:56.236 --> 00:35:57.529 Andrew Buher: I think 164 00:35:57.600 --> 00:36:02.019 Andrew Buher: we have a tendency across the education ecosystem 165 00:36:02.660 --> 00:36:07.699 Andrew Buher: to operate under the adopt or be left behind 166 00:36:07.750 --> 00:36:10.510 Andrew Buher: paradigm, which 167 00:36:11.110 --> 00:36:18.969 Andrew Buher: I think the evidence would indicate is probably a fallacy, particularly as it pertains to education, technology, tools. 168 00:36:19.714 --> 00:36:24.859 Andrew Buher: But what that paradigm does is it forces folks to 169 00:36:24.930 --> 00:36:30.347 Andrew Buher: act to adopt emerging technologies at scale too quickly. 170 00:36:31.110 --> 00:36:35.610 Andrew Buher: And so, I think a better approach. A more practical approach 171 00:36:35.690 --> 00:36:39.200 Andrew Buher: would be small, contained. 172 00:36:39.280 --> 00:36:43.480 Andrew Buher: carefully evaluated classroom-based pilots. 173 00:36:44.390 --> 00:36:45.869 Andrew Buher: That are 174 00:36:46.576 --> 00:36:54.449 Andrew Buher: staffed by cross functional holistic teams, right educators, school-based health staff administrators. 175 00:36:55.750 --> 00:36:57.010 Andrew Buher: and 176 00:36:57.440 --> 00:36:58.549 Andrew Buher: and I think 177 00:36:58.640 --> 00:37:01.649 Andrew Buher: all the folks participating in those pilots. By the way. 178 00:37:02.165 --> 00:37:19.620 Andrew Buher: should have really deep, robust, rigorous AI literacy training, so that they are well informed about what they are going to encounter. As they begin to engage with these tools and technologies. I think the the real challenge here, right is 179 00:37:19.720 --> 00:37:30.360 Andrew Buher: there aren't a lot of pieces of evidence around efficacy related to outcomes of any of these emerging technologies. Yet. 180 00:37:30.830 --> 00:37:38.579 Andrew Buher: and so I think it would be really compelling and potentially, really helpful for the technology itself 181 00:37:38.900 --> 00:37:55.190 Andrew Buher: to get some proof of concept that these tools can really move the needle on student outputs and student outcomes. And I think the best way to do that is, by these small contain pilots that are really carefully evaluated and staffed by cross functional teams. 182 00:37:55.430 --> 00:38:10.950 Amanda Bickerstaff: Yeah. And I, I think what's really fascinating. So like Microsoft right now is is, do is setting up a a pilot or the call to get a private peer review, because it can't be a co-pilot pilot, apparently. For 13 to 17 year olds using co-pilot. 183 00:38:11.020 --> 00:38:39.590 Amanda Bickerstaff: But what's really interesting is that it's so tech led like in the same thing, like, I'm not sure what happened with magic school or school AI or others like we don't have any evidence or any structure or support to actually know how these these pilots are are run before they're released widely. And I think that we had a question about the tools that are pedagogical. And I'm gonna use this because Jen AI is not pedagogical. Everybody, even with the best fine tuning and support. It's not a thinking machine. It's not a human. It's not a teacher. 184 00:38:39.590 --> 00:38:50.030 Amanda Bickerstaff: But I do think that there's a a question of how do we actually start identifying those tools that we know are already impacting students and building space for those cross functional teams because 185 00:38:50.030 --> 00:39:15.030 Amanda Bickerstaff: it has. We have this bad habit of of like not coming together when something new happens and actually creating evidence in a non siloed way. It's happening in tech. It's happening in schools and districts, but in very uneven ways. So maybe this is a call to action to actually have something that is like, we're gonna test the top 5 student facing products and actually see what this means. And we're going to include a mental health and wellbeing component, not just a pedagogy, not just an engagement, not just 186 00:39:15.030 --> 00:39:37.350 Amanda Bickerstaff: student outcome learning outcome piece. I think that would be really unique. And I think they're they're because we're all interested in this. Maybe there's an actual opportunity for it to happen instead of just being talked about. So that's really interesting. Next question is more, this is our last big question. But what do you actually think is the positive that could happen around student wellbeing. There's been research from a typical and others that show that 187 00:39:37.350 --> 00:39:52.719 Amanda Bickerstaff: suit like people are using these tools from mental health and wellbeing. There's been at least I typical said that there was a lowering of social suicidal ideation for some adults using these tools. We know they're using them so like, what do you think is the opportunity for student while being supported by Jen AI. 188 00:39:54.640 --> 00:40:19.569 Dr. Marina A. Badillo-Diaz: So I think about this in 2 ways of the 2 opportunities. So one is actually the practitioner using these tools to support student well-being. So I think about where I role as a school social worker. We know we have high case loads. It might be only one school social worker for several grades or in a school setting. So I have multiple multiple students that I'm supporting for their mental health. So to have these time saving tools to help me be more efficient at my job. 189 00:40:19.570 --> 00:40:42.619 Dr. Marina A. Badillo-Diaz: I could use generative. I to help me create intervention, plan templates. I could have it help me generate ideas for counseling activities in in my office. I could also use it to help create, you know, campaigns for suicide, awareness and anti bullying campaigns. I think they're just immense potential opportunity for these tools really support and enhance the mental health services that our practitioners, like school social workers. 190 00:40:42.620 --> 00:40:57.879 Dr. Marina A. Badillo-Diaz: cool school counselors, have in schools. And then the other piece is something else that we've been talking about, which which is using AI as the companion to really support youth, mental health. So I think about the app like alongside which is a 191 00:40:57.950 --> 00:41:22.060 Dr. Marina A. Badillo-Diaz: chat Bot generated for specifically for adolescents to use alongside with schools. And it's it's created and developed by clinicians. And this sort of chat bot is available in partnership with schools for students to use it. And what I also really appreciate about this model is that students are accessing it to ask questions, to talk about their problems, that they're experiencing. 192 00:41:22.374 --> 00:41:46.940 Dr. Marina A. Badillo-Diaz: Sometimes they just need someone as like a sounding board to validate their feelings. But once it crosses the threshold of a student is reporting ideation that they're suicidal, that they're in crisis. That's when they contact the school. So if there is that human crisis element component of that partnership, so the school staff can then support that that child who's in need, and who's in crisis. So I think about how this won't be replacing our 193 00:41:46.940 --> 00:42:06.320 Dr. Marina A. Badillo-Diaz: school social work role. But it really will enhance the opportunity for us to support. And I think about equity and access. This can be a real potential to support students who may not have access to a school social worker and and counselor, but then also doesn't eliminate the human concern of of the crisis piece. So I think there's a real opportunity there. 194 00:42:06.650 --> 00:42:27.219 Amanda Bickerstaff: Yeah. And I mean, we are so understaffed, right? Especially in high schools around counselors and social workers. That is this a place to be able to to do more of what we have and and uniquely put in the human component. We did. Re, we built a well being tool in Australia. That was my last job into research. And we we found exactly what alongside finds that, like students having the ability to ask for help 195 00:42:27.220 --> 00:42:41.780 Amanda Bickerstaff: and or to identify it even on their own. And then also, even if they don't, having that come and be lifted up and and prioritize and triage to a human is a really really great way to like, actually identify those students that have problems that you do not see. 196 00:42:41.780 --> 00:43:02.860 Amanda Bickerstaff: And I think that this is where, like the the ability to uncover what's below the surface, which we know is has happening, I think, is a really unique opportunity, with with a generative AI tool like alongside, and we love alongside. So that's that's really great. So, Rebecca, to you, what do you think is the the best use case for Jen AI. And how it could potentially support student well, being. 197 00:43:03.250 --> 00:43:24.474 Rebecca Oliver: So I have to say, it's funny. Marina and I have been in multiple meetings even this week together, and we did not talk about our answers to this, but she took a lot of she took a lot of the things that I wanted to say so. But I do again. I think a big thing is we do know, that there are a lot of time consuming tasks and understaffed at schools. 198 00:43:24.750 --> 00:43:36.609 Rebecca Oliver: And so there's the potential for these tools to help us be able to reach more students and be able to connect with more students. So again, I think that that the barrier piece of that question is 199 00:43:36.610 --> 00:43:45.619 Rebecca Oliver: just ensuring that we're at the table to evaluate the various tools, to ask the questions, to determine which ones can be used in which ways. 200 00:43:45.620 --> 00:44:01.650 Rebecca Oliver: and ensuring that we still have, like Marina, said, that human touch of a real human that evaluates some of that critical information. And is that that human point of contact with the students that need the support. 201 00:44:02.340 --> 00:44:27.319 Amanda Bickerstaff: That's great. And I I first of all, I'm so glad that you're having a lot of conversations about this Rebecca and Marina, though, cause. I it makes me so happy that people are asking and listening, and the fact that we have over 200 people hanging out with us at a very last, you know, like it's getting to last weeks of school kind of life is a good indication. But I do. It is interesting that you have so much overlap there, because I think that that's where we can do a focus right? And like it has something that we could start. 202 00:44:27.320 --> 00:44:46.290 Amanda Bickerstaff: And then we could start to to move from there. So Andrew's always gonna have one of the things I love about Andrew is that he is willing to be bracing and also filibuster and all these things. So I would be really interested in knowing what you think are the actual kind of impact potential, positive, most positive impacts of Jenna and Zoom. Well, being. 203 00:44:46.520 --> 00:44:50.341 Andrew Buher: Yeah, so maybe this is gonna be counterintuitive coming from me, Amanda. But 204 00:44:51.220 --> 00:45:02.370 Andrew Buher: you know, I think like, if you broadly agree that one of the primary purposes of school and schooling is to build both social skills and social connection for kids. 205 00:45:02.961 --> 00:45:07.540 Andrew Buher: Which we know has obviously significant bearings on on overall wellbeing. 206 00:45:07.966 --> 00:45:12.729 Andrew Buher: Then I think there's a real opportunity for generative AI to help 207 00:45:12.830 --> 00:45:14.699 Andrew Buher: right? You know, I think most 208 00:45:14.750 --> 00:45:19.000 Andrew Buher: social media media is actually anti social media. 209 00:45:19.160 --> 00:45:35.410 Andrew Buher: But there are corners of those ecosystems that actually have created really rich opportunities for kids to find other kits right? And so I think if we can sort of replicate 210 00:45:35.560 --> 00:45:43.849 Andrew Buher: the good that is happening in pockets and corners of the social media ecosystem 211 00:45:44.000 --> 00:45:53.979 Andrew Buher: and figure out what analogs exist that potentially can be amplified, expanded, leaned into 212 00:45:54.490 --> 00:46:06.540 Andrew Buher: with generative AI powered tools and products. Then there's a real opportunity. Not only to help kids build social skills and awareness. But social connection also. 213 00:46:08.270 --> 00:46:26.000 Amanda Bickerstaff: I think that's great. And yeah, I think you're right. I mean, it's we, even, I mean everything that we think will be like, fully bad for kids like like Covid was going like their kids. That's that flew flew with online education. And like, we'll continue to fly with something different. So I think it's absolutely 214 00:46:26.000 --> 00:46:50.880 Amanda Bickerstaff: true about the of all things. I'm just gonna I'm gonna I'm gonna insert this one thing. I think that we are under like, there's an opportunity for neuro divergent support like hit supporting kids with nerd divergence that's really really high. Whether that's students that might not feel comfortable with written communication because they're dyslexic, or that have unique needs around sensory issues or have not been able to communicate at all. Like because they haven't. 215 00:46:50.880 --> 00:47:15.870 Amanda Bickerstaff: They're they're considered non verbal to be able to now create images, videos, the ability to have an avatar of yourself that could could express things that you've never been able to express, I think, is something that could like significantly impact student wellbeing. I think that if we I would, I would love us to see more in those spaces. First, as a used case, because I think that the the we could really do pilots on that. But also the impact 216 00:47:15.870 --> 00:47:40.689 Amanda Bickerstaff: that could be really, really large, really, really quickly. I mean just the giving of someone, a voice and never had. One is like a unique place that we could do this work together. So we're coming up on time. So we have our last. I know we fixed it all, everybody soon. Well, beings fixed. We got it. But I do wanna stop with like, like, we always like to be as practical as possible, and so like, what is your most practical strategy 217 00:47:40.690 --> 00:47:52.040 Amanda Bickerstaff: that you want people to take away with you in terms of supporting student. Well, being this time, or kind of starting to gather evidence, or whatever you think is most important. So, Marina, what is your what's your idea? 218 00:47:52.540 --> 00:48:17.509 Dr. Marina A. Badillo-Diaz: So my idea really comes from stemming and creating a multi tiered support plan for your for your schools are really thinking about student well being and AI adoption. And in that multi tiered sort of support plan, really thinking about having a cross disciplinary conversation with administration. The It tech person, the the librarian, yours, your school support staff like your school counsel. 219 00:48:17.510 --> 00:48:41.250 Dr. Marina A. Badillo-Diaz: social workers your your educators, and really coming up with tiered interventions from the crisis support to thinking about more of the preventive place of bringing in families into this conversation family workshops, but also student lessons and Scl lessons on digital citizenship and but also cyber bullying. So I think there's a huge opportunity to really create this more thoughtfully 220 00:48:41.250 --> 00:48:44.850 Dr. Marina A. Badillo-Diaz: of a sort of like a multi-tiered intervention plan. 221 00:48:45.360 --> 00:48:59.530 Amanda Bickerstaff: That's great, and if you want to, if you're writing it, we'll support it. If you want to write it together. We're ready. I think that you know, having those resources in hand for the start of the school year that people can use is is an amazing idea. Rebecca. 222 00:49:00.260 --> 00:49:28.989 Rebecca Oliver: So mine's kind of 2 fold one is is that piece of what Marina just touched on making sure people are coming together to ask the questions, to have conversations with families. You know. I think, Amanda, something you had said in our our pre discussion about this webinar is, you know, I think it's also important that we create a safe space where parents can say, I don't know, like, what are my kids interacting with, what are these different things. And if because if we don't give that space. 223 00:49:29.370 --> 00:49:54.070 Rebecca Oliver: you know, then they're not going to come to the table, and and we're not going to be able to have a full discussion. So that's one part of it. And then I think my other piece is balance is going back to what we know is good for kids. Wellbeing in general. Right? All those basic things of making sure students are having balance, rest and sleep, that they're having time outdoors 224 00:49:54.070 --> 00:50:05.500 Rebecca Oliver: that they're in green and blue spaces, that they're drinking water, you know, right? Staying hydrated the basic things connecting with their family and friends, disconnecting from screen time. 225 00:50:05.570 --> 00:50:16.529 Rebecca Oliver: and then learning all sorts of different coping skills that help them, mindfulness and breathing and different things that help them manage and keep balance in their life. 226 00:50:18.000 --> 00:50:19.690 Amanda Bickerstaff: First of all, like. 227 00:50:19.990 --> 00:50:30.890 Amanda Bickerstaff: we see this in the AI literacy work and adoption work we do that seems like we need to throw out the rule book because it's happening so quickly. But I think we need to lean into the role book more. 228 00:50:30.890 --> 00:50:55.869 Amanda Bickerstaff: And you know those spaces because that's gonna give us this, the opportunity to like, do the foundations, and then and then support the ways we know, because it's moving so quickly that if we try to keep up with that and try to change and everything. But like, let's not. Let's do the things that we know work, provide people spaces to be able to learn together to be if I know if you didn't like the word fear. But some people are afraid. But maybe it's okay to be afraid in ways that are 229 00:50:55.870 --> 00:51:08.890 Amanda Bickerstaff: intent like that are not spiraling, but intentional safe spaces that lead to dialogue and next steps. But I think that that's a way to get our hands wrapped around it like the same thing with Marina like, let's do the things we know that really work. 230 00:51:08.960 --> 00:51:20.559 Amanda Bickerstaff: and then talk about how we support this movement with generative AI instead of the other way around, I think, is a really good piece of advice. So, Andrew, you're gonna bring us home. What is your most practical strategy? 231 00:51:21.157 --> 00:51:25.672 Andrew Buher: I'm gonna try to bring us home 2. I would offer to 232 00:51:26.280 --> 00:51:28.450 Andrew Buher: 2 things. One is, I think. 233 00:51:28.550 --> 00:51:35.670 Andrew Buher: following the evidence becomes incredibly critical here, right with the technology advancing so rapidly. 234 00:51:37.463 --> 00:51:43.099 Andrew Buher: understanding what the evidence says, and asking very specific questions 235 00:51:43.130 --> 00:51:51.010 Andrew Buher: about any technology that is being adopted and will be student facing is super critical right? And so those are questions like. 236 00:51:51.634 --> 00:51:58.749 Andrew Buher: what evidence exists that the technology is successfully doing what it is intended to do? 237 00:51:59.339 --> 00:52:03.129 Andrew Buher: What evidence exists that the technology will do no harm? 238 00:52:03.360 --> 00:52:06.200 Andrew Buher: What evidence exists that 239 00:52:06.220 --> 00:52:13.176 Andrew Buher: the builder of these technologies has optimized for student well-being. Okay? And then my second 240 00:52:13.660 --> 00:52:18.113 Andrew Buher: suggestion is a maybe sort of 241 00:52:18.840 --> 00:52:23.439 Andrew Buher: very, very non sexy one. But I think 242 00:52:23.720 --> 00:52:30.220 Andrew Buher: if we are trusting big tech great. You shouldn't have any concerns. 243 00:52:30.280 --> 00:52:39.540 Andrew Buher: If you don't trust big tech then we should be thinking about how we get builders of these products 244 00:52:39.700 --> 00:52:45.220 Andrew Buher: to optimize for student wellbeing, and I think the best way to do that is through procurement. 245 00:52:45.530 --> 00:52:47.900 Andrew Buher: So I would be encouraging folks 246 00:52:47.920 --> 00:52:53.440 Andrew Buher: to get their procurement officers, their districts. To think about building 247 00:52:53.730 --> 00:53:07.239 Andrew Buher: fundamental benchmarks across efficacy and data, privacy and security and interoperability and access. Among other things, that 248 00:53:07.540 --> 00:53:13.130 Andrew Buher: technology providers would need to meet to sell into schools. Right? I think the only way we shift 249 00:53:13.659 --> 00:53:22.740 Andrew Buher: movement towards child. Well, being, when we're building these products is by changing incentives, and the biggest opportunity we have to change incentives in K. 12 250 00:53:22.930 --> 00:53:24.460 Andrew Buher: is our purchasing power. 251 00:53:24.750 --> 00:53:41.790 Amanda Bickerstaff: Okay, yeah. And and I agree. And I think that what you just said, though, isn't, it goes way beyond generative. AI just goes to like, we need to do better with our technology and schools. And I think that that's just across the board. A really important piece. But this might be what's really cool about this moment time. This could be the lever 252 00:53:41.790 --> 00:54:06.780 Amanda Bickerstaff: for getting that for more than just generative AI, because people are paying attention ways. They've never paid attention before. And so I think that's a really great point. I just wanna say, thank you like, can we just give everybody? I know we can't do hearts. But I'm gonna do a heart for everyone that's in the audience today. Thank you so much to our amazing panelists. For also just Andrew and the team at opportunity labs. Rebecca, everyone, Swa Marina, like we just really appreciate you putting this time and effort and evidence into the world 253 00:54:06.780 --> 00:54:31.739 Amanda Bickerstaff: at this important time, like, it's just really, really, really important. And we want to hear more of this. And so I love that that we're able to have this discussion, and this also. Just thank everyone for coming. I mean, we kept, you know, almost the whole group the whole time, a great conversation in the chat. And we just wanna continue this work. So please, like, if you have good ideas and resources, we're always happy to hear them. But I just wanna say thank you to everyone for for taking a moment out of your time to think about students in a meeting. 254 00:54:31.740 --> 00:54:51.210 Amanda Bickerstaff: Meaningful way like that means a lot, and I think that it shows a lot for what we can do to bring that back to our schools and system worldwide at this point. So thanks everybody. I hope you have a good morning. Night lunch. Go to bed. Whatever time it is. We appreciate you all, and just we look forward to having you all for our next webinar in a couple of weeks. Thanks, everybody. 
