Schools are using AI counselors to track students’ mental health. Is it safe? | AI (artificial intelligence)

by Chief Editor

AI as School Counselor: A Growing Trend, But at What Cost?

The alert arrived around 7pm. Brittani Phillips, a middle school counselor in Putnam County, Florida, received a notification from an AI-powered therapy platform used by students outside of school hours. The system flagged an eighth-grader as potentially at risk of self-harm. Phillips spent the evening on the phone with the student’s mother, assessing the situation and, contacting the police. The student is now in ninth grade and, according to Phillips, greets her in the hall – a testament to the trust built through the intervention.

The Rise of AI Mental Health Tools in Schools

Interlachen Jr-Sr High School, where Phillips works, is among a growing number of schools turning to AI to address student mental health needs, particularly in the face of budget shortfalls and limited staffing. The school utilizes Alongside, an automated student monitoring system, which is one of at least nine companies receiving funding since 2022. Alongside boasts over 200 school clients and combines a social-emotional skill-building chat tool – featuring a llama named Kiwi – with AI-generated content monitored by clinicians. The platform aims to provide critical resources, especially in rural areas.

AI is as well becoming a component of national education agendas. Whereas, this increased reliance on technology is meeting resistance from parents, educators, and lawmakers concerned about screen time and student privacy. Some states are beginning to restrict the use of AI in telehealth.

The Appeal of a Digital Ear

Experts suggest students are increasingly comfortable confiding in AI. Speaking with a mental health professional can be intimidating, particularly for adolescents. For a generation accustomed to chat interfaces through social media, AI can feel familiar, and accessible. Students may also avoid the perceived judgment of facial expressions or the inconvenience of scheduling appointments. As Sarah Caliboso-Soto, assistant director of clinical programs at the University of Southern California, notes, “It’s almost more natural than interacting with another human being.”

The demand for accessible mental health support is clear. Crisis text lines and chat lines have seen a rise in usage. While some see AI as a valuable “first line of defense,” regularly checking in with students and directing them to further help, others caution against replacing human counselors.

The Risks and Concerns

A key concern is the potential for students to develop “parasocial relationships” with AI, forming one-sided emotional attachments. Young People’s Alliance advocates for rebuilding human community and cautions against AI use that threatens human companionship. The organization suggests AI should provide feedback or analysis, but avoid conveying emotional states that could encourage attachment.

Privacy is another significant issue. AI chatbots may not offer the same privacy protections as conversations with a licensed therapist, raising “messy” concerns, especially when police involvement is possible.

Caliboso-Soto emphasizes that AI lacks the discernment of a human clinician, unable to interpret subtle cues in voice and body language. “You can’t replace human connection, human judgment,” she adds.

Finding the Right Balance

Alongside representatives maintain their platform is not intended to replace human therapy, but rather serve as a stepping stone to seeking help from adults. However, some students view these tools as temporary fixes.

Phillips’s experience highlights the importance of human oversight. She notes that the AI tool is an improvement over previous monitoring systems that led to in-school discipline rather than mental health support. This year, she identified 19 “severe” alerts from the AI tool (out of 393 active users), some of which were revealed to be students testing the system’s boundaries.

Frequently Asked Questions

Is AI therapy effective?
AI can be a useful tool for initial screening and providing resources, but it should not replace human interaction with qualified mental health professionals.
What are the privacy concerns with AI mental health tools?
AI chatbots may not have the same privacy protections as traditional therapy, and data security is a concern.
How can schools responsibly implement AI mental health tools?
Schools should prioritize human oversight, ensure student data privacy, and use AI as a supplement to, not a replacement for, human counselors.

Pro Tip: If your school is considering implementing an AI mental health tool, involve students, parents, and counselors in the decision-making process.

Explore more articles on the future of technology in education here.

You may also like

Leave a Comment