Key Highlights
- Wysa launches the Safety Assessment for LLMs in Mental Health (SAFE-LMH) on World Mental Health Day.
- The platform will evaluate Large Language Models (LLMs) for safe and empathetic mental health conversations across multiple languages.
- SAFE-LMH aims to set new standards for AI-driven mental health care, addressing gaps in non-English languages.
- Wysa will open-source a dataset featuring 500-800 mental health-related test cases in 20 languages.
- Invites global research partners to collaborate on ensuring safer AI in mental health.
Source: Business Wire
Notable Quote
- “Our goal with the Safety Assessment for LLMs in Mental Health is clear: to ensure that the world’s rapidly advancing AI tools can deliver safe, empathetic, and culturally relevant mental health support, no matter the language.” — Jo Aggarwal, CEO at Wysa
SoHC's Take
Wysa’s launch of the Safety Assessment for LLMs in Mental Health (SAFE-LMH) marks a significant leap forward in the integration of AI with mental health care. By focusing on multilingual models and non-English languages, SAFE-LMH addresses a critical gap in AI-driven mental health tools—ensuring safety, empathy, and cultural sensitivity in a highly personal domain. This initiative invites a global community to collaborate in setting new standards, enhancing the scalability and inclusivity of AI in mental health support systems. Wysa’s commitment to transparency through open-sourcing test cases sets a powerful precedent for ethical AI development in mental health.