Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by Vasanth Sarathy & Laura Hagopian, Vasanth Sarathy, and Laura Hagopian. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Vasanth Sarathy & Laura Hagopian, Vasanth Sarathy, and Laura Hagopian or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.
Player FM - Podcast App
Go offline with the Player FM app!

#22 - Hope, Help, and the Language We Choose

24:58
 
Share
 

Manage episode 523753076 series 3678189
Content provided by Vasanth Sarathy & Laura Hagopian, Vasanth Sarathy, and Laura Hagopian. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Vasanth Sarathy & Laura Hagopian, Vasanth Sarathy, and Laura Hagopian or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

What if the words we use could tip the balance between seeking help and staying silent? In this episode, we explore a fascinating study that compares top-voted Reddit responses with replies generated by large language models (LLMs) to uncover which better reduces stigma around opioid use disorder—and why that distinction matters.

Drawing from Laura’s on-the-ground ER experience and Vasanth’s research on language and moderation, we examine how subtle shifts, like saying “addict” versus “person with OUD, ” can reshape beliefs, impact treatment, and even inform policy. The study zeroes in on three kinds of stigma: skepticism toward medications like Suboxone and methadone, biases against people with OUD, and doubts about the possibility of recovery.

Surprisingly, even with minimal prompting, LLM responses often came across as more supportive, hopeful, and factually accurate. We walk through real examples where personal anecdotes, though well-intended, unintentionally reinforced harmful myths—while AI replies used precise, compassionate language to challenge stigma and foster trust.

But this isn’t a story about AI hype. It’s about how moderation works in online communities, why tone and pronouns matter, and how transparency is key. The takeaway? Language is infrastructure. With thoughtful design and human oversight, AI can help create safer digital spaces, lower barriers to care, and make it easier for people to ask for help, without fear.

If this conversation sparks something for you, follow the show, share it with someone who cares about public health or ethical tech, and leave us a review. Your voice shapes this space: what kind of language do you want to see more of?

Reference:

Exposure to content written by large language models can reduce stigma around opioid use disorder
Shravika Mittal et al.
npj Artificial Intelligence (2025)

Credits:

Theme music: Nowhere Land, Kevin MacLeod (incompetech.com)
Licensed under Creative Commons: By Attribution 4.0
https://creativecommons.org/licenses/by/4.0/

  continue reading

Chapters

1. Opening And Personal Stakes (00:00:00)

2. Stigma And Online Communities (00:01:01)

3. Can AI Moderate For Care? (00:02:31)

4. Three Dimensions Of Stigma (00:04:46)

5. Study Design And Prompt Strategy (00:06:36)

6. Suboxone Example: Human vs LLM (00:09:08)

7. Pronouns, Tone, And Trust (00:11:10)

8. Five-Year Question And Bot Limits (00:13:06)

9. When Human Advice Backfires (00:14:20)

10. Why Language Shapes Policy And Care (00:16:10)

11. Takeaways And Closing (00:18:12)

22 episodes

Artwork
iconShare
 
Manage episode 523753076 series 3678189
Content provided by Vasanth Sarathy & Laura Hagopian, Vasanth Sarathy, and Laura Hagopian. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Vasanth Sarathy & Laura Hagopian, Vasanth Sarathy, and Laura Hagopian or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

What if the words we use could tip the balance between seeking help and staying silent? In this episode, we explore a fascinating study that compares top-voted Reddit responses with replies generated by large language models (LLMs) to uncover which better reduces stigma around opioid use disorder—and why that distinction matters.

Drawing from Laura’s on-the-ground ER experience and Vasanth’s research on language and moderation, we examine how subtle shifts, like saying “addict” versus “person with OUD, ” can reshape beliefs, impact treatment, and even inform policy. The study zeroes in on three kinds of stigma: skepticism toward medications like Suboxone and methadone, biases against people with OUD, and doubts about the possibility of recovery.

Surprisingly, even with minimal prompting, LLM responses often came across as more supportive, hopeful, and factually accurate. We walk through real examples where personal anecdotes, though well-intended, unintentionally reinforced harmful myths—while AI replies used precise, compassionate language to challenge stigma and foster trust.

But this isn’t a story about AI hype. It’s about how moderation works in online communities, why tone and pronouns matter, and how transparency is key. The takeaway? Language is infrastructure. With thoughtful design and human oversight, AI can help create safer digital spaces, lower barriers to care, and make it easier for people to ask for help, without fear.

If this conversation sparks something for you, follow the show, share it with someone who cares about public health or ethical tech, and leave us a review. Your voice shapes this space: what kind of language do you want to see more of?

Reference:

Exposure to content written by large language models can reduce stigma around opioid use disorder
Shravika Mittal et al.
npj Artificial Intelligence (2025)

Credits:

Theme music: Nowhere Land, Kevin MacLeod (incompetech.com)
Licensed under Creative Commons: By Attribution 4.0
https://creativecommons.org/licenses/by/4.0/

  continue reading

Chapters

1. Opening And Personal Stakes (00:00:00)

2. Stigma And Online Communities (00:01:01)

3. Can AI Moderate For Care? (00:02:31)

4. Three Dimensions Of Stigma (00:04:46)

5. Study Design And Prompt Strategy (00:06:36)

6. Suboxone Example: Human vs LLM (00:09:08)

7. Pronouns, Tone, And Trust (00:11:10)

8. Five-Year Question And Bot Limits (00:13:06)

9. When Human Advice Backfires (00:14:20)

10. Why Language Shapes Policy And Care (00:16:10)

11. Takeaways And Closing (00:18:12)

22 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play