Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo
Artwork

Content provided by Daily Security Review. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Daily Security Review or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.
Player FM - Podcast App
Go offline with the Player FM app!

AI Jailbreaks on the Rise: How Hackers Are Extracting Training Data from LLMs

1:26:28
 
Share
 

Manage episode 498285250 series 3645080
Content provided by Daily Security Review. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Daily Security Review or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

In this episode, we examine the rapidly growing threat of AI jailbreaks — a cybersecurity challenge reshaping the landscape of large language models (LLMs) and enterprise chatbots. According to the IBM 2025 Cost of a Data Breach Report, 13% of all data breaches now involve AI systems, with the vast majority stemming from jailbreak attacks that circumvent developer-imposed guardrails.

A highlight of our discussion is Cisco’s “instructional decomposition” jailbreak technique, which shows how attackers can extract original training data — even copyrighted material — by manipulating conversational context and using incremental requests that evade security protocols. We’ll break down how this method works, why it’s so difficult to detect, and what it means for the future of enterprise AI.

Topics we cover include:

  • How Jailbreaks Work: From direct prompt injections to hidden instructions embedded in documents, images, or even ultrasonic audio signals.
  • Data Exfiltration Risks: LLMs trained on proprietary business data can leak PII, intellectual property, or sensitive corporate knowledge.
  • Real-World Cases: From Samsung’s 2023 ChatGPT data leak to the DeepSeek-R1 vulnerabilities and Cisco’s new demonstration of instructional decomposition, proving that what goes into LLMs can come out again.
  • The Human Factor: With 97% of breached organizations lacking proper AI access controls, internal misuse and poor governance remain critical risks.
  • Why Prevention is Hard: Experts warn it’s “very unlikely that LLMs will ever fully prevent jailbreaks,” meaning organizations must shift focus to access control and monitoring.
  • Mitigation Strategies: Multi-factor authentication, strict input/output filtering, network isolation, Zero Trust models, and employee training.
  • Regulatory Pressure: With GDPR, HIPAA, and the EU AI Act enforcing stricter compliance, failure to secure AI systems could mean not only data loss but also severe legal and financial repercussions.

As enterprises accelerate AI adoption, the line between innovation and vulnerability is razor-thin. Jailbreaks prove that guardrails alone are not enough. To safeguard sensitive data and prevent catastrophic breaches, organizations must adopt layered defenses, continuous monitoring, and robust governance frameworks.

#AIJailbreak #LLMSecurity #Cisco #InstructionalDecomposition #ChatbotRisks #DataExfiltration #GenerativeAI #Cybersecurity #AICompliance #IBMDataBreachReport #PromptInjection #EnterpriseAI #SamsungDataLeak #DeepSeekR1 #ZeroTrustAI #AIRegulation

  continue reading

261 episodes

Artwork
iconShare
 
Manage episode 498285250 series 3645080
Content provided by Daily Security Review. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Daily Security Review or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://podcastplayer.com/legal.

In this episode, we examine the rapidly growing threat of AI jailbreaks — a cybersecurity challenge reshaping the landscape of large language models (LLMs) and enterprise chatbots. According to the IBM 2025 Cost of a Data Breach Report, 13% of all data breaches now involve AI systems, with the vast majority stemming from jailbreak attacks that circumvent developer-imposed guardrails.

A highlight of our discussion is Cisco’s “instructional decomposition” jailbreak technique, which shows how attackers can extract original training data — even copyrighted material — by manipulating conversational context and using incremental requests that evade security protocols. We’ll break down how this method works, why it’s so difficult to detect, and what it means for the future of enterprise AI.

Topics we cover include:

  • How Jailbreaks Work: From direct prompt injections to hidden instructions embedded in documents, images, or even ultrasonic audio signals.
  • Data Exfiltration Risks: LLMs trained on proprietary business data can leak PII, intellectual property, or sensitive corporate knowledge.
  • Real-World Cases: From Samsung’s 2023 ChatGPT data leak to the DeepSeek-R1 vulnerabilities and Cisco’s new demonstration of instructional decomposition, proving that what goes into LLMs can come out again.
  • The Human Factor: With 97% of breached organizations lacking proper AI access controls, internal misuse and poor governance remain critical risks.
  • Why Prevention is Hard: Experts warn it’s “very unlikely that LLMs will ever fully prevent jailbreaks,” meaning organizations must shift focus to access control and monitoring.
  • Mitigation Strategies: Multi-factor authentication, strict input/output filtering, network isolation, Zero Trust models, and employee training.
  • Regulatory Pressure: With GDPR, HIPAA, and the EU AI Act enforcing stricter compliance, failure to secure AI systems could mean not only data loss but also severe legal and financial repercussions.

As enterprises accelerate AI adoption, the line between innovation and vulnerability is razor-thin. Jailbreaks prove that guardrails alone are not enough. To safeguard sensitive data and prevent catastrophic breaches, organizations must adopt layered defenses, continuous monitoring, and robust governance frameworks.

#AIJailbreak #LLMSecurity #Cisco #InstructionalDecomposition #ChatbotRisks #DataExfiltration #GenerativeAI #Cybersecurity #AICompliance #IBMDataBreachReport #PromptInjection #EnterpriseAI #SamsungDataLeak #DeepSeekR1 #ZeroTrustAI #AIRegulation

  continue reading

261 episodes

सभी एपिसोड

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play