Secure GenAI
Secure GenAI Podcast
OpenAI safety hub, Anthropic bug bounty, World 1st CPU ransom, Coinbase breach, Agentic AI risk
0:00
-4:40

OpenAI safety hub, Anthropic bug bounty, World 1st CPU ransom, Coinbase breach, Agentic AI risk

GenAI Safety & Security | April 12 - May 19, 2025

Note: my apologies for the absence in two weeks. I caught some incidents after returning to Vietnam from ICLR, 2025 Singapore. This week, I will try to upload more posts to help you get updated.

Secure GenAI is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.

0:00
-4:40

Highlights

    🚨 OpenAI: Safety hub.

    🛡️ Anthropic: Bug bounty launched.

    ⚠️ Palo Alto: Agentic AI threats.

    💰 Coinbase: $20M reward offered.

    💻 World 1st CPU-level ransomware.

Secure GenAI is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.

Dive Deep

🚨 OpenAI - Share AI Safety OpenAI

  • Bug bounty program initiated.

  • Share metrics, evaluations to measure safety.

  • Harmful content, jailbreaks, hallucination and hierarchy.

  • Safety on teaching, testing, sharing.

  • Data privacy & security commitments.

  • SOC 2 Type 2 and CSA STAR Level 1.

Suggestion: AI safety is never enough. The open access is a great way to fix this.

🛡️Anthropic - Stress-Testing Safety Anthropic

  • New bug bounty program launched.

  • Testing Constitutional Classifiers system.

  • Up to $25,000 bounty offered.

  • Focus on CBRN-related jailbreaks.

  • ASL-3 safeguards being iterated.

Headsup: Anthropic is a great example for Bug Bounty.

⚠️ Agentic AI Risks Palo Alto Networks

  • Agentic AI expands attack surface.

  • Nine identified attack scenarios.

  • Prompt injection remains potent.

  • Tool misconfigurations are critical.

  • Layered defense is essential.

    Diagram illustrating the architecture of an application with AI model integration. The diagram includes three main components: Services, Application, and AI Model. The Application is subdivided into Input, Agent, and Output areas, featuring Planning, Execution Loop, and Function Calling within the Agent. Supporting Services include Long-Term Memory and Vector Datastore. The AI Model is marked with a LLM (Large Language Model) and Function Calling. Icons for code, human in the loop, device, and content are shown under Services. Palo Alto Networks and Unit 42 logo lockup.

Headsup: Nvidia agents architecture.

💰 Coinbase - Standing Firm Against Extortion Coinbase

  • Insiders stole customer data for cash.

  • $20M ransom demand refused.

  • $20M reward fund established.

  • Impacted <1% of monthly users.

  • Name, address, phone, email, SS.

  • Bank account, gov ID, account data, corp data.

Mitigation: strong 2FA, “lock first, ask later”.

💻 CPU-Level Ransomware Tom's Hardware

  • First CPU-level ransomware PoC.

  • Bypasses traditional security tech.

  • Exploits AMD Zen chip vulnerability.

  • Zen 1 to Zen 5 AMD chip impacted.

  • Controled BIOS and load 3rd pary loader.

Headsup: Considered avoid AMD chip for encryption.

Thanks for reading Secure GenAI ! This post is public so feel free to share it.

Share

Notice: Y2 GenAI Safety and Security is on GumRoad and Amazon with paperback.

Discussion about this episode

User's avatar