Program

/

CODE BLUE 2025

Time Table

[CyCraft Japan Corporation]AI Security in Practice: From Adversarial Injections to Guardrail Defenses

DAY 1

14:25-14:55

As enterprises adopt Large Language Models (LLMs), attackers are quickly adapting with new adversarial techniques such as prompt injection, prompt extraction, and jailbreaks. Traditional security approaches fall short against these emerging threats. This session focuses on two critical pillars for securing AI systems in practice:

  1. Adversarial Red-Team Testing — how systematic injection scenarios reveal hidden vulnerabilities in LLMs.
  2. Guardrail Defenses — how layered, verifiable controls can mitigate malicious prompts and ensure safer model behavior.

We will share insights from real-world testing, discuss defense strategies, and explore how organizations can build resilience against AI-driven attacks without over-reliance on “black box” trust.

Speakers:
Renata Chang(レナタ チョウ) CyCraft Japan Corporation Architect Consultant

  • Location :

    • Track 2(HALL A)

  • Category :

    • OpenTalks

  • Share :