Skip to main content

Red Teaming GenAI - Let's learn to break or fix Jarvis.

120-minute Workshop

Learn how to build a program that helps deliver GenAI projects through exploratory prompting, and eval frameworks.

Timetable

2:45 p.m. – 4:45 p.m. Thursday 27th

Room

Room E2+E3 - Track 5: Workshops

Other

Audience

Anyone working on GenAI projects

Required

laptop

Key-Learnings

  • Strategy and techniques for building evaluation criteria and promptss
  • Tactical explainable AI approaches for Agentic AI products

Scoring Evaluations of emergent behavior and bias

GenAI is introducing new challenges in many traditional roles in software engineering, but the role of the software tester has never had a brighter future. The very same skills that allowed you to investigate further and apply statistical patterns to search for bugs that have not even presented themselves yet... these skills are desperately needed. But not in the way you approached this in the past.

In this workshop, we will engage in a rival red team session, and build up skills defining LLM and guardrail eval techniques, labeling and scoring criteria... and discover who to begin performing hands on testing for your company projects even before a technical design has been defined.

We will also get more technical exploring how to help design methods to implement LLM-as-a-judge detection systems for agentic AI systems that comply with AI Governance standards.

Prepare to be challenged... prepare to compete, and come out with a fresh perspective on how you can apply your existing testing skills to a truly complex problem space that desperately need you.

Related Sessions

Thu, Nov 27 • 7:45 a.m. – 8:30 a.m.
Room F1+F2+F3 - Plenary

Active Session

Other

Virtual Pass session
Tue, Nov 25 • 2:45 p.m. – 3:30 p.m.
Room F2 - Track 2: Talks

25-minute Talk

Other

Mon, Nov 24 • 8:30 a.m. – 4:30 p.m.
F-,E- & D-Rooms

Full-Day Tutorial (6 hours)

API Other Test Automation