
As AI adoption accelerates, so do the threats targeting those systems.
From adversarial attacks and model theft to data poisoning and API abuse, the risks are real and evolving rapidly.
In this webinar, we will explore how attackers exploit vulnerabilities in AI though adversarial examples, model extraction, and data poisoning, and more importantly, and how Crystal Eye TDIR platform can defend against them.
We will break down the technical foundations and share real-world mitigation techniques across the AI lifecycle - from training and deployment to production monitoring.
What You'll Learn:
Model - Level Security
- What makes ML models vulnerable to adversarial examples.
- Techniques like defensives distillation, input sanitisation, and data augmentation to increase model resilence.
Infrastructure & Deployment Defences
- Securing APIs and model endpoints with authentication, authorisation, and rate limiting.
- Encrypting models and data at rest and in transit.
- Using runtime monitoring and anomaly detection to flag suspicious behaviour in real time.
Advanced Protection and Confidentiality
- Preventing model theft with watermarking and model fingerprinting.
- Enhancing privacy with differential privacy and secure data handling.
- Adopting a layered security strategy across data, model and infrastructure components.
Register Now - Secure Your Spot!
Don't miss this opportunity to strengthen your AI systems against emerging threats.
Event Details
Where: Online via Microsoft Teams
When: Tuesday, 27th of May 2025
Time: 11:30 AM AWST / 1:30 PM AEST
Key Presenters
Ben Aylett - Product Manager at Red Piranha
Krishan Kumar - System Administrator at Red Piranha