Knowledge Sprint - Securing AI 27th May 2025

As AI adoption accelerates, so do the threats targeting those systems.

From adversarial attacks and model theft to data poisoning and API abuse, the risks are real and evolving rapidly.

In this webinar, we will explore how attackers exploit vulnerabilities in AI though adversarial examples, model extraction, and data poisoning, and more importantly, and how Crystal Eye TDIR platform can defend against them.

We will break down the technical foundations and share real-world mitigation techniques across the AI lifecycle - from training and deployment to production monitoring.

What You'll Learn:

Model - Level Security

  • What makes ML models vulnerable to adversarial examples.
  • Techniques like defensives distillation, input sanitisation, and data augmentation to increase model resilence.

Infrastructure & Deployment Defences

  • Securing APIs and model endpoints with authentication, authorisation, and rate limiting.
  • Encrypting models and data at rest and in transit.
  • Using runtime monitoring and anomaly detection to flag suspicious behaviour in real time.

Advanced Protection and Confidentiality

  • Preventing model theft with watermarking and model fingerprinting.
  • Enhancing privacy with differential privacy and secure data handling.
  • Adopting a layered security strategy across data, model and infrastructure components.

Register Now - Secure Your Spot!

Don't miss this opportunity to strengthen your AI systems against emerging threats.


Event Details


Where: Online via Microsoft Teams

When: Tuesday, 27th of May 2025 

Time: 11:30 AM AWST / 1:30 PM AEST
             


Key Presenters


Ben Aylett - Product Manager at Red Piranha

Krishan Kumar - System Administrator at Red Piranha

Category