Session Video
About this session
Rapid advancements in AI technologies have opened up a plethora of opportunities, but they’ve also introduced significant risks and challenges. One crucial yet often overlooked aspect of AI development is threat modeling. Threat modeling AI systems gives companies an edge in secure by design principles. This session aims to illuminate these potential threats within the AI pipeline and provides strategies for identifying, analyzing, and mitigating them.
Presentation Outline: Threat Modeling the AI Pipeline
Introduction: What is Threat Modeling?
Data Collection: What can go wrong? (Bias, poisoning) Mitigations (Diversity in data sources, constant quality checks, skepticism about data origins) Real-world examples (Amazon hiring algorithm, healthcare biases)
Data Preparation: What can go wrong? (Privacy violations, feature engineering errors) Mitigations (Data validation, encryption, regular audits, ongoing monitoring)
Learning Algorithm: What can go wrong? (Overfitting, inadequate training) Mitigations (Choose fair algorithms, diverse perspectives in development, ongoing monitoring)
Performance Evaluation: What can go wrong? (Bias, data drift, etc.) Mitigations (Cross-validation, data augmentation/diversification, fairness metrics, ongoing monitoring)
Deployment Considerations: What can go wrong? (Cybersecurity threats) Mitigations (Standard cybersecurity best practices)
User Input: What can go wrong? (Prompt injections, misalignment) Mitigations (Artful abstraction, contextual input handling, train for robustness, ongoing monitoring) Real-world example (Bing Chat “Sydney”)
Model Output:
What can go wrong? (Bias, misinformation) Mitigations (Fact-checking, authenticity mechanisms, limit access, ongoing monitoring)Fully Trained Model: What can go wrong? (Adversarial attacks, model drift) Mitigations (Adversarial training, encourage good behavior, ongoing monitoring)
Conclusion: Building Responsible AI