Skip to content

Commit

Permalink
Merge pull request #3104 from OpenSecuritySummit/Alone2671-patch-1251
Browse files Browse the repository at this point in the history
Create Threat-Modeling-the-LLM-Pipeline.md
  • Loading branch information
Alone2671 authored Aug 18, 2024
2 parents 2a88676 + b28130f commit 74a1c39
Showing 1 changed file with 66 additions and 0 deletions.
Original file line number Diff line number Diff line change
@@ -0,0 +1,66 @@
---
title : Threat Modeling the LLM Pipeline
track : Threat Modeling
project : Threat Modeling
type : working-session
topics :
featured :
event : mini-summit
when_year : 2024
when_month : Sep
when_day : Mon
when_time : WS-16-17
hey_summit :
session_slack:
#status : draft
description :
banner :
organizers :
- Brandon Green

youtube_link :
zoom_link :
---

## About this session
Rapid advancements in AI technologies have opened up a plethora of opportunities, but they’ve also introduced significant risks and challenges. One crucial yet often overlooked aspect of AI development is threat modeling. Threat modeling AI systems gives companies an edge in secure by design principles. This session aims to illuminate these potential threats within the AI pipeline and provides strategies for identifying, analyzing, and mitigating them.

### Presentation Outline: Threat Modeling the AI Pipeline

1. Introduction: What is Threat Modeling?

2. Data Collection:
What can go wrong? (Bias, poisoning)
Mitigations (Diversity in data sources, constant quality checks, skepticism about data origins)
Real-world examples (Amazon hiring algorithm, healthcare biases)

3. Data Preparation:
What can go wrong? (Privacy violations, feature engineering errors)
Mitigations (Data validation, encryption, regular audits, ongoing monitoring)

4. Learning Algorithm:
What can go wrong? (Overfitting, inadequate training)
Mitigations (Choose fair algorithms, diverse perspectives in development, ongoing monitoring)

5. Performance Evaluation:
What can go wrong? (Bias, data drift, etc.)
Mitigations (Cross-validation, data augmentation/diversification, fairness metrics, ongoing monitoring)

6. Deployment Considerations:
What can go wrong? (Cybersecurity threats)
Mitigations (Standard cybersecurity best practices)

7. User Input:
What can go wrong? (Prompt injections, misalignment)
Mitigations (Artful abstraction, contextual input handling, train for robustness, ongoing monitoring)
Real-world example (Bing Chat "Sydney")

8. Model Output:
What can go wrong? (Bias, misinformation)
Mitigations (Fact-checking, authenticity mechanisms, limit access, ongoing monitoring)

9. Fully Trained Model:
What can go wrong? (Adversarial attacks, model drift)
Mitigations (Adversarial training, encourage good behavior, ongoing monitoring)

10. Conclusion: Building Responsible AI

0 comments on commit 74a1c39

Please sign in to comment.