4.6 End of Chapter Activities
Exercises
Review Questions
- Explain the difference between adversarial attacks and poisoning attacks. Provide examples of each.
- Describe the three primary attack scenarios in poisoning attacks (TS, FT, MT). How do they differ in terms of attacker capabilities and risks?
- What is the goal of an indiscriminate poisoning attack, and how does it differ from a targeted poisoning attack?
- Explain the concept of “feature collision” in clean-label poisoning attacks. How does it exploit the complexity of deep neural networks?
- Discuss the role of outlier detection in mitigating poisoning attacks. Provide examples of techniques that use outlier detection.
- How does differential privacy help in defending against poisoning attacks? What are its limitations?
- Compare and contrast label-flip poisoning and bilevel poisoning. Which one is more effective, and why?
- What are the challenges in detecting and mitigating backdoor attacks during model inspection?
Discussion Questions
- How can organizations balance the need for large, diverse datasets with the risk of poisoning attacks?
- What ethical considerations arise when deploying AI models that may be vulnerable to poisoning attacks?
- How might advancements in explainable AI (XAI) help in detecting and mitigating poisoning attacks?
- What role do regulatory frameworks play in ensuring the security of AI models against poisoning attacks?
Knowledge Check
Quiz Text Description
1. MultiChoice Activity
Which of the following best describes a data poisoning attack?
- Encrypting training data for security
- Manipulating training data to alter model behavior
- Modifying test samples to deceive an ML model
- Reducing model complexity to prevent overfitting
2. MultiChoice Activity
In which scenario does an attacker fine-tune a pre-trained model to introduce vulnerabilities?
- Fine-tuning (FT)
- Model outsourcing (MT)
- Training-from-scratch (TS)
- Reinforcement learning
3. MultiChoice Activity
What is the goal of a targeted poisoning attack?
- To remove adversarial samples from the dataset
- To optimize training data for better performance
- To misclassify a specific target sample while maintaining high overall accuracy
- To decrease the overall model accuracy
4. MultiChoice Activity
Which of the following is NOT a defense against data poisoning attacks?
- Robust training
- Increasing model complexity
- Model inspection
- Training data sanitization
5. MultiChoice Activity
The feature-collision technique in targeted attacks relies on:
- Randomly altering labels of training samples
- Encrypting the training data to prevent access
- Making poisoned samples resemble target samples in feature space
- Using clean data only for training
6. MultiChoice Activity
Which method is commonly used in training data sanitization defenses?
- Gradient descent optimization
- Backpropagation
- Reinforcement learning
- Clustering and outlier detection
Correct Answers:
- b. Manipulating training data to alter model behavior
- a. Fine-tuning (FT)
- c. To misclassify a specific target sample while maintaining high overall accuracy
- b. Increasing model complexity
- c. Making poisoned samples resemble target samples in feature space
- d. Clustering and outlier detection
High Flyer. (2025). Deep Seek. [Large language model]. https://www.deepseek.com/
Prompt: Can you provide end-of-chapter questions for the content? Reviewed and edited by the author.