5.8 End of Chapter Activities
Exercises
Review Questions
- What distinguishes a backdoor attack from other types of adversarial attacks?
- Explain the steps involved in executing a backdoor poisoning attack.
- How do clean-label backdoor attacks differ from traditional backdoor attacks?
- Why are clean-label backdoor attacks harder to detect than traditional backdoor attacks?
- Why are federated learning models particularly vulnerable to backdoor attacks?
- Compare and contrast patch triggers with semantic triggers. Provide real-world examples of each.
- What are the primary differences between semantic and functional triggers?
- Suppose you are training an image classifier. How could an attacker introduce a backdoor into your model without modifying the labels?
- In a transfer learning scenario, what steps can a user take to verify that a pre-trained model is not backdoored?
- What challenges might arise when applying pruning-based defences to mitigate backdoor attacks?
- Given a real-world scenario where a backdoored model is deployed in an autonomous vehicle, what mitigation strategies would you recommend to ensure safety?
- Suppose an attacker poisons 1% of a training dataset with a backdoor trigger. Why might this attack go undetected during data sanitization?
- A facial recognition system is backdoored to misclassify people wearing red hats as a specific target. Propose a mitigation strategy and discuss its potential weaknesses.
- An autonomous vehicle uses a CNN trained on outsourced data. After deployment, it misclassifies stop signs with a small flower sticker as speed limit signs.
- What type of backdoor attack is this?
- How could the manufacturer have detected this attack before deployment?
- A hospital uses federated learning to train a model on patient data from multiple clinics. An attacker introduces a backdoor that misclassifies X-rays with a hidden watermark as “healthy.”
- What defenses could prevent this attack?
- How might the attacker evade detection?
- If a backdoored model is deployed in a critical system (e.g., medical diagnosis, autonomous driving), what are the potential consequences?
- As backdoor attacks become more sophisticated (e.g., adaptive triggers), how should defenses evolve?
- Is it possible to eliminate backdoor risks completely without sacrificing model performance?
Knowledge Check
Quiz Text Description
1. MultiChoice Activity
What is the primary goal of a backdoor poisoning attack?
- To slow down the training process of the model
- To embed a hidden trigger that causes misclassification when present
- To delete training data to make the model unusable
- To reduce the overall accuracy of the model on clean data
2. MultiChoice Activity
In which scenario does a backdoor attack occur when a user downloads a pre-trained model and fine-tunes it?
- Data Augmentation
- Outsourced Training
- Transfer Learning
- Federated Learning
3. MultiChoice Activity
Which of the following is NOT a type of backdoor trigger?
- Semantic Trigger
- Random Noise Injection
- Functional Trigger
- Patch Trigger
4. MultiChoice Activity
What is a key characteristic of a clean-label backdoor attack?
- The attacker changes both the input and its label
- The attack is performed after model deployment
- The attacker only modifies the input but keeps the correct label
- The attacker only modifies the label but not the input
5. MultiChoice Activity
Which defense technique involves identifying and removing poisoned training samples?
- Data Sanitization
- Trigger Reconstruction
- Federated Aggregation
- Model Pruning
6. MultiChoice Activity
NeuralCleanse is a technique used for
- Detecting poisoned samples in the training data
- Encrypting model weights to prevent attacks
- Pruning suspicious neurons in a neural network
- Reconstructing the backdoor trigger via optimization
7. MultiChoice Activity
Why are semantic triggers harder to detect than patch triggers?
- They are invisible to the human eye
- They blend naturally with the input (e.g., glasses on a face)
- They only work in federated learning
- They require changing the model architecture
8. MultiChoice Activity
In federated learning, how can a malicious participant introduce a backdoor?
- By slowing down the training process
- By encrypting the global model
- By submitting poisoned model updates
- By deleting other participants’ data
9. MultiChoice Activity
What is a limitation of Fine-Pruning as a defense?
- It increases model accuracy too much
- It is ineffective against dynamic or semantic triggers
- It only works for patch triggers
- It requires retraining the model from scratch
10. MultiChoice Activity
Which of the following is a post-training defense against backdoors?
- Federated Aggregation
- Trigger Reconstruction
- Data Sanitization
- Label Flipping
Correct Answers:
- b. To embed a hidden trigger that causes misclassification when present
- c. Transfer Learning
- b. Random Noise Injection
- c. The attacker only modifies the input but keeps the correct label
- a. Data Sanitization
- d. Reconstructing the backdoor trigger via optimization
- b. They blend naturally with the input (e.g., glasses on a face)
- c. By submitting poisoned model updates
- b. It is ineffective against dynamic or semantic triggers
- b. Trigger Reconstruction
High Flyer. (2025). Deep Seek. [Large language model]. https://www.deepseek.com/
Prompt: Can you provide end-of-chapter questions for the content? Reviewed and edited by the author.