Week 3
Week 3
1. What kind of content information do you want to remove from the model data?
a. Biased or discriminatory data
b. Useful patterns and trends
c. General public data
d. Random noise
e. Personally identifiable information
f. Valid and accurate data
3. Identify the steps involved in the exact unlearning as discussed in the course.
a. Isolate the data -> shard the data -> slice the data -> aggregate the data
b. Aggregate the data -> isolate the data -> slice the data -> shard the data
c. Shard the data -> Slice the data -> Isolate the data -> Aggregate the data
d. Shard the data -> Isolate the data -> Slice the data -> Aggregate the data
e. Isolate the data -> slice the data -> shard the data -> aggregate the data
5. How should the original model and the model after the below unlearning methods
behave?
1) exact unlearning
2) approximate unlearning
a. 1) distributionally identical 2) distributionally identical
b. 1) distributionally close 2) distributionally close
c. 1) distributionally identical 2) distributionally close
d. 1) distributionally close 2) distributionally identical
6. How does unlearning via differential privacy work?
a. check whether an adversary can reliably tell apart the models before unlearning
and after unlearning
b. check whether the model can output private and sensitive information before and
after unlearning
c. check whether the model's predictions become more consistent and stable for
private information before and after unlearning.
d. check whether an adversary can identify the differences in the distribution of
output data of the model before and after unlearning
10. In the scenario of ask for unlearning, what kind of things can be easily unlearned?
a. Hate speech
b. Toxic content
c. Factual Information
d. Sensitive information
11. When evaluating the quality of unlearning using Membership Inference Attack, which of
the following scenarios implies that the unlearning is successful?
a. The accuracy increases on the forget set
b. The accuracy drops on the forget set
c. The accuracy stays the same on the forget set
d. The accuracy increases on the test set
e. The accuracy drops on the test set
f. The accuracy stays the same on the test set
14. What idea does the paper Corrective Machine Learning build upon?
a. Not all poisoned data can be identified for unlearning
b. Identifying and removing a small subset of poisoned data points is sufficient to
ensure the model's integrity
c. enhancing the model's ability to handle completely new, unseen poisoned data
d. The accuracy of the model improves proportionally with the amount of data
removed, regardless of whether it is poisoned or not
e. adding redundant data to the dataset to counteract the effects of poisoned data.
f. Not all poisoned data can be identified for unlearning
15. Identify all the methods that act as the baseline for the TOFU benchmark dataset
a. Gradient Descent
b. Gradient Ascent
c. Gradient Difference
d. Gradient boosting
e. Gradient Clipping
16. The WMDP benchmark tests on unlearning what kind of information?
a. Biosecurity
b. High-school biology
c. Hate speech on Twitter
d. Crime data
17. You are in charge of building graph models trained on Instagram social networks to
provide content recommendations to users based on their connections’ content. You
realize that a particular user in the network is leading to toxic content recommendations.
What kind of unlearning would you use in this scenario to prevent the recommendation
of toxic content?
a. Node feature unlearning
b. Node unlearning
c. Edge Unlearning
d. Subgraph unlearning