18

Open Problems in the Security of Learning

Embed Size (px)

DESCRIPTION

Open Problems in the Security of Learning. Marco Barreno , Peter L. Bartlett, Fuching Jack Chi, Anthony D. Joseph, Blaine Nelson, Benjamin I. P. Rubinstein, Udam Saini , & J.D. Tygar UC Berkeley. Can we learn securely?. Learning: key to security (intrusion/spam detection, etc) - PowerPoint PPT Presentation

Citation preview

Page 1: Open Problems in the Security of Learning
Page 2: Open Problems in the Security of Learning

Can we learn securely?Learning: key to security (intrusion/spam

detection, etc)

How to learn in a malicious environment? Bad training data

Research Questions: What attacks are feasible? Measurement Defenses

Page 3: Open Problems in the Security of Learning

Secure Learning as a Game

Adversarial capability

Adversarial influence

Secure learning

Example: Spam Detection Adversary knows

lexicon Controls fraction of e-

mail Degrades performance (Poisons classifier)

Information

Actions

Actions Actions Actions

Page 4: Open Problems in the Security of Learning

Adversarial CapabilityThreat model:

Adversary objective Learning algorithm Typical training data What adversary

controls

Research questions: Formal threat

models Measure threat

impact

Information

Actions

Add words to spam

Poison filter

Probe filter

Find weaknesses

Typical training data

Learning algorithm

Strategies

Page 5: Open Problems in the Security of Learning

Attacker

Email Distributio

n

Filter

Contamination

Attacker’s Information

INBOX

Poisoning the Training Set

Learner

Spam

Ham

AttackCorpus

Spam Folder

Page 6: Open Problems in the Security of Learning

Stealthy AttacksGoal: Undetectable, covert attacks

Covert attacks are more effective Untraceable

Research Questions: Measure visibility of attack Provably undetectable attacks Defenses

Page 7: Open Problems in the Security of Learning

Effort-Impact Trade-offsHow much adversary

effort? Depends on attack

type Data contamination Probing learner

Research Questions: Bound on adversarial

effort Bound # of learner

mistakes

Actions

Actions Actions

Best forAdversar

y

Best forLearner

no-op poison

Naive spam filter

Page 8: Open Problems in the Security of Learning

Dictionary AttackMake spam filter unusable

misclassify ham as spam

Spammer

Page 9: Open Problems in the Security of Learning

Dictionary AttackInitial Inbox: 10K

messages

Attacks Black: Optimal

Red: English dictionary

Blue: 90K most common words in Usenet

Page 10: Open Problems in the Security of Learning

ProbingGoal: Adversary wants to reverse engineer

classifier Adversary probes learner state through queries Adversary finds classifier vulnerabilities

Research Questions: Complexity of classifier reverse engineering Develop classifiers resistant to reverse

engineering

Page 11: Open Problems in the Security of Learning

DefensesGoal: secure learning

Malicious environments Range of threats

Research Questions: Build secure learning Robust against

contamination Measure robustness

Actions

Actions Actions

no-op poison

Naive spam filterRobust spam filter

Page 12: Open Problems in the Security of Learning

DefensesReject on Negative Impact (RONI)

Method Assess impact of query message on training Exclude messages with large negative impact

Preliminary Results Perfectly identifies dictionary attacks Unable to differentiate focused attacks

SpamBayes Filter

SpamBayes Learner

Page 13: Open Problems in the Security of Learning

Designing Security Sensitive Learners

Generally difficult to design defenses for a particular adversary

Robust Learning – learners are resilient to limited amount of arbitrary contamination. Robustness measures provide comparison between

procedures

QUESTION: What are useful measures of a procedure’s robustness for designing security sensitive learners? Which learners perform well w.r.t. these measures?

Page 14: Open Problems in the Security of Learning

Orthogonal Learners Idea: use parallel independent learners

Orthogonal learners have different strengths & vulnerabilities

Attack must succeed against entire group

Research Questions: Design orthogonal learners Automatic generation of orthogonal learners

Page 15: Open Problems in the Security of Learning

Future of Secure MLTraditional machine learning: no malicious

environment

Contrast with Secure ML Open research area Vital for use of machine learning in security

Strong preliminary results Showed attack feasibility on Bayesian spam detection Showed attack feasibility on principal component

analysis Evidence of activity in the wild

Prediction: area is ready to move

Page 16: Open Problems in the Security of Learning
Page 17: Open Problems in the Security of Learning
Page 18: Open Problems in the Security of Learning

Characterizing Reasonable Capabilities

What are common adversarial actions? Actions limited by capabilities & information

adversary has

QUESTION: Which forms of adversarial control are tolerable? What sorts of information? Controls?

QUESTION: What are the trade-offs between the learner’s generalization capacity & the adversary’s capabilities?