170 likes | 360 Views
Casting out Demons: Sanitizing Training Data for Anomaly Sensors in IEEE Symp . on S&P 2008 G. F. Cretu , A. Stavrou , M. E. Locasto , S. J. Stolfo and A. D. Keromytis. Po-Ching Lin Dept. of CSIE National Chung Cheng University. Problem definition.
E N D
Casting out Demons:Sanitizing Training Data for Anomaly Sensorsin IEEE Symp. on S&P 2008G. F. Cretu, A. Stavrou, M. E. Locasto, S. J. Stolfo and A. D. Keromytis Po-Ching Lin Dept. of CSIE National Chung Cheng University
Problem definition • Two main approaches to detecting malicious inputs, behavior, network traffic, etc. • Signature matching • Anomaly detection • Challenge of effective anomaly detection of malicious traffic • A highly accurate modeling of normal traffic • Real network traffic is usually polluted or unclean • Using it as the training data can be a problem • How can we sanitize training data for AD sensors? 2
Solution outline • Assumption: • An attack or abnormality appear only in small subsetsin a large training set • The solution: • Test each packet with the micro models using the voting scheme and build a “normal” model. • Data deemed abnormal is used for building an abnormal model. • Abnormal model can be distributed between sites. • A shadow sensor architecture to handle false positives. MM1 MM5 MM4 MM2 MM3 training set noise: attack or non-regularity Mi: micro model i 3
Assumption & micro models • Observation • Over a long period, attacks and abnormalities are a minority class of data. • Deriving the micro models • T = {md1,md2,...,mdN}, where mdi is the micro-dataset starting at time (i − 1) ∗ g, g from 3 to 5 hours • Mi = AD(mdi) : the micro model from mdi 4
Deriving the sanitized training set • Sanitize the training dataset • Test each packet Pj with all the micro-models Mi. • Lj,i = TEST(Pj ,Mi), Lj,i=1 if Pj is abnormal; otherwise 0. • Combine output from the models • SCORE(Pj )= 1/WwiLj,i, where W=wi. • Sanitize the training dataset • Tsan={Pj|SCORE(Pj)V}, Msan = AD(Tsan). • Tabn={Pj|SCORE(Pj)>V}, Msan = AD(Tsan). 5
Evaluation of sanitization • Use two anomaly sensors for evaluation • Anagram and Payl • Experimental corpus • 500 hours of real network traffic • 300 hours of traffic to build the micro-models • the next 100 hours to generate the sanitized model • the remaining 100 hours of data was used for testing • from three different hosts: www, www1, and lists • with cross validation 6
Without sanitizing the training data A: Anagram A-S: Anagram +Snort A-SAN: Anagram + sanitization P: Payl P-SAN: Payl+sanitization V ∈ [0.15, 0.45] 7
Analysis of sanitization parameters • Three parameters for fine-tuning • The granularity of micro-models • The voting algorithm (simple voting vs. weighted voting) • The voting threshold 8
Collaborative Sanitization • Comparing models of abnormality with those generated by other sites • Direct model differencing • Mcross = Msan − {Mabni ∩Msan} • Indirect model differencing • Differencing the sets of packets used to compute the models. • If a packet Pj is considered abnormal by at least one Mabni • features are extracted from the packet for computing the new local abnormal model • used for computing the cross-sanitized model 15
Conclusion & limitations in this paper • The capability of anomaly detection in the micro models • Effectiveness of PAYL and Anagram? • The traffic in the evaluation • “normality” is diverse in a real environment • Deriving packets to form the training set is stateless • Attacks can be across packets or even connections 17