300 likes | 641 Views
UGR Project - Haoyu li, brittany edwards , wei zhang under xiaoxiao xu and arye nehorai. Machine Learning Basics with Applications to Email Spam Detection. General background information about the process of machine learning. The process of email detection. Motivation of this project
E N D
UGR Project - Haoyu li, brittanyedwards, weizhang under xiaoxiaoxu and aryenehorai Machine Learning Basics with Applications to Email Spam Detection
General background information about the process of machine learning
The process of email detection • Motivation of this project • Pre-processing of data • Classifier Models • Evaluation of classifiers
Motivation of this project • Spam email has been annoyed every personal email account • 60% of January 2004 emails were spam • Fraud & Phishing • Spam vs. Ham email
The process of email detection • Motivation of this project • Pre-processing of data • Classifier Models • Evaluation of classifiers
Pre-processing of data • Convert capital letters to lowercase • Remove numbers, and extra white space • Remove punctuations • Remove stop-words • Delete terms with length greater than 20.
Pre-processing of data • Original Email
Pre-processing of data • After pre-processing
Pre-processing of data • Extract Terms
Pre-processing of data • Reduce Terms • Keep word length <20
The process of email detection • Motivation of this project • Pre-processing of data • Classifier Models • Evaluation of classifiers
Different classification methods • K Nearest Neighbor (KNN) • Naive Bayes Classifier • Logistic Regression • Decision Tree Analysis
What is K Nearest Neighbor • Use k "closet" samples (nearest neighbors) to perform classification
Initial outcome and strategies for improvement • KNN accuracy was ~64% - very low • KNN classifier does not fit our project • Term-list is still too large • Try different method to classify and see if evaluation results are better than KNN results • Continue to reduce size of term list by removing terms that are not meaningful
Steps for improvement • Remove sparsity • Reduced length threshold • Created hashtable • Used alternative classifier • Naive- Bayes Classifier
Hashtable • Calculate Hash Key for each term in term-list. • Once collision occurs, use the separate chain
Secondary Results • Correctness increases from 62% to 82.36%
Suggestions for further improvement • Revise pre-processing • Apply additional classifiers
Thank you • Questions?