60 likes | 77 Views
This project aims to enhance Hadoop monitoring by utilizing tracing techniques to capture causality, machine learning to detect problems and unsupervised learning to identify unusual runs. Additionally, it plans to analyze X-Trace reports in real-time and scale data collection for better Hadoop performance.
E N D
Monitoring Hadoop through Tracing Andy Konwinski and Matei Zaharia
Objectives • Debug and profile data center applications • Hadoop file system and map-reduce • Apache Nutch web indexing engine • Automatically detect problems from traces
State-of-the-Art • Unpublished proprietary log management systems at Google, Yahoo, etc • Per-machine logs • Sawzall for mining log data • Node monitoring daemon (System Health Infrastructure)
Our Idea • Capture causality directly by tracing computations across nodes using X-Trace • Use machine learning to detect problems • Detect unusual runs using unsupervised learning • Classify problems using supervised learning • Also want to study Hadoop performance
Risks • Scaling X-Trace data collection • Analyzing X-Trace reports in real time • Identifying features of X-Trace graphs to run machine learning on • Our manually induced errors may not capture all failures that happen in a production cluster