310 likes | 456 Views
Quantitative Methods for Forensic Footwear Analysis. Presented by: Martin Herman Other Team Members: Gunay Dogan, Eve Fleisig, Janelle Henrich, Sarah Hood, Hari I yer, Yooyoung Lee, Steven Lund, Gautham Venkatasubramanian Information Technology Laboratory, NIST January 25, 2018.
E N D
Quantitative Methods for Forensic Footwear Analysis Presented by: Martin Herman Other Team Members: Gunay Dogan, Eve Fleisig, Janelle Henrich, Sarah Hood, Hari Iyer, Yooyoung Lee, Steven Lund, Gautham Venkatasubramanian Information Technology Laboratory, NIST January 25, 2018
Funding Support • National Institute of Justice • IAA# DJO-NIJ-17-RO-0202 • NIST • Internal funds
MOTIVATION • 2009 NAS; 2016 PCAST: • Footwear identifications are largely subjective • Questions about reliability • Questions about scientific validity • Need for quantitative assessments of footwear evidence • Need for increased objectivity of footwear analysis • Need to improve measurement science underpinnings of forensic footwear analysis through quantitative analysis • Need for algorithmic approaches for quantitative analysis by the forensic footwear community
GOALS • Develop quantitative, objective methods for footwear impression comparisons • High degree of repeatability & reproducibility • Easier to measure accuracy with objective methods • Scientifically defensible • Provide software tools for practitioners to use in casework
Quantitative Footwear Impression Comparisons:Approach • For use by examiners in evidence evaluation • FRStat for fingerprints (DFSC) – currently in use • Proof of concept demonstration Current Examiner Comparison Process COMPARISON Crime Scene Impressions Test Impressions Conclusion plus Report Suspect Shoe
Quantitative Comparisons: End-to-End Proof of Concept Crime Scene Impression Image Feature Extraction Feature-based Matching Generate Comparison Score Score Test Impression Casework comparison score Case-relevant ground-truth-known mated and non-mated image pairs (crime scene, test impression) Generate Score Distribution & ROC charts Plot Casework Score on charts Charts, summaries, conclusions, and/or error rates
Proposed Examiner Comparison Process COMPARISON – Examiner Considers Additional Information: Score Distribution/ROC Charts and Error Rates Crime Scene Impressions Test Impressions Conclusion plus Report Suspect Shoe
Crime Scene Impression Image Feature Extraction Feature-based Matching Generate Comparison Score Score Test Impression Casework comparison score Generate Score Distribution & ROC charts Plot Casework Score on charts Case-relevant ground-truth-known mated and non-mated image pairs (crime scene, test impression) Charts, summaries, conclusions, and/or error rates
Image Feature Extraction • Existing algorithms and literature for automated shoeprint matching limited to database retrieval • Performance not adequate for evidence evaluation • Difficulties in automatically identifying outsole features (design, wear, size and RACs) in crime scene images • Partial data, occlusions, smearing, noise, low contrast, cluttered background, multiple impressions, etc. • Our Approach: Hybrid human/computer feature extraction
GUI for Image Mark-UpaaSame SourcekStaged Crime Scene Augmented Test Impression
Crime Scene Impression Image Feature Extraction Feature-based Matching Generate Comparison Score Score Test Impression Casework comparison score Generate Score Distribution & ROC charts Plot Casework Score on charts Case-relevant ground-truth-known mated and non-mated image pairs (crime scene, test impression) Charts, summaries, conclusions, and/or error rates
Feature-based Matching:Three Preliminary Algorithms • MCM_Dist • Maximum Clique Matching based on feature point distance differences between two impressions • DT_MCM_Dist • Delaunay Triangulation Maximum Clique Matching based on feature point distance differences • Comparison score: (Number of maximum cliques) / (total number of cliques) • CT (Comparable Triangles) • Number of comparable triangles based on triangular area and angle differences
Crime Scene Impression Image Feature Extraction Feature-based Matching Generate Comparison Score Score Test Impression Casework comparison score Generate Score Distribution & ROC charts Plot Casework Score on charts Case-relevant ground-truth-known mated and non-mated image pairs (crime scene, test impression) Charts, summaries, conclusions, and/or error rates
Score Distributions (DT_MCM_Dist case) Maximum separation is the goal Kernel Density Estimation
Crime Scene Impression Image Feature Extraction Feature-based Matching Generate Comparison Score Score Test Impression Casework comparison score Generate Score Distribution & ROC charts Plot Casework Score on charts Case-relevant ground-truth-known mated and non-mated image pairs (crime scene, test impression) Charts, summaries, conclusions, and/or error rates
Casework ExampleSimulated Crime Scene Test Impression Comparison score = 0.303 with DT_MCM_DIST algorithm Original images courtesy of Ron Mueller
Kernel Density Estimation At the casework score of 0.303:
Some Possible Ways to Summarize Results • Score distribution and ROC charts are created using a case-relevant reference data set – known pairs of mated and non-mated impressions that are representative of impression pairs obtained under conditions similar to the current crime scene. • The casework score is greater than 75% of mated pair scores, while 2.2% of non-mated pairs have scores higher than the casework score. • If the casework pair were considered a match, then • All pairs with higher scores would also be matches. • Therefore at least 2.2% of the non-mates would be mislabeled. • All non-mates with higher scores would be false matches, giving a FPR ≥ 2.2%. • If the casework pair were considered a non-match, then • All pairs with lower scores would also be non-matches. • This means that at least 75% of the mates would be mislabeled. • All mates with lower scores would be false non-matches, giving a FNR ≥ 75%. • Apply Kernel Density Estimation (or similar method) to histograms and obtain Likelihood Ratio as ratio of heights at casework score. LR = 13.23 Different modeling methods will result in different LRs.
Some Possible Ways to Summarize Results • Score distribution and ROC charts are created using a case-relevant reference data set – known pairs of mated and non-mated impressions that are representative of impression pairs obtained under conditions similar to the current crime scene. • The casework score is greater than 75% of mated pair scores, while 2.2% of non-mated pairs have scores higher than the casework score. • If the casework pair were considered a match, then • All pairs with higher scores would also be matches. • Therefore at least 2.2% of the non-mates would be mislabeled. • All non-mates with higher scores would be false matches, giving a FPR ≥ 2.2%. • If the casework pair were considered a non-match, then • All pairs with lower scores would also be non-matches. • Therefore at least 75% of the mates would be mislabeled. • All mates with lower scores would be false non-matches, giving a FNR ≥ 75%. • Apply Kernel Density Estimation (or similar method) to histograms and obtain Likelihood Ratio as ratio of heights at casework score. LR = 13.23 Different modeling methods will result in different LRs.
Some Possible Ways to Summarize Results • Score distribution and ROC charts are created using a case-relevant reference data set – known pairs of mated and non-mated impressions that are representative of impression pairs obtained under conditions similar to the current crime scene. • The casework score is greater than 75% of mated pair scores, while 2.2% of non-mated pairs have scores higher than the casework score. • If the casework pair were considered a match, then • All pairs with higher scores would also be matches. • Therefore at least 2.2% of the non-mates would be mislabeled. • All non-mates with higher scores would be false matches, giving a FPR ≥ 2.2%. • If the casework pair were considered a non-match, then • All pairs with lower scores would also be non-matches. • Therefore at least 75% of the mates would be mislabeled. • All mates with lower scores would be false non-matches, giving a FNR ≥ 75%. • Apply Kernel Density Estimation (or similar method) to histograms and obtain Score-based Likelihood Ratio as ratio of heights at casework score. SLR = 13.23 Different modeling methods will result in different SLRs.
Understanding the Results The comparison score obtained for the casework pair of impressions, along with score distribution and ROC charts, plus a careful description of the case-relevant reference dataset, can be used to help make weight of evidence assessments.