140 likes | 152 Views
This study investigates the texture bias of Convolutional Neural Networks (CNNs) in object recognition and presents a method to overcome it using shape-based representation. Experiments are conducted using different data sets and stylization techniques, showing improved robustness and accuracy.
E N D
Jiahe Li 2019.1.24 ICLR, 2019
Outlines • Introduction • Experiments • Conclusions
Introduction: Texture versus Shape Classification of a standard ResNet-50
Data Sets • Original (160 natural colour images of objects (10 per category) with white background) • Greyscale • Silhouette • Edges • Texture
Data Sets • Cue conflict • Images generated using iterative style transfer (Gatys et al., 2016) between an image of the Texture data set (as style) and an image from the Original data set (as content). We generated a total of 1280 cue conflict images (80 per category), which allows for presentation to human observers within a single experimental session.
Stylized-ImageNet (SIN) • AdaIN style transfer (Huang & Belongie, 2017) • Different stylization techniques • Take prohibitively long with an iterative approach
Robert Geirhos et al. Generalisation in humans and deep neural networks, 2018 NeurIPS
Conclusions • Texture bias of CNNs trained on IN • A step towards more plausible models of human visual object recognition • Emergent benefits