Zur Hauptnavigation wechseln Zur Suche wechseln Zum Hauptinhalt wechseln

Towards Explainable Ear Recognition Systems Using Deep Residual Networks

Hammam Alshazly*, Christoph Linse, Erhardt Barth, Sahar Ahmed Idris, Thomas Martinetz

*Korrespondierende/r Autor/-in für diese Arbeit

Abstract

This paper presents ear recognition models constructed with Deep Residual Networks (ResNet) of various depths. Due to relatively limited amounts of ear images we propose three different transfer learning strategies to address the ear recognition problem. This is done either through utilizing the ResNet architectures as feature extractors or through employing end-to-end system designs. First, we use pretrained models trained on specific visual recognition tasks, inititalize the network weights and train the fully-connected layer on the ear recognition task. Second, we fine-tune entire pretrained models on the training part of each ear dataset. Third, we utilize the output of the penultimate layer of the fine-tuned ResNet models as feature extractors to feed SVM classifiers. Finally, we build ensembles of networks with various depths to enhance the overall system performance. Extensive experiments are conducted to evaluate the obtained models using ear images acquired under constrained and unconstrained imaging conditions from the AMI, AMIC, WPUT and AWE ear databases. The best performance is obtained by averaging ensembles of fine-tuned networks achieving recognition accuracy of 99.64%, 98.57%, 81.89%, and 67.25% on the AMI, AMIC, WPUT, and AWE databases, respectively. In order to facilitate the interpretation of the obtained results and explain the performance differences on each ear dataset we apply the powerful Guided Grad-CAM technique, which provides visual explanations to unravel the black-box nature of deep models. The provided visualizations highlight the most relevant and discriminative ear regions exploited by the models to differentiate between individuals. Based on our analysis of the localization maps and visualizations we argue that our models make correct prediction when considering the geometrical structure of the ear shape as a discriminative region even with a mild degree of head rotations and the presence of hair occlusion and accessories. However, severe head movements and low contrast images have a negative impact of the recognition performance.

OriginalspracheEnglisch
Aufsatznummer9526589
ZeitschriftIEEE Access
Jahrgang9
Seiten (von - bis)122254-122273
Seitenumfang20
DOIs
PublikationsstatusVeröffentlicht - 2021

Fördermittel

This work was supported by the Deanship of Scientific Research at King Khalid University through the Research Groups Program under Grant RGP.1/95/42. The work of Christoph Linse was supported by the Bundesministeriums für Wirtschaft und Energie (BMWi) through the Mittelstand 4.0-Kompetenzzentrum Kiel Project.

UN SDGs

Dieser Output leistet einen Beitrag zu folgendem(n) Ziel(en) für nachhaltige Entwicklung

  1. SDG 3 – Gesundheit und Wohlergehen
    SDG 3 – Gesundheit und Wohlergehen
  2. SDG 4 – Qualitativ hochwertige Bildung
    SDG 4 – Qualitativ hochwertige Bildung
  3. SDG 9 – Industrie, Innovation und Infrastruktur
    SDG 9 – Industrie, Innovation und Infrastruktur
  4. SDG 11 – Nachhaltige Städte und Gemeinschaften
    SDG 11 – Nachhaltige Städte und Gemeinschaften
  5. SDG 12 – Verantwortungsvoller Konsum und Produktion
    SDG 12 – Verantwortungsvoller Konsum und Produktion
  6. SDG 14 – Lebensraum Wasser
    SDG 14 – Lebensraum Wasser
  7. SDG 15 – Lebensraum Land
    SDG 15 – Lebensraum Land

Strategische Forschungsbereiche und Zentren

  • Zentren: Zentrum für Künstliche Intelligenz Lübeck (ZKIL)
  • Querschnittsbereich: Intelligente Systeme

DFG-Fachsystematik

  • 4.43-05 Bild- und Sprachverarbeitung, Computergraphik und Visualisierung, Human Computer Interaction, Ubiquitous und Wearable Computing

Fingerprint

Untersuchen Sie die Forschungsthemen von „Towards Explainable Ear Recognition Systems Using Deep Residual Networks“. Zusammen bilden sie einen einzigartigen Fingerprint.

Zitieren