• Graduate Programs
    • Tinbergen Institute Research Master in Economics
      • Why Tinbergen Institute?
      • Research Master
      • Admissions
      • All Placement Records
      • PhD Vacancies
    • Facilities
    • Research Master Business Data Science
    • Education for external participants
    • Summer School
    • Tinbergen Institute Lectures
    • PhD Vacancies
  • Research
  • Browse our Courses
  • Events
    • Summer School
      • Applied Public Policy Evaluation
      • Deep Learning
      • Development Economics
      • Economics of Blockchain and Digital Currencies
      • Economics of Climate Change
      • The Economics of Crime
      • Foundations of Machine Learning with Applications in Python
      • From Preference to Choice: The Economic Theory of Decision-Making
      • Inequalities in Health and Healthcare
      • Marketing Research with Purpose
      • Markets with Frictions
      • Modern Toolbox for Spatial and Functional Data
      • Sustainable Finance
      • Tuition Fees and Payment
      • Business Data Science Summer School Program
    • Events Calendar
    • Events Archive
    • Tinbergen Institute Lectures
    • 2026 Tinbergen Institute Opening Conference
    • Annual Tinbergen Institute Conference
  • News
  • Summer School
  • Alumni
    • PhD Theses
    • Master Theses
    • Selected PhD Placements
    • Key alumni publications
    • Alumni Community
Home | Events Archive | Training Neural Networks To Be Insensitve Towards Spurious Concepts
Seminar

Training Neural Networks To Be Insensitve Towards Spurious Concepts


  • Location
    University of Amsterdam, E5.22
    Amsterdam
  • Date and time

    October 28, 2022
    12:30 - 13:30

Abstract
Neural networks are widely used for image recognition. However, a major shortcoming is that they often rely on spurious correlations. Concept activation vectors (Kim et al. 2018) can be used to quantify if a neural network is sensitive with respect to a concept -- e.g. does it use the sea to classify a seagull? The contribution of this thesis is to introduce a method that trains a neural network to be insensitive with respect to a concept, called CAV-penalized training. Users select a set of pictures that capture the spurious concept, and the neural network is trained to not use it for classification. This allows for the incorporation of domain knowledge to deal with a range of spurious correlations. The effectiveness of CAV-penalized training is illustrated across benchmark datasets (MNIST, Waterbirds, CelebA), for both a convolutional neural network (CNN) and a fine-tuned Resnet-50 architecture. Our results indicate that CAV-penalized training performs similar or is competitive with a model trained on a dataset without the spurious correlation. Compared to other methods, CAV-penalized training requires little data annotation (100-250 images of a concept), yet achieves a competitive or better performance.