Back to Search
Start Over
Towards causally linking architectural parametrizations to algorithmic bias in neural networks
- Publication Year :
- 2023
- Publisher :
- arXiv, 2023.
-
Abstract
- Training dataset biases are by far the most scrutinized factors when explaining algorithmic biases of neural networks. In contrast, hyperparameters related to the neural network architecture, e.g., the number of layers or choice of activation functions, have largely been ignored even though different network parameterizations are known to induce different implicit biases over learned features. For example, convolutional kernel size has been shown to bias CNNs towards different frequencies. In order to study the effect of these hyperparameters, we designed a causal framework for linking an architectural hyperparameter to algorithmic bias. Our framework is experimental, in that several versions of a network are trained with an intervention to a specific hyperparameter, and the resulting causal effect of this choice on performance bias is measured. We focused on the causal relationship between sensitivity to high-frequency image details and face analysis classification performance across different subpopulations (race/gender). In this work, we show that modifying a CNN hyperparameter (convolutional kernel size), even in one layer of a CNN, will not only change a fundamental characteristic of the learned features (frequency content) but that this change can vary significantly across data subgroups (race/gender populations) leading to biased generalization performance even in the presence of a balanced dataset.<br />Comment: 23 pages, 15 figures, 2 tables
Details
- Database :
- OpenAIRE
- Accession number :
- edsair.doi.dedup.....35d97e91de6992da433881adc690de20
- Full Text :
- https://doi.org/10.48550/arxiv.2302.03750