Back to Search Start Over

Towards Fair Educational Data Mining: A Case Study on Detecting At-Risk Students

Authors :
Hu, Qian
Rangwala, Huzefa
Source :
International Educational Data Mining Society. 2020.
Publication Year :
2020

Abstract

Over the past decade, machine learning has become an integral part of educational technologies. With more and more applications such as students' performance prediction, course recommendation, dropout prediction and knowledge tracing relying upon machine learning models, there is increasing evidence and concerns about bias and unfairness of these models. Unfair models can lead to inequitable outcomes for some groups of students and negatively impact their learning. We show by real-world examples that educational data has embedded bias that leads to biased student modeling, which urges the development of fairness formalizations and fair algorithms for educational applications. Several formalizations of fairness have been proposed that can be classified into two types: (i) group fairness and (ii) individual fairness. Group fairness guarantees that groups are treated fairly as a whole, which might not be fair to some individuals. Thus individual fairness has been proposed to make sure fairness is achieved on individual level. In this work, we focus on developing an individually fair model for identifying students at-risk of underperforming. We propose a model which is based on the idea that the prediction for a student (identifying at-risk students) should not be influenced by his/her sensitive attributes. The proposed model is shown to effectively remove bias from these predictions and hence, making them useful in aiding all students. [For the full proceedings, see ED607784.]

Details

Language :
English
Database :
ERIC
Journal :
International Educational Data Mining Society
Publication Type :
Conference
Accession number :
ED608050
Document Type :
Speeches/Meeting Papers<br />Reports - Research