When the Past != The Future: Assessing the Impact of Dataset Drift on the Fairness of Learning Analytics Models

dc.contributor.authorDeho, O.B.
dc.contributor.authorLiu, L.
dc.contributor.authorLi, J.
dc.contributor.authorLiu, J.
dc.contributor.authorZhan, C.
dc.contributor.authorJoksimovic, S.
dc.date.issued2024
dc.description.abstractLearning analytics (LA), like much of machine learning, assumes the training and test datasets come from the same distribution. Therefore, LA models built on past observations are (implicitly) expected to work well for future observations. However, this assumption does not always hold in practice because the dataset may drift. Recently, algorithmic fairness has gained significant attention. Nevertheless, algorithmic fairness research has paid little attention to dataset drift. Majority of the existing fairness algorithms are “statically” designed. Put another way, LA models tuned to be “fair” on past data are expected to still be “fair” when dealing with current/future data. However, it is counter-intuitive to deploy a statically fair algorithm to a nonstationary world. There is, therefore, a need to assess the impact of dataset drift on the unfairness of LA models. For this reason, we investigate the relationship between dataset drift and unfairness of LA models. Specifically, we first measure the degree of drift in the features (i.e., covariates) and target label of our dataset. After that, we train predictive models on the dataset and evaluate the relationship between the dataset drift and the unfairness of the predictive models. Our findings suggest a directly proportional relationship between dataset drift and unfairness. Further, we find covariate drift to have the most impact on unfairness of models as compared to target drift, and there are no guarantees that a once fair model would consistently remain fair. Our findings imply that “robustness” of fair LA models to dataset drift is necessary before deployment.
dc.description.statementofresponsibilityOscar Blessed Deho, Lin Liu, Jiuyong Li, Jixue Liu, Chen Zhan, and Srecko Joksimovic
dc.identifier.citationIEEE Transactions on Learning Technologies, 2024; 17:1007-1020
dc.identifier.doi10.1109/TLT.2024.3351352
dc.identifier.issn1939-1382
dc.identifier.issn1939-1382
dc.identifier.orcidLiu, J. [0000-0002-0794-0404]
dc.identifier.orcidZhan, C. [0000-0002-4794-8339]
dc.identifier.urihttps://hdl.handle.net/2440/143934
dc.language.isoen
dc.publisherInstitute of Electrical and Electronics Engineers
dc.relation.granthttp://purl.org/au-research/grants/arc/DP200101210
dc.rights© 2024 IEEE. Personal use is permitted
dc.source.urihttps://doi.org/10.1109/TLT.2024.3351352
dc.subjectDataset drift; ethical learning analytics; fairness; learning analytics (LA); predictive modeling; virtual learning environment
dc.titleWhen the Past != The Future: Assessing the Impact of Dataset Drift on the Fairness of Learning Analytics Models
dc.typeJournal article
pubs.publication-statusPublished

Files

Collections