A General Framework for Fair Regression
Fairness, through its many forms and definitions, has become an important issue facing the machine learning community. In this work, we consider how to incorporate group fairness constraints into kernel regression methods, applicable to Gaussian processes, support vector machines, neural network reg...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2019-07-01
|
Series: | Entropy |
Subjects: | |
Online Access: | https://www.mdpi.com/1099-4300/21/8/741 |
id |
doaj-c4083c91290649f6918da34868affe48 |
---|---|
record_format |
Article |
spelling |
doaj-c4083c91290649f6918da34868affe482020-11-25T02:44:09ZengMDPI AGEntropy1099-43002019-07-0121874110.3390/e21080741e21080741A General Framework for Fair RegressionJack Fitzsimons0AbdulRahman Al Ali1Michael Osborne2Stephen Roberts3Department of Engineering Science, University of Oxford, Oxford OX13PJ, UKFaculty of Business and Law, Northampton University, Northampton NN15PH, UKDepartment of Engineering Science, University of Oxford, Oxford OX13PJ, UKDepartment of Engineering Science, University of Oxford, Oxford OX13PJ, UKFairness, through its many forms and definitions, has become an important issue facing the machine learning community. In this work, we consider how to incorporate group fairness constraints into kernel regression methods, applicable to Gaussian processes, support vector machines, neural network regression and decision tree regression. Further, we focus on examining the effect of incorporating these constraints in decision tree regression, with direct applications to random forests and boosted trees amongst other widespread popular inference techniques. We show that the order of complexity of memory and computation is preserved for such models and tightly binds the expected perturbations to the model in terms of the number of leaves of the trees. Importantly, the approach works on trained models and hence can be easily applied to models in current use and group labels are only required on training data.https://www.mdpi.com/1099-4300/21/8/741machine learningalgorithmic fairnesskernel methodsconstrained learningGaussian processdecision treeneural network |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Jack Fitzsimons AbdulRahman Al Ali Michael Osborne Stephen Roberts |
spellingShingle |
Jack Fitzsimons AbdulRahman Al Ali Michael Osborne Stephen Roberts A General Framework for Fair Regression Entropy machine learning algorithmic fairness kernel methods constrained learning Gaussian process decision tree neural network |
author_facet |
Jack Fitzsimons AbdulRahman Al Ali Michael Osborne Stephen Roberts |
author_sort |
Jack Fitzsimons |
title |
A General Framework for Fair Regression |
title_short |
A General Framework for Fair Regression |
title_full |
A General Framework for Fair Regression |
title_fullStr |
A General Framework for Fair Regression |
title_full_unstemmed |
A General Framework for Fair Regression |
title_sort |
general framework for fair regression |
publisher |
MDPI AG |
series |
Entropy |
issn |
1099-4300 |
publishDate |
2019-07-01 |
description |
Fairness, through its many forms and definitions, has become an important issue facing the machine learning community. In this work, we consider how to incorporate group fairness constraints into kernel regression methods, applicable to Gaussian processes, support vector machines, neural network regression and decision tree regression. Further, we focus on examining the effect of incorporating these constraints in decision tree regression, with direct applications to random forests and boosted trees amongst other widespread popular inference techniques. We show that the order of complexity of memory and computation is preserved for such models and tightly binds the expected perturbations to the model in terms of the number of leaves of the trees. Importantly, the approach works on trained models and hence can be easily applied to models in current use and group labels are only required on training data. |
topic |
machine learning algorithmic fairness kernel methods constrained learning Gaussian process decision tree neural network |
url |
https://www.mdpi.com/1099-4300/21/8/741 |
work_keys_str_mv |
AT jackfitzsimons ageneralframeworkforfairregression AT abdulrahmanalali ageneralframeworkforfairregression AT michaelosborne ageneralframeworkforfairregression AT stephenroberts ageneralframeworkforfairregression AT jackfitzsimons generalframeworkforfairregression AT abdulrahmanalali generalframeworkforfairregression AT michaelosborne generalframeworkforfairregression AT stephenroberts generalframeworkforfairregression |
_version_ |
1724767214303707136 |