Your fairness may vary: Group fairness of pretrained language models in toxic text classification

08/03/2021
by   Ioana Baldini, et al.
0

We study the performance-fairness trade-off in more than a dozen fine-tuned LMs for toxic text classification. We empirically show that no blanket statement can be made with respect to the bias of large versus regular versus compressed models. Moreover, we find that focusing on fairness-agnostic performance metrics can lead to models with varied fairness characteristics.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset