Skip to content

Latest commit

 

History

History
34 lines (27 loc) · 1.48 KB

README.md

File metadata and controls

34 lines (27 loc) · 1.48 KB

Fairness-Aware Structured Pruning in Transformers (AAAI 2024)

We show that certain attention heads are responsible for bias and pruning them improves fairness.

How it works

The figure below illustrates how FASP is applied to a model with $6$ layers and $12$ heads per layer, e.g. DistilGPT-2. We identify and exclude the heads that significantly impact performance from the pruning process (black squares). Subsequently, the remaining heads are prioritized for removal based on their contribution to bias, ensuring that the heads contributing the most to bias are pruned first (red squares).

Running the experiments

Get started with the Colab tutorial, FASP_AAAI24_reproducibility.ipynb, which guides you through the process of downloading the models, understanding the preprocessing steps, and creating the scripts required to run the experiments.

Citation

@inproceedings{zayed2024fairness,
  title={Fairness-aware structured pruning in transformers},
  author={Zayed, Abdelrahman and Mordido, Gon{\c{c}}alo and Shabanian, Samira and Baldini, Ioana and Chandar, Sarath},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={38},
  number={20},
  pages={22484--22492},
  year={2024}
}