About Us#

Mission#

Fairlearn is an open-source, community-driven project to help data scientists improve fairness of AI systems.

The project aspires to include:

  • A Python library for fairness assessment and improvement (fairness metrics, mitigation algorithms, plotting, etc.)

  • Educational resources covering organizational and technical processes for unfairness mitigation (comprehensive user guide, detailed case studies, Jupyter notebooks, white papers, etc.)

Development of Fairlearn is firmly grounded in the understanding that fairness in AI systems is a sociotechnical challenge. Because there are many complex sources of unfairness — some societal and some technical — it is not possible to fully “debias” a system or to guarantee fairness. Our goal is to enable humans to assess fairness-related harms, review the impacts of different mitigation strategies and then make trade-offs appropriate to their scenario.

Fairlearn is a community-driven open source project! The development and growth of Fairlearn are guided by the belief that meaningful progress toward fairer AI systems requires input from a breadth of perspectives, ranging from data scientists, developers, and business decision makers to the people whose lives may be affected by the predictions of AI systems.

Code of conduct#

Fairlearn follows the Fairlearn Organization’s Code of Conduct.

Project focus areas#

Last update: May 16, 2021

As an open-source project, Fairlearn strives to incorporate the best of research and practice. AI is a rapidly evolving field, and fairness in AI is all the more so. We therefore encourage researchers, practitioners, and other stakeholders to contribute fairness metrics and assessment tools, unfairness mitigation algorithms, case studies and other educational materials to Fairlearn as we experiment, learn, and evolve the project together.

Below we list the key areas that we prioritize in the short and medium term, but we are happy to consider other directions if they are aligned with the mission of Fairlearn and there is enough commitment from the contributors. If you want to get involved, please reach out. For concrete opportunities and work in progress please review our issues.

  1. Decrease adoption barriers for the current assessment and mitigation tools in Fairlearn

    • Improve existing use cases and create new ones that make fairness issues concrete. These use cases may or may not use the Fairlearn package. To accelerate this process, we are experimenting with weekly sessions where people can discuss ideas, ongoing projects, and individual example notebooks in detail.

    • Improve the project documentation: critique the current content, expand user guides, improve writing and examples, port the current Python documentation to the numpydoc format. Convert existing example notebooks from .ipynb into .py files that render properly on the website, while still being downloadable as .py or .ipynb files.

    • Improve the usability, relevance, and look of Fairlearn website with the audience of practitioners in mind. Engage by participating in the discussions on community calls or in the corresponding discussions. You can also give us feedback by filing a new issue.

    • Improve the usability and relevance of fairness metrics by critiquing and improving the current metrics API, suggesting new metrics motivated by concrete use cases, and implementing the new metrics. The issues page contains several metrics tasks.

    • Move towards compatibility with scikit-learn: identify incompatible aspects, improve code towards compatibility. While we aim for compatibility there may be aspects that are too restricting for Fairlearn, so this may need to be evaluated on a case-by-case basis.

  2. Grow and nurture a diverse community of contributors

    • Reach out with feedback on what is working and what is not working; suggest how to improve things; point out where documentation, our processes or any other aspect of the projects create barriers of entry.

    • Participate in our weekly community calls. We also work with universities to engage student contributors through course projects and are also open to other forms of collaboration — let us know if you are interested.

    • Improve the Fairlearn website and documentation. See the contributor guide on how to contribute to our docs.

    • Add tests and improve testing infrastructure.

  3. Create metrics, assessment tools, and algorithms to cover more complex ML tasks

    • Create notebooks and use cases that deal with concrete fairness issues in complex ML tasks including ranking, counterfactual estimation, text, computer vision, speech, etc.

    • Lead and participate in contribution efforts around under-researched, but practically relevant ML areas in ranking, counterfactual estimation, text, computer vision, speech, etc. These are likely to be mixed research / practice efforts and we expect substantial contributor commitment before embarking on these.

Governance#

Fairlearn is a project of the Fairlearn Organization and follows the Fairlearn Organization’s Project Governance.

Maintainers#

The current maintainers of the Fairlearn project are


Adrin Jalali


Hilde Weerts


Michael Madaio


Miro Dudik


Richard Edgar


Roman Lutz


Allie Saizan

Project history#

Fairlearn was started in 2018 by Miro Dudik from Microsoft Research as a Python package to accompany the research paper, A Reductions Approach to Fair Classification [1]. The package provided a reduction algorithm for mitigating unfairness in binary classification models — a setting that was commonly studied in the machine learning community. The paper and the Python package were well received, so Miro Dudik and Hanna Wallach with their collaborators sought to translate the research into an industry context. However, they discovered that practitioners typically need to address more fundamental fairness issues before applying specific algorithms, and that mitigating unfairness in binary classification models is a relatively rare use case. They also discovered that fairness assessment is a common need, along with access to domain-specific guides to fairness metrics and unfairness mitigation algorithms. Additionally, many use cases take the form of regression or ranking, rather than classification. As a result of these insights, fairness assessment and use-case notebooks became key components of Fairlearn. Fairlearn also focuses on machine learning tasks beyond binary classification.

The project was greatly expanded in the second half of 2019 thanks to the involvement of many contributors from Azure ML and Microsoft Research. At that time, the project started to have regular releases.

In 2021 Fairlearn adopted neutral governance and since then the project is completely community-driven.

Citing Fairlearn#

If you wish to cite Fairlearn in your work, please use the following:

@misc{weerts2023fairlearn,
      title={Fairlearn: Assessing and Improving Fairness of AI Systems},
      author={Hilde Weerts and Miroslav Dudík and Richard Edgar and Adrin Jalali and Roman Lutz and Michael Madaio},
      journal={Journal of Machine Learning Research},
      year={2023},
      volume={24},
      number={257},
      pages={1--8},
      url={http://jmlr.org/papers/v24/23-0389.html}
}

Frequently asked questions#

See our Frequently asked questions page.

Funding#

Fairlearn is a community-driven project. However, several companies and academic institutions help to assure its sustainability. We would like to thank the following funders.

Infrastructure support#

We would also like to thank the following for free CPU time on their continuous integration servers: