-
Notifications
You must be signed in to change notification settings - Fork 903
Description
Proposal: Species as a Protected Attribute Category
Summary
Add species as a protected attribute category in AIF360, with metrics for evaluating speciesist bias in NLP models. This extends AIF360's fairness framework to cover a documented, measurable form of bias that has been established across 5+ peer-reviewed publications.
Academic Foundation
Multiple peer-reviewed papers have documented speciesist bias as a measurable phenomenon in AI systems:
-
Hagendorff, Bossert, Tse & Singer (2023). "Speciesist bias in AI: how AI applications perpetuate the exploitation of animals." AI and Ethics. DOI: 10.1007/s43681-023-00380-w
- Found GPT-3 associates farmed animals with violence
- The Delphi moral AI judged killing a factory-farmed pig as "okay" but killing a dog as "wrong"
- Explicitly calls for fairness frameworks to "widen their scope and include mitigation measures for speciesist biases"
-
Takeshita, Rzepka & Araki (2022). "Speciesist language and nonhuman animal bias in English masked language models." Information Processing & Management.
- Demonstrated BERT, DistilBERT, RoBERTa, and ALBERT all associate harmful words with nonhuman animals
-
Hagendorff et al. (2025). "SpeciesismBench." arXiv:2508.11534.
- 1,003-item benchmark showing LLMs "frequently normalized harm toward farmed animals while refusing to do so for non-farmed animals"
-
AI-for-Animals (2025). "AHA Benchmark." arXiv:2503.04804.
- 4,350-item dataset finding species-dependent risks of harm in LLM outputs
-
Takeshita et al. (2024). Found speciesism among NLP researchers themselves and in annotated NLP data. AI and Ethics.
What This Would Look Like
- A new
SpeciesistBiasMetricclass that evaluates differential treatment of animal species in model outputs - Species as a recognized protected attribute alongside race, gender, age, etc.
- Integration with existing AIF360 metrics infrastructure
- Test datasets based on SpeciesismBench and AHA Benchmark
Why AIF360
Hagendorff et al. (2023) specifically identified fairness toolkits as the right venue for speciesist bias metrics. AIF360 is the most widely-used open-source AI fairness toolkit and would set the standard for other frameworks.
Willingness to Contribute
We (Open Paws) are willing to implement this module. We've already built speciesist bias evaluation tools and have experience with the relevant benchmarks.