Skip to content

Commit 7f903bf

Browse files
authored
Merge pull request #126 from NGO-Algorithm-Audit/feature/structural_edits
Update BDT.md
2 parents ef139c8 + 3941a64 commit 7f903bf

File tree

1 file changed

+9
-14
lines changed
  • content/english/technical-tools

1 file changed

+9
-14
lines changed

content/english/technical-tools/BDT.md

Lines changed: 9 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -117,33 +117,28 @@ The bias detection tool currently works for tabular numerical and categorical da
117117

118118
##### Why this bias detection tool?
119119

120-
* No data needed on protected attributes of users (unsupervised bias detection);
121-
* Model-agnostic
122-
* Informs human experts which characteristics of AI-sytem behavior should manually be scrutinized;
123-
* Connecting quantitative, statistical tools with the qualitative doctrine of law and ethics to assess fair AI;
124-
* Developed open-source and not-for-profit.
120+
* **Quantitative-qualitative joint method**: Data-driven bias testing combined with the balanced and context-sensitive judgment of human experts;
121+
* **Unsupervised bias detection**: No user data needed on protected attributes;
122+
* **Bias scan tool**: Scalable method based on statistical learning to detect algorithmic bias;
123+
* **Detects complex bias**: Identifies unfairly treated groups characterized by mixture of features, detects intersectional bias;
124+
* **Model-agnostic**: Works for all AI systems;
125+
* **Open-source and not-for-profit**: Easy to use and available for the entire AI auditing community.
125126

126127
##### By whom can the bias detection tool be used? 
127128

128129
The bias detection tool allows the entire ecosystem involved in auditing AI, e.g., data scientists, journalists, policy makers, public- and private auditors, to use quantitative methods to detect bias in AI systems.
129130

130131
##### What does the tool compute? 
131132

132-
A statistical method is used to compute which clusters are relatively often misclassified by an AI system. A cluster is a group of data points sharing similar features. On these features the AI-system is initially trained. The tool identifies and visualizes the found clusters automatically. The tool also assesses how individuals in a deviating cluster differ (in terms of the provided features) from others outside the cluster. If the differences are statistically significant is directly tested by means of [Welch’s two-samples t-test](https://en.wikipedia.org/wiki/Welch%27s_t-test) for unequal variances. All results kan directly be downloaded as a pdf file.
133+
A statistical method is used to compute for which clusters an AI system underperforms. A cluster is a group of data points sharing similar features. On these features the AI system is initially trained. The tool identifies and visualizes the found clusters automatically. The tool also assesses how individuals in a deviating cluster differ (in terms of the provided features) from other data points outside the cluster. The differences between these clusters are tested on statistical significance. All results can directly be downloaded as a pdf file.
133134

134135
#### The tool detects prohibited discrimination in AI? 
135136

136137
No. The bias detection tool serves as a starting point to assess potentially unfair AI classifiers with the help of subject-matter expertise. The features of identified clusters are examined on critical links with protected grounds, and whether the measured disparities are legitimate. This is a qualitative assessment for which the context-sensitive legal doctrine provides guidelines, i.e., to assess the legitimacy of the aim pursued and whether the means of achieving that aim are appropriate and necessary. In a [case study](/algoprudence/cases/bert-based-disinformation-classifier-aa202301/) of Algorithm Audit – in which the bias detection tool was tested on a BERT-based disinformation classifier – a normative advice commission argued that the measured quantitative deviations could be legitimised. Legitimisation of unequal treatment is a context-sensitive taks for which legal frameworks exist, such an assessment of proportionality, necessity and suitability. This qualitative judgement will always be a human task.
137138

138-
##### For what type of AI does the tool work? 
139-
140-
Currently, only binary classification algorithms can be reviewed. For instance, prediction of loan approval (yes/no), disinformation detection (true/false) or disease detection (positive/negative).
141-
142-
##### What happens with my data?
143-
144-
Your csv file is uploaded to a Amazon Web Services (AWS) bucket, where it is processed by Python code. Once the HBAC-algorithm has identified clusters, the results are sent back to the browser and the data is immediately deleted. Usually, your data is stored only for 5-10 seconds in the cloud environment. The web application is built according to the below architecture diagram.
139+
##### How is my data processed?
145140

146-
{{< image id="architecture-diagram" width_desktop="12" width_mobile="12" image1="/images/BDT/architecture.png" alt1="Architecture diagram bias detection tool web app" caption1="Architecture diagram bias detection tool web app" >}}
141+
The tool is privacy preserving. It uses computing power of your own computer to analyze a dataset. In this architectural setup, data is processed entirely on your device and it not uploaded to any third party, such as cloud providers. This local-only feature allows organisations to securely use the tool with proprietary data. The used software is also available as <a href="https://pypi.org/project/unsupervised-bias-detection/" target="_blank">pip package</a> `unsupervised-bias-detection`. [![!pypi](https://img.shields.io/pypi/v/unsupervised-bias-detection?logo=pypi\&color=blue)](https://pypi.org/project/unsupervised-bias-detection/)
147142

148143
##### In sum 
149144

0 commit comments

Comments
 (0)