You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The tool identifies groups where an algorithm or AI system shows variations in performance. This type of monitoring is referred to as _anomaly detection_. To identify anomalous patterns, the tool uses <ahref="https://en.wikipedia.org/wiki/Cluster_analysis"target="_blank">clustering</a>. Clustering is a form of _unsupervised learning_. This means detecting disparate treatment (bias) does not require any data on protected attributes of users, such as gender, nationality, or ethnicity. The metric used to measure bias can be manually selected and is referred to as the `bias metric`.
90
+
The tool identifies groups where an algorithm or AI system shows variations in performance. This type of monitoring is referred to as *anomaly detection*. To identify anomalous patterns, the tool uses <ahref="https://en.wikipedia.org/wiki/Cluster_analysis"target="_blank">clustering</a>. Clustering is a form of *unsupervised learning*. This means detecting disparate treatment (bias) does not require any data on protected attributes of users, such as gender, nationality, or ethnicity. The metric used to measure bias can be manually selected and is referred to as the `bias metric`.
90
91
91
92
#### What data can be processed?
92
93
@@ -125,7 +126,7 @@ The tool identifies deviating clusters. A summary of the results is made availab
125
126
126
127
#### How is my data processed?
127
128
128
-
The tool is privacy-friendly because the data is processed entirely within the browser. The data does not leave your computer or the environment of your organization. The tool utilizes the computing power of your own computer to analyze the data. This type of browser-based software is referred to as _local-first_. The tool does not upload data to third parties, such as cloud providers. Instructions on how to host the tool and local-first architecture can be hosted locally within your own organization can be found on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
129
+
The tool is privacy-friendly because the data is processed entirely within the browser. The data does not leave your computer or the environment of your organization. The tool utilizes the computing power of your own computer to analyze the data. This type of browser-based software is referred to as *local-first*. The tool does not upload data to third parties, such as cloud providers. Instructions on how to host the tool and local-first architecture can be hosted locally within your own organization can be found on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
- The source code of the anolamy detection-algorithm is available on <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection"target="_blank">Github</a> and as a <ahref="https://pypi.org/project/unsupervised-bias-detection/"target="_blank">pip package</a>: `pip install unsupervised-bias-detection`.
- The architecture to run web apps local-first is also available on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
147
+
* The source code of the anolamy detection-algorithm is available on <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection"target="_blank">Github</a> and as a <ahref="https://pypi.org/project/unsupervised-bias-detection/"target="_blank">pip package</a>: `pip install unsupervised-bias-detection`.
* The architecture to run web apps local-first is also available on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
The tool uses the _Hierarchical Bias-Aware Clustering_ (HBAC) algorithm. HBAC processes input data according to the k-means (for numerical data) or k-modes (for categorical data) clustering algorithm. The HBAC-algorithm is introduced by Misztal-Radecka and Indurkya in a [scientific article](https://www.sciencedirect.com/science/article/abs/pii/S0306457321000285) as published in *Information Processing and Management* (2021). Our implementation of the HBAC-algorithm, including additional methodological checks to distinguish real bias from noise, such as sample splitting, statistical hypothesis testing and measuring cluster stability, can be found in the <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection/blob/master/README.md"target="_blank">unsupervised-bias-detection</a> pip package.
157
+
The tool uses the *Hierarchical Bias-Aware Clustering* (HBAC) algorithm. HBAC processes input data according to the k-means (for numerical data) or k-modes (for categorical data) clustering algorithm. The HBAC-algorithm is introduced by Misztal-Radecka and Indurkya in a [scientific article](https://www.sciencedirect.com/science/article/abs/pii/S0306457321000285) as published in *Information Processing and Management* (2021). Our implementation of the HBAC-algorithm, including additional methodological checks to distinguish real bias from noise, such as sample splitting, statistical hypothesis testing and measuring cluster stability, can be found in the <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection/blob/master/README.md"target="_blank">unsupervised-bias-detection</a> pip package.
158
158
159
159
{{< container_close >}}
160
160
@@ -188,11 +188,11 @@ Local-first computing is the opposite of cloud computing: the data is not upload
{{< accordion_item_open title="Innovation grant Dutch Ministry of the Interior" image="/images/supported_by/BZK.jpg" tag1="2024-25" >}}
198
198
@@ -218,7 +218,7 @@ In 2024, the SIDN Fund <a href="https://www.sidnfonds.nl/projecten/open-source-a
218
218
219
219
<!-- Awards and acknowledgements -->
220
220
221
-
{{< container_open title="Awards and acknowledgements" icon="fas fa-medal" id="awards-acknowledgements">}}
221
+
{{< container_open title="Awards and acknowledgements" icon="fas fa-medal" id="awards-acknowledgements">}}
222
222
223
223
This tool has received awards and is acknowledged by various <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection?tab=readme-ov-file#contributing-members"target="_blank">stakeholders</a>, including civil society organisations, industry representatives and academics.
224
224
@@ -246,16 +246,16 @@ The unsupervised bias detection tool is part of OECD's <a href="https://oecd.ai/
Key take-aways about unsupervised bias detection tool:
252
252
253
-
- <spanstyle="color:#005AA7">Quantitative-qualitative research method</span>: Data-driven bias testing combined with the balanced and context-sensitive judgment of human experts;
254
-
- <spanstyle="color:#005AA7">Unsupervised bias detection</span>: No user data needed on protected attributes (_unsupervised learning_);
255
-
- <spanstyle="color:#005AA7">Anolamy detection</span>: Scalable method based on statistical analysis;
256
-
- <spanstyle="color:#005AA7">Detects complex bias</span>: Identifies unfairly treated groups characterized by mixture of features, detects intersectional bias;
257
-
- <spanstyle="color:#005AA7">Model-agnostic</span>: Works for all binary classification algorithms and AI systems;
258
-
- <spanstyle="color:#005AA7">Open-source and not-for-profit</span>: User friendly and free to use for the entire AI auditing community.
253
+
* <spanstyle="color:#005AA7">Quantitative-qualitative research method</span>: Data-driven bias testing combined with the balanced and context-sensitive judgment of human experts;
254
+
* <spanstyle="color:#005AA7">Unsupervised bias detection</span>: No user data needed on protected attributes (*unsupervised learning*);
255
+
* <spanstyle="color:#005AA7">Anolamy detection</span>: Scalable method based on statistical analysis;
256
+
* <spanstyle="color:#005AA7">Detects complex bias</span>: Identifies unfairly treated groups characterized by mixture of features, detects intersectional bias;
257
+
* <spanstyle="color:#005AA7">Model-agnostic</span>: Works for all binary classification algorithms and AI systems;
258
+
* <spanstyle="color:#005AA7">Open-source and not-for-profit</span>: User friendly and free to use for the entire AI auditing community.
0 commit comments