You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The tool identifies groups where an algorithm or AI system shows variations in performance. This type of monitoring is referred to as *anomaly detection*. To identify anomalous patterns, the tool uses <ahref="https://en.wikipedia.org/wiki/Cluster_analysis"target="_blank">clustering</a>. Clustering is a form of _unsupervised learning_. This means detecting disparate treatment (bias) does not require any data on protected attributes of users, such as gender, nationality, or ethnicity. The metric used to measure bias can be manually selected and is referred to as the `bias metric`.
88
+
89
+
The tool identifies groups where an algorithm or AI system shows variations in performance. This type of monitoring is referred to as _anomaly detection_. To identify anomalous patterns, the tool uses <ahref="https://en.wikipedia.org/wiki/Cluster_analysis"target="_blank">clustering</a>. Clustering is a form of _unsupervised learning_. This means detecting disparate treatment (bias) does not require any data on protected attributes of users, such as gender, nationality, or ethnicity. The metric used to measure bias can be manually selected and is referred to as the `bias metric`.
90
90
91
91
#### What data can be processed?
92
-
The tool processes all data in table format. The type of data (numerical, categorical, time, etc.) is automatically detected. One column must be selected as the `bias metric` – which should be a numerical value. The user must specify whether a high or low value of the `bias metric` is considered better. For example: for an error rate, a low value is better, while for accuracy, a high value is better.
92
+
93
+
The tool processes all data in table format. The type of data (numerical, categorical, time, etc.) is automatically detected. One column must be selected as the `bias metric` – which should be a numerical value. The user must specify whether a high or low value of the `bias metric` is considered better. For example: for an error rate, a low value is better, while for accuracy, a high value is better.
93
94
94
95
The tool contains a demo data for which output is generated. Hit the 'Try it out' button.
95
96
@@ -113,6 +114,7 @@ The tool contains a demo data for which output is generated. Hit the 'Try it out
113
114
<br>
114
115
115
116
#### What does the tool return?
117
+
116
118
The tool identifies deviating clusters. A summary of the results is made available in a bias analysis report that can be downloaded as a pdf. All identified clusters can be downloaded in a .json file. The tool specifically focuses on the most negatively deviating cluster and provides a description of this cluster. These results serve as a starting point for further investigation by domain experts, who can assess whether the observed disparities are indeed undesirable. The tool also visualizes the outcomes.
117
119
118
120
#### Overview of process
@@ -122,39 +124,32 @@ The tool identifies deviating clusters. A summary of the results is made availab
122
124
</div>
123
125
124
126
#### How is my data processed?
125
-
The tool is privacy-friendly because the data is processed entirely within the browser. The data does not leave your computer or the environment of your organization. The tool utilizes the computing power of your own computer to analyze the data. This type of browser-based software is referred to as *local-first*. The tool does not upload data to third parties, such as cloud providers. Instructions on how to host the tool and local-first architecture can be hosted locally within your own organization can be found on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
127
+
128
+
The tool is privacy-friendly because the data is processed entirely within the browser. The data does not leave your computer or the environment of your organization. The tool utilizes the computing power of your own computer to analyze the data. This type of browser-based software is referred to as _local-first_. The tool does not upload data to third parties, such as cloud providers. Instructions on how to host the tool and local-first architecture can be hosted locally within your own organization can be found on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
- The source code of the anolamy detection-algorithm is available on <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection"target="_blank">Github</a> and as a <ahref="https://pypi.org/project/unsupervised-bias-detection/"target="_blank">pip package</a>: `pip install unsupervised-bias-detection`.
- The source code of the anolamy detection-algorithm is available on <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection"target="_blank">Github</a> and as a <ahref="https://pypi.org/project/unsupervised-bias-detection/"target="_blank">pip package</a>: `pip install unsupervised-bias-detection`.
- The architecture to run web apps local-first is also available on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
Local-first computing is the opposite of cloud computing: the data is not uploaded to third-parties, such as a cloud providers, and is processed by your own computer. The data attached to the tool therefore doesn't leave your computer or the environment of your organization. The tool is privacy-friendly because the data can be processed within the mandate of your organisation and doesn't need to be shared with new parties. The unsupervised bias detection tool can also be hosted locally within your organization. Instructions, including the source code or the web app, can be found on <ahref="https://github.com/NGO-Algorithm-Audit/local-first-web-tool"target="_blank">Github</a>.
188
180
189
181
#### Overview of local-first architecture
@@ -194,19 +186,18 @@ Local-first computing is the opposite of cloud computing: the data is not upload
{{< accordion_item_open title="Innovation grant Dutch Ministry of the Interior" image="/images/supported_by/BZK.jpg" tag1="2024-25" >}}
208
198
209
199
##### Description
200
+
210
201
In partnership with the Dutch Executive Agency for Education and the Dutch Ministry of the Interior, Algorithm Audit has been developing and testing this tool from July 2024 to July 2025, supported by an <ahref="https://www.digitaleoverheid.nl/overzicht-van-alle-onderwerpen/innovatie/innovatiebudget/toekenning-innovatiebudget-2024/"target="_blank">Innovation grant</a> from the annual competition hosted by the Dutch Ministry of the Interior. Project progress was shared at a community gathering on 13-02-2025.
211
202
212
203

@@ -225,19 +216,18 @@ In 2024, the SIDN Fund <a href="https://www.sidnfonds.nl/projecten/open-source-a
225
216
226
217
{{< container_close >}}
227
218
228
-
229
-
230
219
<!-- Awards and acknowledgements -->
231
220
232
221
{{< container_open title="Awards and acknowledgements" icon="fas fa-medal" id="awards-acknowledgements">}}
233
222
234
223
This tool has received awards and is acknowledged by various <ahref="https://github.com/NGO-Algorithm-Audit/unsupervised-bias-detection?tab=readme-ov-file#contributing-members"target="_blank">stakeholders</a>, including civil society organisations, industry representatives and academics.
Under the name Joint Fairness Assessment Method (JFAM) the unsupervised bias detection tool has been selected as a finalist in <ahref="https://hai.stanford.edu/ai-audit-challenge-2023-finalists"target="_blank">Stanford’s AI Audit Competition 2023</a>.
242
232
243
233
{{< accordion_item_close >}}
@@ -254,8 +244,6 @@ The unsupervised bias detection tool is part of OECD's <a href="https://oecd.ai/
0 commit comments