You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* Update model-card-annotated.md
Food for thought: Adding a section that deals directly with Redteaming (as that is becoming a big interest/concern). Hit specifically on CSAM, as we have been inspired by Thorn and Rebecca Portnoff to be more proactive about this.
* Update model-card-annotated.md
* Update docs/hub/model-card-annotated.md
Co-authored-by: Julien Chaumond <[email protected]>
* Update model-card-annotated.md
Suggesting updates based on multiple discussions.
This includes:
- The github discussion here
- Discussions in the Ethics meetings about pros and cons of different documentation requirements
- Discussions with Thorn and All Tech is Human on responsible AI development
* Update model-card-annotated.md
Iterating on wording to introduce a new Evaluation section.
---------
Co-authored-by: Julien Chaumond <[email protected]>
Copy file name to clipboardExpand all lines: docs/hub/model-card-annotated.md
+17-7Lines changed: 17 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -132,7 +132,7 @@ _List how the model may foreseeably be misused (used in a way it will not work f
132
132
133
133
**Section Overview:** This section identifies foreseeable harms, misunderstandings, and technical and sociotechnical limitations. It also provides information on warnings and potential mitigations. Bias, risks, and limitations can sometimes be inseparable/refer to the same issues. Generally, bias and risks are sociotechnical, while limitations are technical:
134
134
- A **bias** is a stereotype or disproportionate performance (skew) for some subpopulations.
135
-
- A **risk** is a socially-relevant issue that the model might cause.
135
+
- A **risk** is a socially-relevant issue that the model might cause.
136
136
- A **limitation** is a likely failure mode that can be addressed following the listed Recommendations.
**Section Overview:** This section describes the evaluation protocols, what is being measured in the evaluation, and provides the results. Evaluation is ideally constructed with factors, such as domain and demographic subgroup, and metrics, such as accuracy, which are prioritized in light of foreseeable error contexts and groups. Target fairness metrics should be decided based on which errors are more likely to be problematic in light of the model use. You can also specify your model's evaluation results in a structured way in the model card metadata. Results are parsed by the Hub and displayed in a widget on the model page. See https://huggingface.co/docs/hub/model-cards#evaluation-results.
185
+
**Section Overview:** This section describes the evaluation protocols, what is being measured in the evaluation, and provides the results. Evaluation ideally has at least two parts, with one part looking at quantitative measurement of general performance ([Testing Data, Factors & Metrics](#testing-data-factors--metrics)), such as may be done with benchmarking; and another looking at performance with respect to specific social safety issues ([Societal Impact Assessment](#societal-impact-assessment-optional)), such as may be done with red-teaming. You can also specify your model's evaluation results in a structured way in the model card metadata. Results are parsed by the Hub and displayed in a widget on the model page. See https://huggingface.co/docs/hub/model-cards#evaluation-results.
186
186
187
187
188
188
## Testing Data, Factors & Metrics
189
189
190
+
_Evaluation is ideally **disaggregated** with respect to different factors, such as task, domain and population subgroup; and calculated with metrics that are most meaningful for foreseeable contexts of use. Equal evaluation performance across different subgroups is said to be "fair" across those subgroups; target fairness metrics should be decided based on which errors are more likely to be problematic in light of the model use. However, this section is most commonly used to report aggregate evaluation performance on different task benchmarks._
191
+
190
192
### Testing Data
191
193
192
194
`testing_data`
193
195
194
-
_Ideally this links to a Dataset Card for the testing data._
196
+
_Describe testing data or link to its Dataset Card._
195
197
196
198
### Factors
197
199
198
200
`testing_factors`
199
201
200
-
_What are the foreseeable characteristics that will influence how the model behaves? This includes domain and context, as well as population subgroups. Evaluation should ideally be **disaggregated** across factors in order to uncover disparities in performance._
202
+
_What are the foreseeable characteristics that will influence how the model behaves? Evaluation should ideally be disaggregated across these factors in order to uncover disparities in performance._
201
203
202
204
### Metrics
203
205
204
206
`testing_metrics`
205
207
206
-
_What metrics will be used for evaluation in light of tradeoffs between different errors?_
208
+
_What metrics will be used for evaluation?_
207
209
208
210
## Results
209
211
210
212
`results`
211
213
212
-
213
214
_Results should be based on the Factors and Metrics defined above._
214
215
215
216
### Summary
216
217
217
-
218
218
`results_summary`
219
219
220
220
_What do the results say? This can function as a kind of tl;dr for general audiences._
221
221
222
+
## Societal Impact Assessment [optional]
223
+
224
+
_Use this free text section to explain how this model has been evaluated for risk of societal harm, such as for child safety, NCII, privacy, and violence. This might take the form of answers to the following questions:_
225
+
226
+
-_Is this model safe for kids to use? Why or why not?_
227
+
-_Has this model been tested to see if it can be used for non-consensual intimate imagery (including CSEM)?_
228
+
-_Has this model been tested to see if it can be used for violent activities, or to depict violence? What were the results?_
229
+
230
+
_Quantitative numbers on each issue may also be provided._
231
+
222
232
# Model Examination [optional]
223
233
224
234
**Section Overview:** This is an experimental section some developers are beginning to add, where work on explainability/interpretability may go.
0 commit comments