Replies: 2 comments 2 replies
-
I'd like the advocate for use-case based prioritisation of what is validated, using #384 as a starting point. |
Beta Was this translation helpful? Give feedback.
-
Philosophically I think this is great, but I'd love to see more atomized and detailed steps. I think there's a conversation we often get into about improving quality which then leads into the murky "how do we define quality?" question which then leads us into the awkward place where we try to quantify quality. I'll assert that trying to quantify quality is a bad idea and leads people to cynical states of being which is less than ideal.
This is great. Can we (you) start iterating these fields? Maybe open a PR with attached (minimal) RFD to make the change on a per-field basis? I'd be happy pair on these too to help get the ball rolling. If these are small changes it may be worth making the case that they can go into 5.x rather than waiting for 6 👍 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I would like to propose a discussion about implementing minimal validation for all populated fields within the 6.0.0 CVE JSON schema.
Currently, while some fields have well-defined requirements, others lack basic checks for data quality. This proposal is not about making optional fields mandatory, but rather ensuring that if a field is populated with data, that data is substantive and useful to consumers of the CVE List. This applies not only to strings but also to arrays, numbers, and other data types.
Why this is important:
Improved Data Quality: This approach helps prevent the submission of fields with meaningless content, such as empty arrays, zero values for counts, or single-character strings. By setting a minimum standard for all data types, we can ensure the information is always useful.
Enhanced Reliability: Enforcing minimal validation across all fields makes the data more reliable and easier to process for automated systems and tools. It reduces the need to handle a wide range of edge cases and invalid data formats.
Consistency Across the Board: Applying this principle universally creates a more consistent and robust schema. It ensures that the integrity of the data is maintained, regardless of the field type.
Proposed Action:
I suggest we review all fields in the schema and establish a set of minimal validation rules for them. These rules would only apply if a field is present in the record. For example:
Strings: Require a minLength greater than zero (e.g., minLength: 1) to prevent empty strings.
Arrays: Require a minItems greater than zero (e.g., minItems: 1) to prevent empty arrays.
Numbers/Integers: For fields that represent a count or a measure, require a minimum value greater than zero (e.g., minimum: 1) to ensure the value is meaningful.
This change would be a significant step toward improving the overall quality and usability of the CVE data. I look forward to the community's thoughts on this proposal.
Beta Was this translation helpful? Give feedback.
All reactions