Trained machine learning models are increasingly used to perform high-impact tasks in areas such as law enforcement, medicine, education, and employment
In order to clarify the intended use cases of machine learning models and minimize their usage in contexts for which they are not well suited, we recommend that released models be accompanied by documentation detailing their performance characteristics
Short documents accompanying trained machine learning models that provide benchmarked evaluation in a variety of conditions, such as across different cultural, demographic, or phenotypic groups (e.g., race, geographic location, sex, Fitzpatrick skin type) and intersectional groups (e.g., age and race, or sex and Fitzpatrick skin type) that are relevant to the intended application domains
While the focus is primarily on human-centered machine learning models in the application fields of computer vision and natural language processing, this framework can be used to document any trained machine learning model
Model cards are proposed as a step towards the responsible democratization of machine learning and related artificial intelligence technology, increasing transparency into how well artificial intelligence technology works
The goal is to encourage those releasing trained machine learning models to accompany model releases with similar detailed evaluation numbers and other relevant documentation
Not only does this practice improve model understanding and help to standardize decision making processes for invested stakeholders, but it also encourages forward-looking model analysis techniques.
Slicing the evaluation across groups functions to highlight errors that may fall disproportionately on some groups of people, and accords with many recent notions of mathematical fairness.
Including group analysis as part of the reporting procedure prepares stakeholders to begin to gauge the fairness and inclusion of future outcomes of the machine learning system.
Future research could include creating robust evaluation datasets and protocols for the types of disaggregated evaluation we advocate for in this work, for example, by including differential privacy mechanisms so that individuals in the testing set cannot be uniquely identified by their characteristics.
Discloses information about a trained machine learning model, including how it was built, what assumptions were made during its development, what type of model behavior different cultural, demographic, or phenotypic population groups may experience, and an evaluation of how well the model performs with respect to those groups
Basic information about the model, including the person or organization developing it, model date, version, type, training algorithms, parameters, fairness constraints, features, paper/resource for more information, citation details, license, and where to send questions or comments
Details on the dataset(s) used for training the model, mirroring the Evaluation Data section if possible, or providing minimal allowable information on the distribution over various factors