{"iri":"https://folio.openlegalstandard.org/RCZdG7bztfQT8BuzFadwgir","label":"Opacity","sub_class_of":["https://folio.openlegalstandard.org/RCafxux1zDpXGZzqVBKNVpa"],"parent_class_of":[],"is_defined_by":null,"see_also":[],"comment":null,"deprecated":false,"preferred_label":null,"alternative_labels":[],"translations":{},"hidden_label":null,"definition":"Opacity in an AI system refers to the condition where one or more features, such as processes, the provenance of datasets, functions, output, or behavior, are unavailable or incomprehensible to all stakeholders. Opacity is usually considered an antonym for transparency and can hinder understanding, trust, and accountability in AI systems.","examples":[],"notes":[],"history_note":null,"editorial_note":null,"in_scheme":null,"identifier":null,"description":null,"source":null,"country":null}