Information technology — Artificial intelligence — Overview of trustworthiness in artificial intelligence

This document surveys topics related to trustworthiness in AI systems, including the following: — approaches to establish trust in AI systems through transparency, explainability, controllability, etc.; — engineering pitfalls and typical associated threats and risks to AI systems, along with possible mitigation techniques and methods; and — approaches to assess and achieve availability, resiliency, reliability, accuracy, safety, security and privacy of AI systems. The specification of levels of trustworthiness for AI systems is out of the scope of this document.

Technologies de l'information — Intelligence artificielle — Examen d'ensemble de la fiabilité en matière d'intelligence artificielle

General Information

Status
Published
Publication Date
27-May-2020
Current Stage
6060 - International Standard published
Start Date
28-May-2020
Completion Date
28-May-2020
Ref Project

Buy Standard

Technical report
ISO/IEC TR 24028:2020 - Information technology -- Artificial intelligence -- Overview of trustworthiness in artificial intelligence
English language
43 pages
sale 15% off
Preview
sale 15% off
Preview
Technical report
ISO/IEC TR 24028:2020 - Information technology -- Artificial intelligence -- Overview of trustworthiness in artificial intelligence
English language
43 pages
sale 15% off
Preview
sale 15% off
Preview

Standards Content (Sample)

TECHNICAL ISO/IEC TR
REPORT 24028
First edition
2020-05
Information technology —
Artificial intelligence — Overview
of trustworthiness in artificial
intelligence
Technologies de l'information — Intelligence artificielle — Examen
d'ensemble de la fiabilité en matière d'intelligence artificielle
Reference number
ISO/IEC TR 24028:2020(E)
©
ISO/IEC 2020

---------------------- Page: 1 ----------------------
ISO/IEC TR 24028:2020(E)

COPYRIGHT PROTECTED DOCUMENT
© ISO/IEC 2020
All rights reserved. Unless otherwise specified, or required in the context of its implementation, no part of this publication may
be reproduced or utilized otherwise in any form or by any means, electronic or mechanical, including photocopying, or posting
on the internet or an intranet, without prior written permission. Permission can be requested from either ISO at the address
below or ISO’s member body in the country of the requester.
ISO copyright office
CP 401 • Ch. de Blandonnet 8
CH-1214 Vernier, Geneva
Phone: +41 22 749 01 11
Fax: +41 22 749 09 47
Email: copyright@iso.org
Website: www.iso.org
Published in Switzerland
ii © ISO/IEC 2020 – All rights reserved

---------------------- Page: 2 ----------------------
ISO/IEC TR 24028:2020(E)

Contents Page
Foreword .v
Introduction .vi
1 Scope . 1
2 Normative references . 1
3 Terms and definitions . 1
4 Overview . 7
5 Existing frameworks applicable to trustworthiness . 7
5.1 Background . 7
5.2 Recognition of layers of trust. 8
5.3 Application of software and data quality standards . 8
5.4 Application of risk management.10
5.5 Hardware-assisted approaches .10
6 Stakeholders .11
6.1 General concepts .11
6.2 Types.12
6.3 Assets .12
6.4 Values .13
7 Recognition of high-level concerns .13
7.1 Responsibility, accountability and governance .13
7.2 Safety .14
8 Vulnerabilities, threats and challenges .14
8.1 General .14
8.2 AI specific security threats .15
8.2.1 General.15
8.2.2 Data poisoning .15
8.2.3 Adversarial attacks.15
8.2.4 Model stealing .16
8.2.5 Hardware-focused threats to confidentiality and integrity .16
8.3 AI specific privacy threats .16
8.3.1 General.16
8.3.2 Data acquisition .16
8.3.3 Data pre-processing and modelling .17
8.3.4 Model query .17
8.4 Bias .17
8.5 Unpredictability .17
8.6 Opaqueness .18
8.7 Challenges related to the specification of AI systems .18
8.8 Challenges related to the implementation of AI systems .19
8.8.1 Data acquisition and preparation.19
8.8.2 Modelling .19
8.8.3 Model updates .21
8.8.4 Software defects .21
8.9 Challenges related to the use of AI systems .21
8.9.1 Human-computer interaction (HCI) factors .21
8.9.2 Misapplication of AI systems that demonstrate realistic human behaviour.22
8.10 System hardware faults .22
9 Mitigation measures .23
9.1 General .23
9.2 Transparency .23
9.3 Explainability .24
9.3.1 General.24
© ISO/IEC 2020 – All rights reserved iii

---------------------- Page: 3 ----------------------
ISO/IEC TR 24028:2020(E)

9.3.2 Aims of explanation .24
9.3.3 Ex-ante vs ex-post explanation .24
9.3.4 Approaches to explainability .25
9.3.5 Modes of ex-post explanation .25
9.3.6 Levels of explainability .26
9.3.7 Evaluation of the explanations .27
9.4 Controllability .27
9.4.1 General.27
9.4.2 Human-in-the-loop control points .28
9.5 Strategies for reducing bias .28
9.6 Privacy .28
9.7 Reliability, resilience and robustness .28
9.8 Mitigating system hardware faults .29
9.9 Functional safety .29
9.10 Testing and evaluation .30
9.10.1 General.30
9.10.2 Software validation and verification methods .30
9.10.3 Robustness considerations .32
9.10.4 Privacy-related considerations .33
9.10.5 System predictability considerations.33
9.11 Use and applicability .34
9.11.1 Compliance .34
9.11.2 Managing expectations .34
9.11.3 Product labelling .34
9.11.4 Cognitive science research .34
10 Conclusions .34
Annex A (informative) Related work on societal issues .36
Bibliography .37
iv © ISO/IEC 2020 – All rights reserved

---------------------- Page: 4 ----------------------
ISO/IEC TR 24028:2020(E)

Foreword
ISO (the International Organization for Standardization) and IEC (the International Electrotechnical
Commission) form the specialized system for worldwide standardization. National bodies that
are members of ISO or IEC participate in the development of International Standards through
technical committees established by the respective organization to deal with particular fields of
technical activity. ISO and IEC technical committees collaborate in fields of mutual interest. Other
international organizations, governmental and non-governmental, in liaison with ISO and IEC, also
take part in the work.
The procedures used to develop this document and those intended for its further maintenance are
described in the ISO/IEC Directives, Part 1. In particular, the different approval criteria needed for
the different types of document should be noted. This document was drafted in accordance with the
editorial rules of the ISO/IEC Directives, Part 2 (see www .iso .org/ directives).
Attention is drawn to the possibility that some of the elements of this document may be the subject
of patent rights. ISO and IEC shall not be held responsible for identifying any or all such patent
rights. Details of any patent rights identified during the development of the document will be in the
Introduction and/or on the ISO list of patent declarations received (see www .iso .org/ patents) or the IEC
list of patent declarations received (see http:// patents .iec .ch).
Any trade name used in this document is information given for the convenience of users and does not
constitute an endorsement.
For an explanation of the voluntary nature of standards, the meaning of ISO specific terms and
expressions related to conformity assessment, as well as information about ISO's adherence to the
World Trade Organization (WTO) principles in the Technical Barriers to Trade (TBT), see www .iso .org/
iso/ foreword .html.
This document was prepared by Joint Technical Committee ISO/IEC JTC 1, Information Technology,
Subcommittee SC 42, Artificial Intelligence.
Any feedback or questions on this document should be directed to the user’s national standards body. A
complete listing of these bodies can be found at www .iso .org/ members .html.
© ISO/IEC 2020 – All rights reserved v

---------------------- Page: 5 ----------------------
ISO/IEC TR 24028:2020(E)

Introduction
The goal of this document is to analyse the factors that can impact the trustworthiness of systems
providing or using AI, called hereafter artificial intelligence (AI) systems. The document briefly
surveys the existing approaches that can support or improve trustworthiness in technical systems
and discusses their potential application to AI systems. The document discusses possible approaches
to mitigating AI system vulnerabilities that relate to trustworthiness. The document also discusses
approaches to improving the trustworthiness of AI systems.
vi © ISO/IEC 2020 – All rights reserved

---------------------- Page: 6 ----------------------
TECHNICAL REPORT ISO/IEC TR 24028:2020(E)
Information technology — Artificial intelligence —
Overview of trustworthiness in artificial intelligence
1 Scope
This document surveys topics related to trustworthiness in AI systems, including the following:
— approaches to establish trust in AI systems through transparency, explainability, controllability, etc.;
— engineering pitfalls and typical associated threats and risks to AI systems, along with possible
mitigation techniques and methods; and
— approaches to assess and achieve availability, resiliency, reliability, accuracy, safety, security and
privacy of AI systems.
The specification of levels of trustworthiness for AI systems is out of the scope of this document.
2 Normative references
There are no normative references in this document.
3 Terms and definitions
For the purposes of this document, the following terms and definitions apply.
ISO and IEC maintain terminological databases for use in standardization at the following addresses:
— ISO Online browsing platform: available at https:// www .iso .org/ obp
— IEC Electropedia: available at http:// www .electropedia .org/
3.1
accountability
property that ensures that the actions of an entity (3.16) may be traced uniquely to that entity
[SOURCE: ISO/IEC 2382:2015, 2126250, modified — The Notes to entry have been removed.]
3.2
actor
entity (3.16) that communicates and interacts
[SOURCE: ISO/IEC TR 22417:2017, 3.1]
3.3
algorithm
set of rules for transforming the logical representation of data (3.11)
[SOURCE: ISO/IEC 11557:1992, 4.3]
3.4
artificial intelligence
AI
capability of an engineered system (3.38) to acquire, process and apply knowledge and skills
Note 1 to entry: Knowledge are facts, information (3.20) and skills acquired through experience or education.
© ISO/IEC 2020 – All rights reserved 1

---------------------- Page: 7 ----------------------
ISO/IEC TR 24028:2020(E)

3.5
asset
anything that has value (3.46) to a stakeholder (3.37)
Note 1 to entry: There are many types of assets, including:
a) information (3.20);
b) software, such as a computer program;
c) physical, such as computer;
d) services;
e) people and their qualifications, skills and experience; and
f) intangibles, such as reputation and image.
[SOURCE: ISO/IEC 21827:2008, 3.4, modified — In the definition, “the organization” has been changed
to “a stakeholder”. Note 1 to entry has been removed.]
3.6
attribute
property or characteristic of an object that can be distinguished quantitatively or qualitatively by
human or automated means
[SOURCE: ISO/IEC/IEEE 15939:2017, 3.2]
3.7
autonomy
autonomous
characteristic of a system (3.38) governed by its own rules as the result of self-learning
Note 1 to entry: Such systems are not subject to external control (3.10) or oversight.
3.8
bias
favouritism towards some things, people or groups over others
3.9
consistency
degree of uniformity, standardization and freedom from contradiction among the documents or parts
of a system (3.38) or component
[SOURCE: ISO/IEC 21827:2008, 3.14]
3.10
control
purposeful action on or in a process (3.29) to meet specified objectives
[SOURCE: IEC 61800-7-1:2015, 3.2.6]
3.11
data
re-interpretable representation of information (3.20) in a formalized manner suitable for
communication, interpretation or processing
Note 1 to entry: Data (3.11) can be processed by human or automatic means.
[SOURCE: ISO/IEC 2382:2015, 2121272, modified — Notes 2 and 3 to entry have been removed.]
2 © ISO/IEC 2020 – All rights reserved

---------------------- Page: 8 ----------------------
ISO/IEC TR 24028:2020(E)

3.12
data subject
individual about whom personal data (3.27) are recorded
[SOURCE: ISO 5127:2017, 3.13.4.01, modified — Note 1 to entry has been removed.]
3.13
decision tree
supervised-learning model for which inference can be represented by traversing one or more tree-like
structures
3.14
effectiveness
extent to which planned activities are realized and planned results achieved
[SOURCE: ISO 9000:2015, 3.7.11, modified — Note 1 to entry has been removed.]
3.15
efficiency
relationship between the results achieved and the resources used
[SOURCE: ISO 9000:2015, 3.7.10]
3.16
entity
any concrete or abstract thing of interest
[SOURCE: ISO/IEC 10746-2:2009, 6.1]
3.17
harm
injury or damage to the health of people or damage to property or the environment
[SOURCE: ISO/IEC Guide 51:2014, 3.1]
3.18
hazard
potential source of harm (3.17)
[SOURCE: ISO/IEC Guide 51:2014, 3.2]
3.19
human factors
environmental, organizational and job factors, in conjunction with cognitive human characteristics,
which influence the behaviour of persons or organizations
3.20
information
meaningful data (3.11)
[SOURCE: ISO 9000:2015, 3.8.2]
3.21
integrity
property of protecting the accuracy and completeness of assets (3.5)
[SOURCE: ISO/IEC 27000:2018, 3.36, modified — In the definition, "protecting the" has been added
before "accuracy" and "of assets" has been added after "completeness".]
© ISO/IEC 2020 – All rights reserved 3

---------------------- Page: 9 ----------------------
ISO/IEC TR 24028:2020(E)

3.22
intended use
use in accordance with information (3.20) provided with a product or system (3.38) or, in the absence of
such information, by generally understood patterns (3.26) of usage.
[SOURCE: ISO/IEC Guide 51:2014, 3.6]
3.23
machine learning
ML
process (3.29) by which a functional unit improves its performance by acquiring new knowledge or
skills or by reorganizing existing knowledge or skills
[SOURCE: ISO/IEC 2382:2015, 2123789]
3.24
machine learning model
mathematical construct that generates an inference or prediction, based on input data (3.11)
3.25
neural network
computational model utilizing distributed, parallel local processing and consisting of a network of
simple processing elements called artificial neurons, which can exhibit complex global behaviour
[SOURCE: ISO 18115-1:2013, 8.1]
3.26
pattern
set of features and their relationships used to recognize an entity (3.16) within a given context
[SOURCE: ISO/IEC 2382:2015, 2123798]
3.27
personal data
data (3.11) relating to an identified or identifiable individual
[SOURCE: ISO 5127:2017, 3.1.10.14, modified — The admitted terms and Notes 1 and 2 to entry have
been removed.]
3.28
privacy
freedom from intrusion into the private life or affairs of an individual when that intrusion results from
undue or illegal gathering and use of data (3.11) about that individual
[SOURCE: ISO/IEC 2382:2015, 2126263, modified — Notes 1 and 2 to entry have been removed.]
3.29
process
set of interrelated or interacting activities that use inputs to deliver an intended result
[SOURCE: ISO 9000:2015, 3.4.1, modified — The notes to entry have been omitted.]
3.30
reliability
property of consistent intended behaviour and results
[SOURCE: ISO/IEC 27000:2018, 3.55]
4 © ISO/IEC 2020 – All rights reserved

---------------------- Page: 10 ----------------------
ISO/IEC TR 24028:2020(E)

3.31
risk
effect of uncertainty on objectives
Note 1 to entry: An effect is a deviation from the expected. It can be positive, negative or both and can address,
create or result in opportunities and threats (3.39).
Note 2 to entry: Objectives can have different aspects and categories and can be applied at different levels.
Note 3 to entry: Risk is usually expressed in terms of risk sources, potential events, their consequences and their
likelihood.
[SOURCE: ISO 31000:2018, 3.1]
3.32
robot
programmed actuated mechanism with a degree of autonomy (3.7), moving within its environment, to
perform intended tasks
Note 1 to entry: A robot includes the control (3.10) system and interface of the control system (3.38).
Note 2 to entry: The classification of robot into industrial robot or service robot is done according to its intended
application.
[SOURCE: ISO 18646-2:2019, 3.1]
3.33
robotics
science and practice of designing, manufacturing and applying robots (3.32)
[SOURCE: ISO 8373:2012, 2.16]
3.34
safety
freedom from risk (3.31) which is not tolerable
[SOURCE: ISO/IEC Guide 51:2014, 3.14]
3.35
security
degree to which a product or system (3.38) protects information (3.20) and data (3.11) so that persons
or other products or systems have the degree of data access appropriate to their types and levels of
authorization
[SOURCE: ISO/IEC 25010:2011, 4.2.6]
3.36
sensitive data
data (3.11) with potentially harmful effects in the event of disclosure or misuse
[SOURCE: ISO 5127:2017, 3.1.10.16]
3.37
stakeholder
any individual, group or organization that can affect, be affected by or perceive itself to be affected by
a decision or activity
[SOURCE: ISO/IEC 38500:2015, 2.24]
© ISO/IEC 2020 – All rights reserved 5

---------------------- Page: 11 ----------------------
ISO/IEC TR 24028:2020(E)

3.38
system
combination of interacting elements organized to achieve one or more stated purposes
Note 1 to entry: A system is sometimes considered as a product or as the services it provides.
[SOURCE: ISO/IEC/IEEE 15288:2015, 3.38]
3.39
threat
potential cause of an unwanted incident, which may result in harm (3.17) to systems (3.38), organizations
or individuals
3.40
training
process (3.29) to establish or to improve the parameters of a machine learning model (3.24) based on a
machine learning algorithm (3.3) by using training data (3.11)
3.41
trust
degree to which a user (3.43) or other stakeholder (3.37) has confidence that a product or system (3.38)
will behave as intended
[SOURCE: ISO/IEC 25010:2011, 4.1.3.2]
3.42
trustworthiness
ability to meet stakeholders' (3.37) expectations in a verifiable way
Note 1 to entry: Depending on the context or sector and also on the specific product or service, data (3.11) and
technology used, different characteristics apply and need verification (3.47) to ensure stakeholders expectations
are met.
Note 2 to entry: Characteristics of trustworthiness include, for instance, reliability (3.30), availability, resilience,
security (3.35), privacy (3.28), safety (3.34), accountability (3.1), transparency, integrity (3.21), authenticity,
quality, usability.
Note 3 to entry: Trustworthiness is an attribute (3.6) that can be applied to services, products, technology, data
and information (3.20) as well as, in the context of governance, to organizations.
3.43
user
individual or group that interacts with a system (3.38) or benefits from a system during its utilization
[SOURCE: ISO/IEC/IEEE 15288:2015, 4.1.52, modified — Note 1 to entry has been removed.]
3.44
validation
confirmation, through the provision of objective evidence, that the requirements for a specific intended
use (3.22) or application have been fulfilled
Note 1 to entry: The right system (3.38) was
...

TECHNICAL ISO/IEC TR
REPORT 24028
First edition
Information technology —
Artificial intelligence — Overview
of trustworthiness in artificial
intelligence
Technologies de l'information — Intelligence artificielle — Examen
d'ensemble de la fiabilité en matière d'intelligence artificielle
PROOF/ÉPREUVE
Reference number
ISO/IEC TR 24028:2020(E)
©
ISO/IEC 2020

---------------------- Page: 1 ----------------------
ISO/IEC TR 24028:2020(E)

COPYRIGHT PROTECTED DOCUMENT
© ISO/IEC 2020
All rights reserved. Unless otherwise specified, or required in the context of its implementation, no part of this publication may
be reproduced or utilized otherwise in any form or by any means, electronic or mechanical, including photocopying, or posting
on the internet or an intranet, without prior written permission. Permission can be requested from either ISO at the address
below or ISO’s member body in the country of the requester.
ISO copyright office
CP 401 • Ch. de Blandonnet 8
CH-1214 Vernier, Geneva
Phone: +41 22 749 01 11
Fax: +41 22 749 09 47
Email: copyright@iso.org
Website: www.iso.org
Published in Switzerland
ii PROOF/ÉPREUVE © ISO/IEC 2020 – All rights reserved

---------------------- Page: 2 ----------------------
ISO/IEC TR 24028:2020(E)

Contents Page
Foreword .v
Introduction .vi
1 Scope . 1
2 Normative references . 1
3 Terms and definitions . 1
4 Overview . 7
5 Existing frameworks applicable to trustworthiness . 7
5.1 Background . 7
5.2 Recognition of layers of trust. 8
5.3 Application of software and data quality standards . 8
5.4 Application of risk management.10
5.5 Hardware-assisted approaches .10
6 Stakeholders .11
6.1 General concepts .11
6.2 Types.12
6.3 Assets .12
6.4 Values .13
7 Recognition of high-level concerns .13
7.1 Responsibility, accountability and governance .13
7.2 Safety .14
8 Vulnerabilities, threats and challenges .14
8.1 General .14
8.2 AI specific security threats .15
8.2.1 General.15
8.2.2 Data poisoning .15
8.2.3 Adversarial attacks.15
8.2.4 Model stealing .16
8.2.5 Hardware-focused threats to confidentiality and integrity .16
8.3 AI specific privacy threats .16
8.3.1 General.16
8.3.2 Data acquisition .16
8.3.3 Data pre-processing and modelling .17
8.3.4 Model query .17
8.4 Bias .17
8.5 Unpredictability .17
8.6 Opaqueness .18
8.7 Challenges related to the specification of AI systems .18
8.8 Challenges related to the implementation of AI systems .19
8.8.1 Data acquisition and preparation.19
8.8.2 Modelling .19
8.8.3 Model updates .21
8.8.4 Software defects .21
8.9 Challenges related to the use of AI systems .21
8.9.1 Human-computer interaction (HCI) factors .21
8.9.2 Misapplication of AI systems that demonstrate realistic human behaviour.22
8.10 System hardware faults .22
9 Mitigation measures .23
9.1 General .23
9.2 Transparency .23
9.3 Explainability .24
9.3.1 General.24
© ISO/IEC 2020 – All rights reserved PROOF/ÉPREUVE iii

---------------------- Page: 3 ----------------------
ISO/IEC TR 24028:2020(E)

9.3.2 Aims of explanation .24
9.3.3 Ex-ante vs ex-post explanation .24
9.3.4 Approaches to explainability .25
9.3.5 Modes of ex-post explanation .25
9.3.6 Levels of explainability .26
9.3.7 Evaluation of the explanations .27
9.4 Controllability .27
9.4.1 General.27
9.4.2 Human-in-the-loop control points .28
9.5 Strategies for reducing bias .28
9.6 Privacy .28
9.7 Reliability, resilience and robustness .28
9.8 Mitigating system hardware faults .29
9.9 Functional safety .29
9.10 Testing and evaluation .30
9.10.1 General.30
9.10.2 Software validation and verification methods .30
9.10.3 Robustness considerations .32
9.10.4 Privacy-related considerations .33
9.10.5 System predictability considerations.33
9.11 Use and applicability .34
9.11.1 Compliance .34
9.11.2 Managing expectations .34
9.11.3 Product labelling .34
9.11.4 Cognitive science research .34
10 Conclusions .34
Annex A (informative) Related work on societal issues .36
Bibliography .37
iv PROOF/ÉPREUVE © ISO/IEC 2020 – All rights reserved

---------------------- Page: 4 ----------------------
ISO/IEC TR 24028:2020(E)

Foreword
ISO (the International Organization for Standardization) and IEC (the International Electrotechnical
Commission) form the specialized system for worldwide standardization. National bodies that
are members of ISO or IEC participate in the development of International Standards through
technical committees established by the respective organization to deal with particular fields of
technical activity. ISO and IEC technical committees collaborate in fields of mutual interest. Other
international organizations, governmental and non-governmental, in liaison with ISO and IEC, also
take part in the work.
The procedures used to develop this document and those intended for its further maintenance are
described in the ISO/IEC Directives, Part 1. In particular, the different approval criteria needed for
the different types of document should be noted. This document was drafted in accordance with the
editorial rules of the ISO/IEC Directives, Part 2 (see www .iso .org/ directives).
Attention is drawn to the possibility that some of the elements of this document may be the subject
of patent rights. ISO and IEC shall not be held responsible for identifying any or all such patent
rights. Details of any patent rights identified during the development of the document will be in the
Introduction and/or on the ISO list of patent declarations received (see www .iso .org/ patents) or the IEC
list of patent declarations received (see http:// patents .iec .ch).
Any trade name used in this document is information given for the convenience of users and does not
constitute an endorsement.
For an explanation of the voluntary nature of standards, the meaning of ISO specific terms and
expressions related to conformity assessment, as well as information about ISO's adherence to the
World Trade Organization (WTO) principles in the Technical Barriers to Trade (TBT), see www .iso .org/
iso/ foreword .html.
This document was prepared by Joint Technical Committee ISO/IEC JTC 1, Information Technology,
Subcommittee SC 42, Artificial Intelligence.
Any feedback or questions on this document should be directed to the user’s national standards body. A
complete listing of these bodies can be found at www .iso .org/ members .html.
© ISO/IEC 2020 – All rights reserved PROOF/ÉPREUVE v

---------------------- Page: 5 ----------------------
ISO/IEC TR 24028:2020(E)

Introduction
The goal of this document is to analyse the factors that can impact the trustworthiness of systems
providing or using AI, called hereafter artificial intelligence (AI) systems. The document briefly
surveys the existing approaches that can support or improve trustworthiness in technical systems
and discusses their potential application to AI systems. The document discusses possible approaches
to mitigating AI system vulnerabilities that relate to trustworthiness. The document also discusses
approaches to improving the trustworthiness of AI systems.
vi PROOF/ÉPREUVE © ISO/IEC 2020 – All rights reserved

---------------------- Page: 6 ----------------------
TECHNICAL REPORT ISO/IEC TR 24028:2020(E)
Information technology — Artificial intelligence —
Overview of trustworthiness in artificial intelligence
1 Scope
This document surveys topics related to trustworthiness in AI systems, including the following:
— approaches to establish trust in AI systems through transparency, explainability, controllability, etc.;
— engineering pitfalls and typical associated threats and risks to AI systems, along with possible
mitigation techniques and methods; and
— approaches to assess and achieve availability, resiliency, reliability, accuracy, safety, security and
privacy of AI systems.
The specification of levels of trustworthiness for AI systems is out of the scope of this document.
2 Normative references
There are no normative references in this document.
3 Terms and definitions
For the purposes of this document, the following terms and definitions apply.
ISO and IEC maintain terminological databases for use in standardization at the following addresses:
— ISO Online browsing platform: available at https:// www .iso .org/ obp
— IEC Electropedia: available at http:// www .electropedia .org/
3.1
accountability
property that ensures that the actions of an entity (3.16) may be traced uniquely to that entity
[SOURCE: ISO/IEC 2382:2015, 2126250, modified — The Notes to entry have been removed.]
3.2
actor
entity (3.16) that communicates and interacts
[SOURCE: ISO/IEC TR 22417:2017, 3.1]
3.3
algorithm
set of rules for transforming the logical representation of data (3.11)
[SOURCE: ISO/IEC 11557:1992, 4.3]
3.4
artificial intelligence
AI
capability of an engineered system (3.38) to acquire, process and apply knowledge and skills
Note 1 to entry: Knowledge are facts, information (3.20) and skills acquired through experience or education.
© ISO/IEC 2020 – All rights reserved PROOF/ÉPREUVE 1

---------------------- Page: 7 ----------------------
ISO/IEC TR 24028:2020(E)

3.5
asset
anything that has value (3.46) to a stakeholder (3.37)
Note 1 to entry: There are many types of assets, including:
a) information (3.20);
b) software, such as a computer program;
c) physical, such as computer;
d) services;
e) people and their qualifications, skills and experience; and
f) intangibles, such as reputation and image.
[SOURCE: ISO/IEC 27000:2016, 2.4, modified - from “the organization” to “a stakeholder”]
3.6
attribute
property or characteristic of an object that can be distinguished quantitatively or qualitatively by
human or automated means
[SOURCE: ISO/IEC 15939:2016, 3.2]
3.7
autonomy
autonomous
characteristic of a system (3.38) governed by its own rules as the result of self-learning
Note 1 to entry: Such systems are not subject to external control (3.10) or oversight.
3.8
bias
favouritism towards some things, people or groups over others
3.9
consistency
degree of uniformity, standardization and freedom from contradiction among the documents or parts
of a system (3.38) or component
[SOURCE: ISO/IEC 21827:2008, 3.14]
3.10
control
purposeful action on or in a process (3.29) to meet specified objectives
[SOURCE: IEC 61800-7-1:2015, 3.2.6]
3.11
data
re-interpretable representation of information (3.20) in a formalized manner suitable for
communication, interpretation or processing
Note 1 to entry: Data (3.11) can be processed by human or automatic means.
[SOURCE: ISO/IEC 2382:2015, 2121272, modified — Notes 2 and 3 to entry have been removed.]
2 PROOF/ÉPREUVE © ISO/IEC 2020 – All rights reserved

---------------------- Page: 8 ----------------------
ISO/IEC TR 24028:2020(E)

3.12
data subject
individual about whom personal data (3.27) are recorded
[SOURCE: ISO 5127:2017, 3.13.4.01, modified — Note 1 to entry has been removed.]
3.13
decision tree
supervized-learning model for which inference can be represented by traversing one or more tree-like
structures
3.14
effectiveness
extent to which planned activities are realized and planned results achieved
[SOURCE: ISO 9000:2005, 3.2.14]
3.15
efficiency
relationship between the results achieved and the resources used
[SOURCE: ISO 9000:2005]
3.16
entity
any concrete or abstract thing of interest
[SOURCE: ISO/IEC 10746-2]
3.17
harm
injury or damage to the health of people or damage to property or the environment
[SOURCE: ISO/IEC Guide 51:2014, 3.1]
3.18
hazard
potential source of harm (3.17)
[SOURCE: ISO/IEC Guide 51:2014, 3.2]
3.19
human factors
environmental, organizational and job factors, in conjunction with cognitive human characteristics,
which influence the behaviour of persons or organizations
3.20
information
meaningful data (3.11)
[SOURCE: ISO 9000:2015, 3.8.2]
3.21
integrity
property of protecting the accuracy and completeness of assets (3.5)
[SOURCE: ISO/IEC 27000:2016, 2.36]
© ISO/IEC 2020 – All rights reserved PROOF/ÉPREUVE 3

---------------------- Page: 9 ----------------------
ISO/IEC TR 24028:2020(E)

3.22
intended use
use in accordance with information (3.20) provided with a product or system (3.38) or, in the absence of
such information, by generally understood patterns (3.26) of usage.
[SOURCE: ISO/IEC Guide 51:2014, 3.6]
3.23
machine learning
ML
process (3.29) by which a functional unit improves its performance by acquiring new knowledge or
skills or by reorganizing existing knowledge or skills
[SOURCE: ISO/IEC 2382:2015, 2123789]
3.24
machine learning model
mathematical construct that generates an inference or prediction, based on input data (3.11)
3.25
neural network
computational model utilizing distributed, parallel local processing and consisting of a network of
simple processing elements called artificial neurons, which can exhibit complex global behaviour
[SOURCE: ISO 18115-1:2013, 8.1]
3.26
pattern
set of features and their relationships used to recognize an entity (3.16) within a given context
[SOURCE: ISO/IEC 2382:2015, 2123798]
3.27
personal data
data (3.11) relating to an identified or identifiable individual
[SOURCE: ISO 5127:2017, 3.1.10.14, modified — The admitted terms and Notes 1 and 2 to entry have
been removed.]
3.28
privacy
freedom from intrusion into the private life or affairs of an individual when that intrusion results from
undue or illegal gathering and use of data (3.11) about that individual
[SOURCE: ISO/IEC 2382-8:1998, 08.01.23]
3.29
process
set of interrelated or interacting activities that use inputs to deliver an intended result
[SOURCE: ISO 9001:2015]
3.30
reliability
property of consistent intended behaviour and results
[SOURCE: ISO/IEC 27000:2016, 2.56]
4 PROOF/ÉPREUVE © ISO/IEC 2020 – All rights reserved

---------------------- Page: 10 ----------------------
ISO/IEC TR 24028:2020(E)

3.31
risk
effect of uncertainty on objectives
Note 1 to entry: An effect is a deviation from the expected. It can be positive, negative or both and can address,
create or result in opportunities and threats (3.39).
Note 2 to entry: Objectives can have different aspects and categories and can be applied at different levels.
Note 3 to entry: Risk is usually expressed in terms of risk sources, potential events, their consequences and their
likelihood.
[SOURCE: ISO 31000:2018, 3.1]
3.32
robot
programmed actuated mechanism with a degree of autonomy (3.7), moving within its environment, to
perform intended tasks
Note 1 to entry: A robot includes the control (3.10) system and interface of the control system (3.38).
Note 2 to entry: The classification of robot into industrial robot or service robot is done according to its intended
application.
[SOURCE: ISO 18646-2:2019, 3.1]
3.33
robotics
science and practice of designing, manufacturing and applying robots (3.32)
[SOURCE: ISO 8373:2012, 2.16]
3.34
safety
freedom from risk (3.31)which is not tolerable
[SOURCE: ISO/IEC Guide 51:2014]
3.35
security
degree to which a product or system (3.38) protects information (3.20) and data (3.11) so that persons
or other products or systems have the degree of data access appropriate to their types and levels of
authorization
[SOURCE: ISO/IEC 25010:2011]
3.36
sensitive data
data (3.11) with potentially harmful effects in the event of disclosure or misuse
[SOURCE: ISO 5127:2017, 3.1.10.16]
3.37
stakeholder
any individual, group or organization that can affect, be affected by or perceive itself to be affected by
a decision or activity
[SOURCE: ISO/IEC 38500:2015, 2.24]
© ISO/IEC 2020 – All rights reserved PROOF/ÉPREUVE 5

---------------------- Page: 11 ----------------------
ISO/IEC TR 24028:2020(E)

3.38
system
combination of interacting elements organized to achieve one or more stated purposes
Note 1 to entry: A system is sometimes considered as a product or as the services it provides.
[SOURCE: ISO/IEC/IEEE 15288:2015, 3.38]
3.39
threat
potential cause of an unwanted incident, which may result in harm (3.17) to systems (3.38) organizations
or individuals
3.40
training
process (3.29) to establish or to improve the parameters of a machine learning model (3.24) based on a
machine learning algorithm (3.3) by using training data (3.11)
3.41
trust
degree to which a user (3.43) or other stakeholder (3.37) has confidence that a product or system (3.38)
will behave as intended
[SOURCE: ISO/IEC 25010:2011, 4.1.3.2]
3.42
trustworthiness
ability to meet stakeholders' (3.37) expectations in a verifiable way
Note 1 to entry: Depending on the context or sector and also on the specific product or service, data (3.11) and
technology used, different characteristics apply and need verification (3.47) to ensure stakeholders expectations
are met.
Note 2 to entry: Characteristics of trustworthiness include, for instance, reliability (3.30), availability, resilience,
security (3.35), privacy (3.28), safety (3.34), accountability (3.1), transparency, integrity (3.21), authenticity,
quality, usability.
Note 3 to entry: Trustworthiness is an attribute (3.6) that can be applied to services, products, technology, data
and information (3.20) as well as, in the context of governance, to organizations.
3.43
user
individual or group that interacts with a system (3.38) or benefits from a system during its utilization
[SOURCE: ISO/IEC/IEEE 15288:2015]
3.44
validation
confirmation, through the provision of objective evidence, that the requirements for a specific intended
use (3.22) or application have been fulfilled
Note 1 to entry: The right system (3.38) was built.
[SOURCE: ISO/IEC TR 29110-1:2016, 3.73, modified — Only the last sentence of Note 1 to entry has been
retained and Note 2 to entry has been removed.]
3.45
value
unit of data (3.11)
[SOURCE: ISO/IEC 15939:2016, 3.41]
6 PROOF/ÉPREUVE © ISO/IEC 2020 – All rights reserved

---------------------- P
...

Questions, Comments and Discussion

Ask us and Technical Secretary will try to provide an answer. You can facilitate discussion about the standard in here.