ai explainability sharing by hong kong privacy commissioner · 4 source: report on risk analysis of...

12
52nd Asia Pacific Privacy Authorities Forum 2019: CIPL and TTC Labs Mini Design Jam on AI Explainability AI Explainability – Sharing by Hong Kong Privacy Commissioner 3 December 2019, Ocean Pavilion, Shangri-La Mactan, Cebu, Philippines Stephen Kai-yi WONG, Barrister Privacy Commissioner for Personal Data, Hong Kong, China

Upload: others

Post on 06-Oct-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

52nd Asia Pacific Privacy Authorities Forum 2019:

CIPL and TTC Labs Mini Design Jam on AI

Explainability

AI Explainability – Sharing by Hong Kong

Privacy Commissioner

3 December 2019, Ocean Pavilion, Shangri-La Mactan, Cebu,

Philippines Stephen Kai-yi WONG, Barrister Privacy Commissioner for Personal Data,

Hong Kong, China

Page 2: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

2

Transparency Personal Data Protection

Explainability

Importance of Explainability

• Part and parcel of personal data protection and transparency

• Avoid distrust of and grievance against organisations

• Respect human dignity • Ensure fairness

Page 3: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

3

Explainability:

Core of AI governance principles worldwide

• Ability to explain technical process of AI and human decisions

• Automated decisions be understood by humans and be traceable

• To enable the

affected individuals to understand and to challenge the outcomes

• To verify continuous alignment with individuals’ expectation

• To enable overall human control

• To ensure reliability, controllability and safety of AI

Explainability

ICDPPC - “Declaration on Ethics and Data

Protection in Artificial

Intelligence” (2018)

European Commission –

“Ethics Guidelines for Trustworthy AI”

(2019)

OECD – “Principles for Responsible Stewardship of Trustworthy AI”

(2019)

China – “Governance

Principles for AI” (2019)

Page 4: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

4

Source: Report on Risk Analysis of AI Ethics (April 2019)

China’s National AI Standardization Group:

Indicators of algorithmic ethics

• Disclose source code or operation rules of AI (provided that IP rights are not compromised)

Transparency

• Produce results without errors within specified timeframe

Reliability

• Be able to explain the reasons for certain results

Explicability

• Be able to reproduce the same results under specified conditions

Verifiability

Page 5: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

5

Source: Lilian Edwards & Michael Veale, Slave to the Algorithm? Why a “Right to an Explanation‘’is probably not the remedy you are looking for https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2972855

Provide more meaningful

explanation to individuals

Good for consistency and

procedural regularity of

decision-making

Expalainability of AI

Model-centric explainability

Disclose information about an AI model in general, e.g.:

- process of development

- logics involved

- training data used

Subject-centric explainability

Disclose about specific cases, e.g.:

- What changes in input data would affect the decision about me?

- What are the characteristics of individuals who receive similar treatment as me?

Page 6: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

Lighting up Hong Kong as a Smart City

6 Source: OGCIO, Hong Kong

Page 7: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

AI applications in Smart Lampposts

Panoramic cameras with AI • Collect real-time traffic data for sharing to the public as

well as traffic monitoring and incident management. Vehicle speed and types can be analyzed.

• Collect illegal dumping data and analyze patterns of illegal dumping activities. There will be auto-detection of illegal dumping activities.

7

Page 8: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

To be explainable! • Why and how the system generates a particular output or

decision? What’s the logic and rationale ?

• What is the process of machine learning? What combination of input factors contributed to the decision?

• Avoid “Black Box” algorithms that are opaque and complicated

• Explainability is crucial for building and maintaining the public’s trust in the AI system

8

Page 9: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

Proceed with caution: Regulation for AI explainability ?

• Not yet a standalone regulation in Hong Kong

• The idea of explainability is embedded in data ethics and accountability and reflected in the some industry.

• E.g. “Guidelines on Online Distribution and Advisory Platforms” by the Securities and Futures Commission, Hong Kong, July 2019

• E.g. Ethical Accountability Framework for AI (Draft), to be published by PCPD

9

Page 10: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

Guidelines on Online Distribution and Advisory Platforms

• The Guideline applies to: • Provision of financial advice in an online

environment using algorithms and other technology tools

• Uses of data and algorithms to profile clients and devise responses

• Requirements: • Information about the algorithm and its

limitations must be provided to clients • Internal controls in place to supervise

algorithm, prevent unauthorised access 10

Page 11: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

Building upon… • Ethical Accountability

Framework for Hong Kong, China – PCPD (2018) •Declaration on Ethics and

Data Protection in Artificial Intelligence – ICDPPC (2018) • Ethics Guidelines for

Trustworthy AI – EC (2019)

11

Page 12: AI Explainability Sharing by Hong Kong Privacy Commissioner · 4 Source: Report on Risk Analysis of AI Ethics (April 2019) China’s National AI Standardization Group: Indicators

Thank you

Stephen Kai-yi WONG, Barrister Privacy Commissioner for Personal Data,

Hong Kong, China