SlideShare une entreprise Scribd logo
1  sur  83
Télécharger pour lire hors ligne
© 2018 International Business Machines Corporation
IBM Research AI
Developing Trust in Artificial Intelligence
and Machine Learning for High-Stakes
Applications
Kush R. Varshney

krvarshn@us.ibm.com

http://krvarshney.github.io
© 2018 International Business Machines Corporation IBM Research AI
▪ Pneumonia patients with a history of asthma predicted to have lower mortality risk
than general population
© 2018 International Business Machines Corporation IBM Research AI
© 2018 International Business Machines Corporation IBM Research AI
© 2018 International Business Machines Corporation IBM Research AI
Machine Learning in High-Stakes Applications
▪ Machine learning algorithms have started influencing every part of our lives
– Health and wellness, law and order, commerce, entertainment, finance, human capital management,
communication, transportation, philanthropy, …
– Technological components of larger sociotechnical systems
▪ Many consequential decisions are being supported by machine learning models:
Credit Employment Admission Sentencing
© 2018 International Business Machines Corporation IBM Research AI
Trust
© 2018 International Business Machines Corporation IBM Research AI
Maslow’s Hierarchy of Needs
https://mrsomarali.files.wordpress.com/2015/10/maslow1.jpg
© 2018 International Business Machines Corporation IBM Research AI
Hierarchy of Needs for AI
https://mrsomarali.files.wordpress.com/2015/10/maslow1.jpg
basic accuracy and reliability
safety and security
transparency and trust
© 2018 International Business Machines Corporation IBM Research AI
• Medical diagnosis
• Prison sentencing
• Loan approval
• (Enterprise AI)
• Streaming services deciding on the
compression level of video packets
• Web portals deciding which news
story to show on top
• Phoneme classification within
speech transcription systems
• (Consumer AI)
Decision Sciences Data Products
Application Domains
© 2018 International Business Machines Corporation IBM Research AI
• High human cost of errors
• Uncertainty of training set being
representative of test set, and few
predictions made
• Safety and security issues
• Consideration of strategies for
achieving safety and security
beyond basic accuracy
• Quality of service has low human cost
• Large training set, large test set, ability
to explore the feature space, so little
epistemic uncertainty
• Not safety and security issues
• Focus squarely on basic accuracy
Decision Sciences Data Products
Application Domains
© 2018 International Business Machines Corporation IBM Research AI
• Self-driving cars
• Robotic surgery
• Autonomous weapons
• Humongous state space
• Difficult value alignment
Cyber-Physical Systems
Application Domains
© 2018 International Business Machines Corporation IBM Research AI
Some of Our Focus Areas – Enterprise AI vs. Consumer AI
▪ Transferability
▪ Learning + Reasoning
▪ Safety and Security
© 2018 International Business Machines Corporation IBM Research AI
Agenda
▪ Formal definition of safety in machine learning
▪ Strategies for achieving safety
▪ Details on interpretable machine learning
▪ Details on algorithmic fairness
– AI Fairness 360
▪ Transparency via supplier’s declarations of conformity (SDoCs)
▪ Conclusion
© 2018 International Business Machines Corporation IBM Research AI
K. R. Varshney and H. Alemzadeh, “On the Safety of Machine Learning:
Cyber-Physical Systems, Decision Sciences, and Data Products,” Big
Data, vol. 5, no. 3, p. 246–255, Sept. 2017.
© 2018 International Business Machines Corporation IBM Research AI
Safety
▪ Commonly used term across engineering disciplines connoting the absence of
failures or conditions that render a system dangerous (Ferrell, 2010)
– Safe food and water, safe vehicles and roads, safe medical treatments, safe toys, safe
neighborhoods, safe industrial plants, …
▪ Each domain has specific design principles and regulations applicable only to it
▪ Few works attempt a precise definition applicable broadly
▪ Definition based on harm, risk, and epistemic uncertainty (Möller, 2012)
© 2018 International Business Machines Corporation IBM Research AI
Harm
▪ A system yields an outcome based on its state and the inputs it receives
▪ The outcome event may be desired or undesired
▪ Outcomes have associated costs that can be measured and quantified by society
▪ An undesired outcome is a harm if its cost exceeds some threshold
▪ Unwanted events of small severity are not counted as safety issues
© 2018 International Business Machines Corporation IBM Research AI
Risk
▪ We do not know what the outcome will be, but its distribution is known and we can
calculate the expectation of its cost
▪ Risk is the expected value of the cost of harm
© 2018 International Business Machines Corporation IBM Research AI
Epistemic Uncertainty
▪ We still do not know what the outcome will be, but in contrast to risk, its probability
distribution is also unknown
▪ Epistemic uncertainty, in contrast to aleatoric uncertainty, results from lack of
knowledge that could be obtained in principle, but may be practically intractable to
gather
▪ Some decision theorists argue that all uncertainty can be captured probabilistically,
but we maintain the distinction between risk and uncertainty, following Möller
(2012)
© 2018 International Business Machines Corporation IBM Research AI
Safety
▪ Safety is the reduction or minimization of risk and epistemic uncertainty of harmful
events
▪ Costs have to be sufficiently high in some human sense for events to be harmful
▪ Safety involves reducing both the probability of expected harms and the possibility
of unexpected harms
© 2018 International Business Machines Corporation IBM Research AI
Risk Minimization in Statistical Machine Learning
▪ Risk minimization is the basis of statistical machine learning theory and practice
– Features and labels with probability density
– Function mapping
– Loss function
– Find to minimize risk
▪ Given i.i.d. training samples, not
– Empirical risk minimization
– converges to uniformly for all as goes to infinity (Glivenko-Cantelli)
▪ When is small, minimizing may not yield an with small
– Restrict complexity of based on some inductive bias (Vapnik, 1992)
© 2018 International Business Machines Corporation IBM Research AI
Epistemic Uncertainty in Machine Learning
▪ Risk minimization has many strengths but does not capture uncertainty
▪ Not always the case that training samples are drawn from true underlying
probability distribution of
– The distribution the samples come from cannot always be known
Precludes covariate shift and domain adaptation techniques
– Training on a data set from a different distribution can cause much harm
▪ Even when drawn from true distribution, training samples may be absent from
large parts of due to small probability density there
© 2018 International Business Machines Corporation IBM Research AI
Epistemic Uncertainty in Machine Learning
▪ Statistical learning theory analysis utilizes laws of large numbers to study the effect
of finite training data and the convergence of to
▪ In considering safety, should also be cognizant that in deployment, a machine
learning system only encounters a finite number of test samples
– The actual operational risk is an empirical quantity of the test set
– This operational risk may be much larger than the true risk for small cardinality test sets, even if is
risk-optimal
– This uncertainty caused by the instantiation of the test set can have large safety implications on
individual test samples
© 2018 International Business Machines Corporation IBM Research AI
Loss Functions in Machine Learning
▪ In risk minimization, the domain of the loss function is and the range is an abstract
quantity representing prediction error
▪ In real-world applications, the value of the loss function may be endowed with
some human cost
– That human cost may imply a loss function that also includes in the domain
▪ The cost may be severe enough to be harmful and thus a safety issue in some
parts of the domain and not in others
© 2018 International Business Machines Corporation IBM Research AI
Value Alignment
▪ Obtaining is non-trivial and context-dependent
▪ Eliciting the values of society
▪ Encapsulation of morals
© 2018 International Business Machines Corporation IBM Research AI
Agenda
▪ Formal definition of safety in machine learning
▪ Strategies for achieving safety
▪ Details on interpretable machine learning
▪ Details on algorithmic fairness
– AI Fairness 360
▪ Transparency via supplier’s declarations of conformity (SDoCs)
▪ Conclusion
© 2018 International Business Machines Corporation IBM Research AI
Four Categories of Strategies for Achieving Safety
▪ Möller and Hansson (2008) have identified four main
categories of approaches to achieve safety
1. Inherently safe design
– Exclusion of a potential hazard from the system, e.g. excluding
hydrogen from buoyant material of blimp
2. Safe fail
– System remains safe when it fails in its intended operation, e.g.
dead man’s switches
3. Safety reserves
– Safety factors and safety margins, e.g. vessel wall is designed
thicker than the largest pressure it will experience by some ratio
4. Procedural safeguards
– Audits, training, posted warnings
© 2018 International Business Machines Corporation IBM Research AI
Inherently Safe Design in Machine Learning
▪ Want robustness against the uncertainty of the training set not being sampled from the test
distribution
– The training set may have various quirks or biases that are unknown to the user and that will not be present
during the test phase
– Highly complex modeling techniques used today, including extreme gradient boosting and deep neural networks,
may pick up on those data vagaries in the learned models they produce to achieve high accuracy, but might fail
due to an unknown shift in the data domain
– The models are so complex that it is very difficult to understand how they will react to such shifts and whether
they will produce harmful outcomes as a result
▪ Interpretability of models and causality of predictors
– Extra constraints on
▪ May be performance loss in accuracy by doing so when measuring accuracy with a
common training and testing data probability distribution, but the reduction in epistemic
uncertainty by doing so increases safety
© 2018 International Business Machines Corporation IBM Research AI
Safe Fail in Machine Learning
▪ When predictions cannot be given confidently, a reject option is used
– A human operator intervenes and provides a manual prediction
▪ Models are reported to be least confident and in the reject regime near the
decision boundary in classification problems
– Implicitly assumed that distance from the boundary is inversely related to confidence
– Reasonable in parts of with high probability density and many training samples
▪ In parts of with low density and no training samples, decision boundary may be
completely based on inductive bias with much epistemic uncertainty
– Distance from boundary as a trigger for reject option is meaningless
▪ For a rare combination of features in a test sample, always elect manual
examination (Attenberg et al., 2015)
© 2018 International Business Machines Corporation IBM Research AI
Safety Reserves in Machine Learning (Uncertainty)
▪ Parametrize all uncertainty using
– Uncertainty in training set being from true distribution
Uncertain class priors, label noise, etc.
– Uncertainty in instantiation of test set
Do not care as much about average test error as maximum test error in certain applications
▪ Let the risk of the risk-optimal model if were known be
▪ Robust formulations find while constraining or minimizing either:
© 2018 International Business Machines Corporation IBM Research AI
Safety Reserves in Machine Learning (Loss Function)
▪ Fairness
– The risk of harm for members of protected groups should not be much worse than the risk of harm for
others
– Feature space partition contains unprivileged group members (e.g. defined by caste or gender)
– Feature space partition contains priviliged group members
▪ Disparate impact constraint (Feldman et al., 2015)
© 2018 International Business Machines Corporation IBM Research AI
Procedural Safeguards in Machine Learning
▪ Defining the training set and setting up evaluation procedures have certain
subtleties that can cause harm if done incorrectly
– User experience design can be used to guide and warn users and thereby increase safety
▪ Open source software and open data allows for the possibility of public audit
– Safety hazards can be discovered through public examination
© 2018 International Business Machines Corporation IBM Research AI
Agenda
▪ Formal definition of safety in machine learning
▪ Strategies for achieving safety
▪ Details on interpretable machine learning
▪ Details on algorithmic fairness
– AI Fairness 360
▪ Transparency via supplier’s declarations of conformity (SDoCs)
▪ Conclusion
© 2018 International Business Machines Corporation IBM Research AI
Three Dimensions of Explainability

One explanation does not fit all: There are many ways to explain things
directly interpretable vs. post hoc interpretation
global (model-level) vs. local (instance-level)
static vs. interactive (visual analytics)
The oldest AI formats, such as decision rule sets,
decision trees, and decision tables are simple enough
for people to understand. Supervised learning of these
models is directly interpretable.
Start with a black box model and probe into it with a
companion model to create interpretations. The black box
model continues to provide the actual prediction while
interpretation improve human interactions.
Show the entire predictive model to the user to help
them understand it (e.g. a small decision tree,
whether obtained directly or in a post hoc manner).
Only show the explanations associated with individual
predictions (i.e. what was it about the features of this
particular person that made her loan denied).
The interpretation is simply presented to the user. The user can interact with interpretation.
© 2018 International Business Machines Corporation IBM Research AI
D. M. Malioutov and K. R. Varshney, “Exact Rule Learning via Boolean
Compressed Sensing,” Int. Conf. Machine Learning, p. 765-773, Jun. 2013.
© 2018 International Business Machines Corporation IBM Research AI
Directly Interpretable Supervised Binary Classification
▪ Binary classification: special case of supervised learning when
▪ As discussed, different inductive biases lead to different ℋ
▪ The ℋ containing only sparse rule-based classifiers is interpretable (Freitas, 2014)
▪ Interpretability implies inherently safe design
© 2018 International Business Machines Corporation IBM Research AI
Rule-Based Classifiers
▪ Single rule is an AND clause of a few Boolean terms
▪ Predict Federer to defeat Murray if he:
– Wins more than 59% of 4 to 9 shot rallies AND
– Wins more than 78% of points when serving at 30-30 or Deuce AND
– Serves less than 20% of serves into the body
▪ Rule set is a collection of single rules
© 2018 International Business Machines Corporation IBM Research AI
Challenges of Rule Learning
▪ Finding compact decision rules involving few Boolean terms that best approximate
a given data set is an NP hard combinatorial optimization problem
▪ Most existing approaches for finding maximize criteria such as information gain,
support, confidence, lift, Gini impurity, etc.
– Decision trees, decision lists, RIPPER, SLIPPER, etc.
– Greedy heuristics with ad hoc pruning
▪ Renewed interest in rule learning driven by optimizing a principled objective, but
which retains interpretability (Rudin, 2015)
© 2018 International Business Machines Corporation IBM Research AI
Group Testing Problem
▪ Discover a sparse subset of faulty items in a large set of mostly good items
using a few pooled tests
– Blood screening of large groups of army recruits
– Computational biology
– Fault discovery in computer networks
▪ Mix together the blood of several recruits
– If test is negative, none of the recruits are diseased
– If test is positive, at least one of the recruits is diseased
– Logical OR operation
▪ Construct the pools in an intelligent way to require a small number of tests with
perfect recovery of diseased individuals
© 2018 International Business Machines Corporation IBM Research AI
Rule Learning as Group Testing
▪ Standard supervised binary classification problem
– with features and Boolean labels
▪ Construct individual Boolean clauses from features , for
– Months since promoted > 13
– Compensation plan == quota-based
– For continuous dimensions of X, make comparisons to set of thresholds
▪ Calculate the truth value of each Boolean term for each training sample to
construct an truth table matrix with entries
© 2018 International Business Machines Corporation IBM Research AI
Rule Learning as Group Testing (continued)
▪ The positive training samples are now equivalent to diseased pools of army
recruits
▪ Determine an Boolean coefficient vector that specifies which Boolean terms to
OR together in a decision rule to recover the positive samples
▪ Learn so that , where Boolean notation means:
▪ AND rules are more natural than OR rules for people
– DeMorgan’s laws
© 2018 International Business Machines Corporation IBM Research AI
Discussion
▪ Efficient optimization by relaxation with provable guarantees
▪ Achieve competitive accuracy and high interpretability
© 2018 International Business Machines Corporation IBM Research AI
Selected

Recent Explainability
Innovations from IBM
Research AI
Global, Directly Interpretable
Boolean Decision Rules via Column Generation
NIPS 2018 (accepted). Dash, Gunluk, Wei.
Global, Post-Hoc
Improving Simple Models with Confidence Profiles
NIPS 2018 (accepted). Dhurandhar, Shanmugam, Luss, Olsen.
Local, Directly Interpretable
Teaching Meaningful Explanations
AAAI 2019 (submitted). Codella, Hind, Ramamurthy, Campbell, Dhurandhar, Varshney, Wei, Mojsilovic.
Local, Post-Hoc
Explanations Based on the Missing: Towards Contrastive Explanations with Pertinent
Negatives
NIPS 2018 (accepted). Dhurandhar, Chen, Luss, Tu, Ting, Shanmugam, Das.
Interactive Model Visualization
Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models
IEEE VAST 2018 (accepted). Strobelt, Gehrmann, Behrisch, Perer, Pfister, Rush.
© 2018 International Business Machines Corporation IBM Research AI
Global, Post-Hoc
Improving Simple Models with Confidence Profiles
NIPS 2018. Dhurandhar, Shanmugam, Luss, Olsen.
Domain experts preference: SMEs many times want to use a model they are comfortable
with and trust.
Small Data settings: Small client data where complex neural networks might overfit so
simple models are preferred. However, possible to improve performance if we have a
pretrained network on a large public or private corpora belonging to the same feature space.
Resource Constrained Settings: In memory and power constrained settings only small
models can be deployed. Here improving small neural networks with the help of larger
neural network can be useful.
Challenge: Transfer information from complex models with higher accuracy to the decision
trees in a manner that enhances performance and adds valuable insight.
Method: Find area under curve (AUC) based on confidence scores at each probe for each
example based on complex neural network. Weight each training example by corresponding
AUC (or optimal weighting procedure) and retrain simple model such as a decision tree.
© 2018 International Business Machines Corporation IBM Research AI
Local, Directly Interpretable





Teaching Meaningful Explanations
AAAI 2019 (submitted). Codella, Hind, Ramamurthy, Campbell, Dhurandhar, Varshney, Wei, Mojsilovic.
• Ask SME to provide explanations (e) with their training
data (x, y, e)
• Train ML algorithm on features to jointly predict label (y)
and explanations (e)
• Apply classifier on new features inputs to produce label
(y) and explanation (e)
© 2018 International Business Machines Corporation IBM Research AI
Local, Post-Hoc
Explanations Based on the Missing: Towards Contrastive Explanations with Pertinent Negatives
NIPS 2018. Dhurandhar, Chen, Luss, Tu, Ting, Shanmugam, Das.
3
• Current methods focus mostly on what is present
based on positive and negative relevance.
• Our method highlights what should be minimally
sufficiently present and necessarily absent.
© 2018 International Business Machines Corporation IBM Research AI
Interactive Model Visualization





Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models
IEEE VAST 2018. Strobelt, Gehrmann, Behrisch, Perer, Pfister, Rush.
© 2018 International Business Machines Corporation IBM Research AI
Agenda
▪ Formal definition of safety in machine learning
▪ Strategies for achieving safety
▪ Details on interpretable machine learning
▪ Details on algorithmic fairness
– AI Fairness 360
▪ Transparency via supplier’s declarations of conformity (SDoCs)
▪ Conclusion
© 2018 International Business Machines Corporation IBM Research AI
Unwanted bias and algorithmic fairness

Machine learning, by its very nature, is always a form of statistical discrimination
Discrimination becomes
objectionable when it places certain
privileged groups at systematic
advantage and certain unprivileged
groups at systematic disadvantage
Illegal in certain contexts
© 2018 International Business Machines Corporation IBM Research AI
Unwanted bias and algorithmic fairness

Machine learning, by its very nature, is always a form of statistical discrimination
Unwanted bias in training data
yields models with unwanted bias
that scale out
Prejudice in labels
Undersampling or oversampling
© 2018 International Business Machines Corporation IBM Research AI
Fairness in building and deploying models
(d’Alessandro et al., 2017)
© 2018 International Business Machines Corporation IBM Research AI
Metrics, Algorithms
dataset
metric
pre-
processing
algorithm in-processing
algorithm
post-
processing
algorithm
classifier
metric
© 2018 International Business Machines Corporation IBM Research AI
Metrics, Algorithms, and Explainers
dataset
metric
pre-
processing
algorithm in-processing
algorithm
post-
processing
algorithm
classifier
metric
classifier
metric
explainer
dataset
metric
explainer
© 2018 International Business Machines Corporation IBM Research AI
21 (or more) definitions of fairness

Group and individual fairness
Disparate impact, statistical parity, equality of odds, equality of opportunity
There is no one definition of fairness applicable in all contexts
Some definitions even conflict
© 2018 International Business Machines Corporation IBM Research AI
Bias mitigation is not easy

Cannot simply drop protected attributes because features are correlated with them

© 2018 International Business Machines Corporation IBM Research AI
(Hardt, 2017)
Research

Algorithmic fairness is one of the hottest topics in the ML/AI research community

© 2018 International Business Machines Corporation IBM Research AI
F. P. Calmon, D. Wei, B. Vinzamuri, K. N. Ramamurthy, and K. R. Varshney,
“Optimized Data Pre-Processing for Discrimination Prevention,” Advances in
Neural Information Processing Systems, Dec. 2017.
© 2018 International Business Machines Corporation
Pre-Processing for Flexible Discrimination Prevention
(Biased) Data Model Decision
Pre-processing
In-processing
Post-processing
Published Data
© 2018 International Business Machines Corporation IBM Research AI
© 2018 International Business Machines Corporation
Our Approach
▪ Given dataset containing
• – discriminatory (protected) variables
• – other (non-protected) variables
• – outcome
▪ Learn optimized (randomized) mapping
▪ Map both train and test data before model
Race Age Degree Priors Recid
Black >45 Misd. 0 N
White 25-45 Felony 0 N
Black <25 Felony 1-3 Y
White >45 Misd. 0 N
Race Age Degree Priors Recid
Black >45 Felony 0 N
White >45 Felony 0 N
Black <25 Felony 1-3 N
White >45 Misd. 0 N
© 2018 International Business Machines Corporation
Optimization Criteria
1. Group discrimination
Control dependence of
transformed outcome on
3. Utility preservation
Retain joint distribution so model
can still learn task
2. Individual distortion
Avoid large changes in individual
features
𝑥, 𝑦
^𝑥, ^𝑦
𝛿
𝑑1 𝑑2
© 2018 International Business Machines Corporation
Original COMPAS Dataset
re-commit crime
race
red = misdemeanor

blue = felony
© 2018 International Business Machines Corporation
Pre-Processed COMPAS Dataset
race
red = misdemeanor

blue = felony
re-commit crime
© 2018 International Business Machines Corporation
Interpretable Classifier Learned from Original Dataset
prior crimes
race
> 0
predicted not to 

re-offend
= 0
prior crimes age
Caucasian
African-American
predicted not to 

re-offend
predicted to 

re-offend
> 3 ≤ 3
predicted not to 

re-offend
predicted to 

re-offend
< 45 ≥ 45
© 2018 International Business Machines Corporation
Interpretable Classifier Learned from Pre-Processed Dataset
prior crimes
predicted to 

re-offend
> 3
predicted not to 

re-offend
≤ 3
© 2018 International Business Machines Corporation IBM Research AI
Balanced Accuracy vs. Fairness
For a classifier trained with original training data, at the
best classification rate, this measure is quite high. This
implies unfairness.
abs(1-disparate impact) must be close to zero (typically < 0.2) for classifier predictions to be fair.
For a classifier trained with transformed training data,
at the best classification rate, the measure drops
significantly, implying bias mitigation.
© 2018 International Business Machines Corporation IBM Research AI
P. Sattigeri, S. C. Hoffman, V. Chenthamarakshan, and K. R. Varshney,
“Fairness GAN,” AAAI 2019, submitted.
© 2018 International Business Machines Corporation
Our Approach
An auxiliary classifier generative adversarial network (AC-GAN) with one main twist: try
to predict the protected attribute as poorly as possible from the features
© 2018 International Business Machines Corporation
1. Joint Conditional Generative
Model
Learn a generative model that
matches the true distribution .
3. GAN Formulation
Posed as an adversarial learning
problem and solved using novel
GAN formulation
2. Label and Sensitive
Attribute Independence
Maximization
Make outcome and sensitive
variable () independent in the
learned distribution.
Discriminator Training:
Generator Training:
Fairness GAN
© 2018 International Business Machines Corporation IBM Research AI
Fairness GAN
© 2018 International Business Machines Corporation IBM Research AI
Fairness GAN
© 2018 International Business Machines Corporation IBM Research AI
Agenda
▪ Formal definition of safety in machine learning
▪ Strategies for achieving safety
▪ Details on interpretable machine learning
▪ Details on algorithmic fairness
– AI Fairness 360
▪ Transparency via supplier’s declarations of conformity (SDoCs)
▪ Conclusion
05/03/18 Facebook says it has a tool to detect bias in its artificial intelligence Quartz
05/25/18 Microsoft is creating an oracle for catching biased AI algorithms MIT Technology Review
05/31/18 Pymetrics open-sources Audit AI, an algorithm bias detection tool VentureBeat
06/07/18 Google Education Guide to Responsible AI Practices – Fairness Google
06/09/18 Accenture wants to beat unfair AI with a professional toolkit TechCrunch
Fairness Measures
Framework to test given algorithm on variety of datasets and fairness
metrics
https://github.com/megantosh/
fairness_measures_code
Fairness
Comparison
Extensible test-bed to facilitate direct comparisons of algorithms with
respect to fairness measures. Includes raw & preprocessed datasets
https://github.com/algofairness/
fairness-comparison
Themis-ML
Python library built on scikit-learn that implements fairness-aware
machine learning algorithms
https://github.com/cosmicBboy/themis-
ml
FairML
Looks at significance of model inputs to quantify prediction
dependence on inputs
https://github.com/adebayoj/fairml
Aequitas
Web audit tool as well as python lib. Generates bias report for given
model and dataset
https://github.com/dssg/aequitas
Fairtest
Tests for associations between algorithm outputs and protected
populations
https://github.com/columbia/fairtest
Themis
Takes a black-box decision-making procedure and designs test cases
automatically to explore where the procedure might be exhibiting
group-based or causal discrimination
https://github.com/LASER-UMASS/
Themis
Audit-AI
Python library built on top of scikit-learn with various statistical tests
for classification and regression tasks
https://github.com/pymetrics/audit-ai
© 2018 International Business Machines Corporation IBM Research AI
AI Fairness 360
 Differentiation
Datasets
Toolbox
Fairness metrics (30+)
Fairness metric explanations
Bias mitigation algorithms (9+)
Guidance
Industry-specific tutorials
Comprehensive bias mitigation toolbox (including
unique algorithms from IBM Research)
Several metrics and algorithms that have no available
implementations elsewhere
Extensible
Designed to translate new research from the lab to
industry practitioners

(e.g. scikit-learn’s fit/predict paradigm)
© 2018 International Business Machines Corporation IBM Research AI
© 2018 International Business Machines Corporation IBM Research AI
© 2018 International Business Machines Corporation IBM Research AI
Agenda
▪ Formal definition of safety in machine learning
▪ Strategies for achieving safety
▪ Details on interpretable machine learning
▪ Details on algorithmic fairness
– AI Fairness 360
▪ Transparency via supplier’s declarations of conformity (SDoCs)
▪ Conclusion
© 2018 International Business Machines Corporation IBM Research AI
Supplier’s Declarations of Conformity
▪ A common factsheet voluntarily released with products in many different industries
▪ Intended use, provenance, etc.
▪ Reliability, safety, and security test results
▪ Transparency for the final step of achieving trust
▪ Information asymmetry / market for lemons
© 2018 International Business Machines Corporation IBM Research AI
Supplier’s Declarations of Conformity
▪ A common factsheet voluntarily released with products in many different industries
▪ Intended use, provenance, etc.
▪ Reliability, safety, and security test results
▪ Transparency for the final step of achieving trust
▪ Information asymmetry / market for lemons
© 2018 International Business Machines Corporation IBM Research AI
Supplier’s Declarations of Conformity
▪ What is the intended use of the service output?
▪ What algorithms or techniques does this service implement?
▪ Which datasets was the service tested on? (Provide links to datasets that were used for testing, along with corresponding
datasheets.)
▪ Describe the testing methodology.
▪ Describe the test results.
▪ Are you aware of possible examples of bias, ethical issues, or other safety risks as a result of using the service?
▪ Are the service outputs explainable and/or interpretable?
▪ For each dataset used by the service: Was the dataset checked for bias? What efforts were made to ensure that it is fair and
representative?
▪ Does the service implement and perform any bias detection and remediation?
▪ What is the expected performance on unseen data or data with different distributions?
▪ Was the service checked for robustness against adversarial attacks?
▪ When were the models last updated?
© 2018 International Business Machines Corporation IBM Research AI
Agenda
▪ Formal definition of safety in machine learning
▪ Strategies for achieving safety
▪ Details on interpretable machine learning
▪ Details on algorithmic fairness
– AI Fairness 360
▪ Transparency via supplier’s declarations of conformity (SDoCs)
▪ Conclusion
© 2018 International Business Machines Corporation IBM Research AI
Summary and Conclusion
▪ Considerations beyond accuracy in enterprise AI
▪ Very basic definition of safety in terms of harm, risk, uncertainty
▪ Minimization of epistemic uncertainty is missing from standard modes of machine
learning developed around risk minimization
▪ Research agenda motivated by trust
– Risk minimization close to the limits of what is achievable
– Epistemic uncertainty minimization has new and exciting problems
– Some examples of strategies for increasing safety in machine learning, but far from complete
– Interpretability, fairness, SDoCs
– Open source toolbox: AI Fairness 360 (aif360)
© 2018 International Business Machines Corporation
http://aif360.mybluemix.net

https://github.com/ibm/aif360

https://pypi.org/project/aif360

Contenu connexe

Tendances

Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Krishnaram Kenthapadi
 
Fairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsFairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsKrishnaram Kenthapadi
 
Responsible AI
Responsible AIResponsible AI
Responsible AIAnand Rao
 
Fairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsFairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsKrishnaram Kenthapadi
 
Trusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open SourceTrusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open SourceAnimesh Singh
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedResponsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedKrishnaram Kenthapadi
 
Fairness in Machine Learning
Fairness in Machine LearningFairness in Machine Learning
Fairness in Machine LearningDelip Rao
 
Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)Krishnaram Kenthapadi
 
Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Krishnaram Kenthapadi
 
Ethics of Analytics and Machine Learning
Ethics of Analytics and Machine LearningEthics of Analytics and Machine Learning
Ethics of Analytics and Machine LearningMark Underwood
 
Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...
Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...
Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...Patrick Van Renterghem
 
"Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons...
"Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons..."Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons...
"Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons...Edge AI and Vision Alliance
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AIBill Liu
 
20190528 - Guidelines for Trustworthy AI
20190528 - Guidelines for Trustworthy AI20190528 - Guidelines for Trustworthy AI
20190528 - Guidelines for Trustworthy AIBrussels Legal Hackers
 
Frameworks for Algorithmic Bias
Frameworks for Algorithmic BiasFrameworks for Algorithmic Bias
Frameworks for Algorithmic BiasJonathan Stray
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedResponsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedKrishnaram Kenthapadi
 
Responsible AI: An Example AI Development Process with Focus on Risks and Con...
Responsible AI: An Example AI Development Process with Focus on Risks and Con...Responsible AI: An Example AI Development Process with Focus on Risks and Con...
Responsible AI: An Example AI Development Process with Focus on Risks and Con...Patrick Van Renterghem
 

Tendances (20)

Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
Responsible AI in Industry (Tutorials at AAAI 2021, FAccT 2021, and WWW 2021)
 
Fairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsFairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML Systems
 
Responsible AI
Responsible AIResponsible AI
Responsible AI
 
Fairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML SystemsFairness and Privacy in AI/ML Systems
Fairness and Privacy in AI/ML Systems
 
Trusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open SourceTrusted, Transparent and Fair AI using Open Source
Trusted, Transparent and Fair AI using Open Source
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedResponsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons Learned
 
Fairness in Machine Learning
Fairness in Machine LearningFairness in Machine Learning
Fairness in Machine Learning
 
Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)Explainable AI in Industry (KDD 2019 Tutorial)
Explainable AI in Industry (KDD 2019 Tutorial)
 
Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)
 
Ethics of Analytics and Machine Learning
Ethics of Analytics and Machine LearningEthics of Analytics and Machine Learning
Ethics of Analytics and Machine Learning
 
Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...
Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...
Fairness and Transparency: Algorithmic Explainability, some Legal and Ethical...
 
"Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons...
"Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons..."Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons...
"Practical Approaches to Training Data Strategy: Bias, Legal and Ethical Cons...
 
Amazon SageMaker Clarify
Amazon SageMaker ClarifyAmazon SageMaker Clarify
Amazon SageMaker Clarify
 
Implementing Ethics in AI
Implementing Ethics in AIImplementing Ethics in AI
Implementing Ethics in AI
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AI
 
20190528 - Guidelines for Trustworthy AI
20190528 - Guidelines for Trustworthy AI20190528 - Guidelines for Trustworthy AI
20190528 - Guidelines for Trustworthy AI
 
Frameworks for Algorithmic Bias
Frameworks for Algorithmic BiasFrameworks for Algorithmic Bias
Frameworks for Algorithmic Bias
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedResponsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons Learned
 
Responsible AI: An Example AI Development Process with Focus on Risks and Con...
Responsible AI: An Example AI Development Process with Focus on Risks and Con...Responsible AI: An Example AI Development Process with Focus on Risks and Con...
Responsible AI: An Example AI Development Process with Focus on Risks and Con...
 
Introduction to AI Governance
Introduction to AI GovernanceIntroduction to AI Governance
Introduction to AI Governance
 

Similaire à NYAI #24: Developing Trust in Artificial Intelligence and Machine Learning for High-Stakes Applications w/ Dr. Kush Varshney (IBM Research AI)

How Facility Controls Systems Present Cybersecurity Challenges - OSIsoft
How Facility Controls Systems Present Cybersecurity Challenges - OSIsoftHow Facility Controls Systems Present Cybersecurity Challenges - OSIsoft
How Facility Controls Systems Present Cybersecurity Challenges - OSIsoftOSIsoft, LLC
 
Smarter cyber security v8
Smarter cyber security v8Smarter cyber security v8
Smarter cyber security v8John Palfreyman
 
Evolution security controls towards Cloud Services
Evolution security controls towards Cloud ServicesEvolution security controls towards Cloud Services
Evolution security controls towards Cloud ServicesHugo Rodrigues
 
Building Risk Management into Enterprise Architecture
Building Risk Management into Enterprise ArchitectureBuilding Risk Management into Enterprise Architecture
Building Risk Management into Enterprise Architectureiasaglobal
 
Accenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber ResilienceAccenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber Resilienceaccenture
 
Accenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber ResilienceAccenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber Resilienceaccenture
 
Irv Badr: Managing Risk Safety and Security Compliance
Irv Badr: Managing Risk Safety and Security Compliance Irv Badr: Managing Risk Safety and Security Compliance
Irv Badr: Managing Risk Safety and Security Compliance EnergyTech2015
 
New technologies - Amer Haza'a
New technologies - Amer Haza'aNew technologies - Amer Haza'a
New technologies - Amer Haza'aFahmi Albaheth
 
Cyber security framework
Cyber security frameworkCyber security framework
Cyber security frameworkYann Lecourt
 
From checkboxes to frameworks
From checkboxes to frameworksFrom checkboxes to frameworks
From checkboxes to frameworksVincent Bellamy
 
Car Cybersecurity: The Gap Still Exists
Car Cybersecurity: The Gap Still ExistsCar Cybersecurity: The Gap Still Exists
Car Cybersecurity: The Gap Still ExistsSecurity Innovation
 
10 Security Essentials Every CxO Should Know
10 Security Essentials Every CxO Should Know10 Security Essentials Every CxO Should Know
10 Security Essentials Every CxO Should KnowIBM Security
 
Operationalizing Big Data Security Analytics - IANS Forum Toronto Keynote
Operationalizing Big Data Security Analytics - IANS Forum Toronto KeynoteOperationalizing Big Data Security Analytics - IANS Forum Toronto Keynote
Operationalizing Big Data Security Analytics - IANS Forum Toronto KeynoteInterset
 
Aalto cyber-10.4.18
Aalto cyber-10.4.18Aalto cyber-10.4.18
Aalto cyber-10.4.18japijapi
 
Cyber Threat Intelligence: Transforming Data into Relevant Intelligence
Cyber Threat Intelligence: Transforming Data into Relevant IntelligenceCyber Threat Intelligence: Transforming Data into Relevant Intelligence
Cyber Threat Intelligence: Transforming Data into Relevant IntelligenceEnterprise Management Associates
 
Operationalizing Big Data Security Analytics - IANS Forum Dallas
Operationalizing Big Data Security Analytics - IANS Forum DallasOperationalizing Big Data Security Analytics - IANS Forum Dallas
Operationalizing Big Data Security Analytics - IANS Forum DallasInterset
 
Bridging the Gap between Privacy and Security: Using Technology to Manage Com...
Bridging the Gap between Privacy and Security: Using Technology to Manage Com...Bridging the Gap between Privacy and Security: Using Technology to Manage Com...
Bridging the Gap between Privacy and Security: Using Technology to Manage Com...IBM Security
 

Similaire à NYAI #24: Developing Trust in Artificial Intelligence and Machine Learning for High-Stakes Applications w/ Dr. Kush Varshney (IBM Research AI) (20)

How Facility Controls Systems Present Cybersecurity Challenges - OSIsoft
How Facility Controls Systems Present Cybersecurity Challenges - OSIsoftHow Facility Controls Systems Present Cybersecurity Challenges - OSIsoft
How Facility Controls Systems Present Cybersecurity Challenges - OSIsoft
 
Infosec russia cnemeth_v1.2.ppt
Infosec russia cnemeth_v1.2.pptInfosec russia cnemeth_v1.2.ppt
Infosec russia cnemeth_v1.2.ppt
 
Smarter cyber security v8
Smarter cyber security v8Smarter cyber security v8
Smarter cyber security v8
 
Evolution security controls towards Cloud Services
Evolution security controls towards Cloud ServicesEvolution security controls towards Cloud Services
Evolution security controls towards Cloud Services
 
Building Risk Management into Enterprise Architecture
Building Risk Management into Enterprise ArchitectureBuilding Risk Management into Enterprise Architecture
Building Risk Management into Enterprise Architecture
 
Accenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber ResilienceAccenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber Resilience
 
Accenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber ResilienceAccenture Security CG&S Cyber Resilience
Accenture Security CG&S Cyber Resilience
 
Irv Badr: Managing Risk Safety and Security Compliance
Irv Badr: Managing Risk Safety and Security Compliance Irv Badr: Managing Risk Safety and Security Compliance
Irv Badr: Managing Risk Safety and Security Compliance
 
Cyber Risk in the Energy Industry
Cyber Risk in the Energy IndustryCyber Risk in the Energy Industry
Cyber Risk in the Energy Industry
 
New technologies - Amer Haza'a
New technologies - Amer Haza'aNew technologies - Amer Haza'a
New technologies - Amer Haza'a
 
Automotive Cybersecurity: The Gap Still Exists
Automotive Cybersecurity: The Gap Still ExistsAutomotive Cybersecurity: The Gap Still Exists
Automotive Cybersecurity: The Gap Still Exists
 
Cyber security framework
Cyber security frameworkCyber security framework
Cyber security framework
 
From checkboxes to frameworks
From checkboxes to frameworksFrom checkboxes to frameworks
From checkboxes to frameworks
 
Car Cybersecurity: The Gap Still Exists
Car Cybersecurity: The Gap Still ExistsCar Cybersecurity: The Gap Still Exists
Car Cybersecurity: The Gap Still Exists
 
10 Security Essentials Every CxO Should Know
10 Security Essentials Every CxO Should Know10 Security Essentials Every CxO Should Know
10 Security Essentials Every CxO Should Know
 
Operationalizing Big Data Security Analytics - IANS Forum Toronto Keynote
Operationalizing Big Data Security Analytics - IANS Forum Toronto KeynoteOperationalizing Big Data Security Analytics - IANS Forum Toronto Keynote
Operationalizing Big Data Security Analytics - IANS Forum Toronto Keynote
 
Aalto cyber-10.4.18
Aalto cyber-10.4.18Aalto cyber-10.4.18
Aalto cyber-10.4.18
 
Cyber Threat Intelligence: Transforming Data into Relevant Intelligence
Cyber Threat Intelligence: Transforming Data into Relevant IntelligenceCyber Threat Intelligence: Transforming Data into Relevant Intelligence
Cyber Threat Intelligence: Transforming Data into Relevant Intelligence
 
Operationalizing Big Data Security Analytics - IANS Forum Dallas
Operationalizing Big Data Security Analytics - IANS Forum DallasOperationalizing Big Data Security Analytics - IANS Forum Dallas
Operationalizing Big Data Security Analytics - IANS Forum Dallas
 
Bridging the Gap between Privacy and Security: Using Technology to Manage Com...
Bridging the Gap between Privacy and Security: Using Technology to Manage Com...Bridging the Gap between Privacy and Security: Using Technology to Manage Com...
Bridging the Gap between Privacy and Security: Using Technology to Manage Com...
 

Plus de Maryam Farooq

Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)
Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)
Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)Maryam Farooq
 
AI & COVID19: Ethics & Data Rights (NYAI x AISC)
AI & COVID19: Ethics & Data Rights (NYAI x AISC)AI & COVID19: Ethics & Data Rights (NYAI x AISC)
AI & COVID19: Ethics & Data Rights (NYAI x AISC)Maryam Farooq
 
NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...
NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...
NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...Maryam Farooq
 
NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...
NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...
NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...Maryam Farooq
 
NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...
NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...
NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...Maryam Farooq
 
NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)
NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)
NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)Maryam Farooq
 
NYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell Rebo
NYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell ReboNYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell Rebo
NYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell ReboMaryam Farooq
 
NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...
NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...
NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...Maryam Farooq
 
NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...
NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...
NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...Maryam Farooq
 
NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...
NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...
NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...Maryam Farooq
 
NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)
NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)
NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)Maryam Farooq
 
"Understanding Humans with Machines" (Arthur Tisi)
"Understanding Humans with Machines" (Arthur Tisi)"Understanding Humans with Machines" (Arthur Tisi)
"Understanding Humans with Machines" (Arthur Tisi)Maryam Farooq
 
NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...
NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...
NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...Maryam Farooq
 
NYAI #13: "AI and Business Transformation" - Josh Sutton
NYAI #13: "AI and Business Transformation" - Josh SuttonNYAI #13: "AI and Business Transformation" - Josh Sutton
NYAI #13: "AI and Business Transformation" - Josh SuttonMaryam Farooq
 
NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...
NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...
NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...Maryam Farooq
 

Plus de Maryam Farooq (15)

Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)
Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)
Women in AI Social: Fall Edition (NYAI x Aggregate Intellect x AI Geeks)
 
AI & COVID19: Ethics & Data Rights (NYAI x AISC)
AI & COVID19: Ethics & Data Rights (NYAI x AISC)AI & COVID19: Ethics & Data Rights (NYAI x AISC)
AI & COVID19: Ethics & Data Rights (NYAI x AISC)
 
NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...
NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...
NLP Community Conference - Dr. Catherine Havasi (ConceptNet/MIT Media Lab/Lum...
 
NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...
NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...
NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...
 
NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...
NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...
NYAI #26: Federated Learning: Machine Learning on Edge Devices w/ Alice Albre...
 
NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)
NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)
NYAI #23: Using Cognitive Neuroscience to Create AI (w/ Dr. Peter Olausson)
 
NYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell Rebo
NYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell ReboNYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell Rebo
NYAI #25: Evolution Strategies: An Alternative Approach to AI w/ Maxwell Rebo
 
NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...
NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...
NYAI #19: AI & UI - "AI + Emotion: It's all about Trust" by Steph Hay (VP Des...
 
NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...
NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...
NYAI #19: AI & UI - "Designing Intelligent Agents & a New Class of 'Perceived...
 
NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...
NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...
NYAI #18: Team Alignment for Human-Centered AI (Chris Butler - Director of AI...
 
NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)
NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)
NYAI #18: Designing for AI (Rob Strati & Jesse Schifano of ECHO)
 
"Understanding Humans with Machines" (Arthur Tisi)
"Understanding Humans with Machines" (Arthur Tisi)"Understanding Humans with Machines" (Arthur Tisi)
"Understanding Humans with Machines" (Arthur Tisi)
 
NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...
NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...
NYAI #13: "Designing AI by Learning from Enterprise" - Nicholas Borge (Impart...
 
NYAI #13: "AI and Business Transformation" - Josh Sutton
NYAI #13: "AI and Business Transformation" - Josh SuttonNYAI #13: "AI and Business Transformation" - Josh Sutton
NYAI #13: "AI and Business Transformation" - Josh Sutton
 
NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...
NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...
NYAI #10: Building an AI Autonomous Agent Using Supervised Learning with Denn...
 

Dernier

Principles and Practices of Data Visualization
Principles and Practices of Data VisualizationPrinciples and Practices of Data Visualization
Principles and Practices of Data VisualizationKianJazayeri1
 
Semantic Shed - Squashing and Squeezing.pptx
Semantic Shed - Squashing and Squeezing.pptxSemantic Shed - Squashing and Squeezing.pptx
Semantic Shed - Squashing and Squeezing.pptxMike Bennett
 
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...Amil Baba Dawood bangali
 
Cyber awareness ppt on the recorded data
Cyber awareness ppt on the recorded dataCyber awareness ppt on the recorded data
Cyber awareness ppt on the recorded dataTecnoIncentive
 
Networking Case Study prepared by teacher.pptx
Networking Case Study prepared by teacher.pptxNetworking Case Study prepared by teacher.pptx
Networking Case Study prepared by teacher.pptxHimangsuNath
 
Data Factory in Microsoft Fabric (MsBIP #82)
Data Factory in Microsoft Fabric (MsBIP #82)Data Factory in Microsoft Fabric (MsBIP #82)
Data Factory in Microsoft Fabric (MsBIP #82)Cathrine Wilhelmsen
 
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...Dr Arash Najmaei ( Phd., MBA, BSc)
 
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdfEnglish-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdfblazblazml
 
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...Boston Institute of Analytics
 
SMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptxSMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptxHaritikaChhatwal1
 
Bank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis ProjectBank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis ProjectBoston Institute of Analytics
 
What To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptxWhat To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptxSimranPal17
 
Learn How Data Science Changes Our World
Learn How Data Science Changes Our WorldLearn How Data Science Changes Our World
Learn How Data Science Changes Our WorldEduminds Learning
 
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Boston Institute of Analytics
 
Real-Time AI Streaming - AI Max Princeton
Real-Time AI  Streaming - AI Max PrincetonReal-Time AI  Streaming - AI Max Princeton
Real-Time AI Streaming - AI Max PrincetonTimothy Spann
 
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024Susanna-Assunta Sansone
 
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...Thomas Poetter
 
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptxThe Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptxTasha Penwell
 
Decoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis ProjectDecoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis ProjectBoston Institute of Analytics
 

Dernier (20)

Insurance Churn Prediction Data Analysis Project
Insurance Churn Prediction Data Analysis ProjectInsurance Churn Prediction Data Analysis Project
Insurance Churn Prediction Data Analysis Project
 
Principles and Practices of Data Visualization
Principles and Practices of Data VisualizationPrinciples and Practices of Data Visualization
Principles and Practices of Data Visualization
 
Semantic Shed - Squashing and Squeezing.pptx
Semantic Shed - Squashing and Squeezing.pptxSemantic Shed - Squashing and Squeezing.pptx
Semantic Shed - Squashing and Squeezing.pptx
 
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
 
Cyber awareness ppt on the recorded data
Cyber awareness ppt on the recorded dataCyber awareness ppt on the recorded data
Cyber awareness ppt on the recorded data
 
Networking Case Study prepared by teacher.pptx
Networking Case Study prepared by teacher.pptxNetworking Case Study prepared by teacher.pptx
Networking Case Study prepared by teacher.pptx
 
Data Factory in Microsoft Fabric (MsBIP #82)
Data Factory in Microsoft Fabric (MsBIP #82)Data Factory in Microsoft Fabric (MsBIP #82)
Data Factory in Microsoft Fabric (MsBIP #82)
 
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
6 Tips for Interpretable Topic Models _ by Nicha Ruchirawat _ Towards Data Sc...
 
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdfEnglish-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
English-8-Q4-W3-Synthesizing-Essential-Information-From-Various-Sources-1.pdf
 
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
 
SMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptxSMOTE and K-Fold Cross Validation-Presentation.pptx
SMOTE and K-Fold Cross Validation-Presentation.pptx
 
Bank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis ProjectBank Loan Approval Analysis: A Comprehensive Data Analysis Project
Bank Loan Approval Analysis: A Comprehensive Data Analysis Project
 
What To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptxWhat To Do For World Nature Conservation Day by Slidesgo.pptx
What To Do For World Nature Conservation Day by Slidesgo.pptx
 
Learn How Data Science Changes Our World
Learn How Data Science Changes Our WorldLearn How Data Science Changes Our World
Learn How Data Science Changes Our World
 
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
 
Real-Time AI Streaming - AI Max Princeton
Real-Time AI  Streaming - AI Max PrincetonReal-Time AI  Streaming - AI Max Princeton
Real-Time AI Streaming - AI Max Princeton
 
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
 
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
Minimizing AI Hallucinations/Confabulations and the Path towards AGI with Exa...
 
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptxThe Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
The Power of Data-Driven Storytelling_ Unveiling the Layers of Insight.pptx
 
Decoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis ProjectDecoding Patterns: Customer Churn Prediction Data Analysis Project
Decoding Patterns: Customer Churn Prediction Data Analysis Project
 

NYAI #24: Developing Trust in Artificial Intelligence and Machine Learning for High-Stakes Applications w/ Dr. Kush Varshney (IBM Research AI)

  • 1. © 2018 International Business Machines Corporation IBM Research AI Developing Trust in Artificial Intelligence and Machine Learning for High-Stakes Applications Kush R. Varshney
 krvarshn@us.ibm.com
 http://krvarshney.github.io
  • 2. © 2018 International Business Machines Corporation IBM Research AI ▪ Pneumonia patients with a history of asthma predicted to have lower mortality risk than general population
  • 3. © 2018 International Business Machines Corporation IBM Research AI
  • 4. © 2018 International Business Machines Corporation IBM Research AI
  • 5. © 2018 International Business Machines Corporation IBM Research AI Machine Learning in High-Stakes Applications ▪ Machine learning algorithms have started influencing every part of our lives – Health and wellness, law and order, commerce, entertainment, finance, human capital management, communication, transportation, philanthropy, … – Technological components of larger sociotechnical systems ▪ Many consequential decisions are being supported by machine learning models: Credit Employment Admission Sentencing
  • 6. © 2018 International Business Machines Corporation IBM Research AI Trust
  • 7. © 2018 International Business Machines Corporation IBM Research AI Maslow’s Hierarchy of Needs https://mrsomarali.files.wordpress.com/2015/10/maslow1.jpg
  • 8. © 2018 International Business Machines Corporation IBM Research AI Hierarchy of Needs for AI https://mrsomarali.files.wordpress.com/2015/10/maslow1.jpg basic accuracy and reliability safety and security transparency and trust
  • 9. © 2018 International Business Machines Corporation IBM Research AI • Medical diagnosis • Prison sentencing • Loan approval • (Enterprise AI) • Streaming services deciding on the compression level of video packets • Web portals deciding which news story to show on top • Phoneme classification within speech transcription systems • (Consumer AI) Decision Sciences Data Products Application Domains
  • 10. © 2018 International Business Machines Corporation IBM Research AI • High human cost of errors • Uncertainty of training set being representative of test set, and few predictions made • Safety and security issues • Consideration of strategies for achieving safety and security beyond basic accuracy • Quality of service has low human cost • Large training set, large test set, ability to explore the feature space, so little epistemic uncertainty • Not safety and security issues • Focus squarely on basic accuracy Decision Sciences Data Products Application Domains
  • 11. © 2018 International Business Machines Corporation IBM Research AI • Self-driving cars • Robotic surgery • Autonomous weapons • Humongous state space • Difficult value alignment Cyber-Physical Systems Application Domains
  • 12. © 2018 International Business Machines Corporation IBM Research AI Some of Our Focus Areas – Enterprise AI vs. Consumer AI ▪ Transferability ▪ Learning + Reasoning ▪ Safety and Security
  • 13. © 2018 International Business Machines Corporation IBM Research AI Agenda ▪ Formal definition of safety in machine learning ▪ Strategies for achieving safety ▪ Details on interpretable machine learning ▪ Details on algorithmic fairness – AI Fairness 360 ▪ Transparency via supplier’s declarations of conformity (SDoCs) ▪ Conclusion
  • 14. © 2018 International Business Machines Corporation IBM Research AI K. R. Varshney and H. Alemzadeh, “On the Safety of Machine Learning: Cyber-Physical Systems, Decision Sciences, and Data Products,” Big Data, vol. 5, no. 3, p. 246–255, Sept. 2017.
  • 15. © 2018 International Business Machines Corporation IBM Research AI Safety ▪ Commonly used term across engineering disciplines connoting the absence of failures or conditions that render a system dangerous (Ferrell, 2010) – Safe food and water, safe vehicles and roads, safe medical treatments, safe toys, safe neighborhoods, safe industrial plants, … ▪ Each domain has specific design principles and regulations applicable only to it ▪ Few works attempt a precise definition applicable broadly ▪ Definition based on harm, risk, and epistemic uncertainty (Möller, 2012)
  • 16. © 2018 International Business Machines Corporation IBM Research AI Harm ▪ A system yields an outcome based on its state and the inputs it receives ▪ The outcome event may be desired or undesired ▪ Outcomes have associated costs that can be measured and quantified by society ▪ An undesired outcome is a harm if its cost exceeds some threshold ▪ Unwanted events of small severity are not counted as safety issues
  • 17. © 2018 International Business Machines Corporation IBM Research AI Risk ▪ We do not know what the outcome will be, but its distribution is known and we can calculate the expectation of its cost ▪ Risk is the expected value of the cost of harm
  • 18. © 2018 International Business Machines Corporation IBM Research AI Epistemic Uncertainty ▪ We still do not know what the outcome will be, but in contrast to risk, its probability distribution is also unknown ▪ Epistemic uncertainty, in contrast to aleatoric uncertainty, results from lack of knowledge that could be obtained in principle, but may be practically intractable to gather ▪ Some decision theorists argue that all uncertainty can be captured probabilistically, but we maintain the distinction between risk and uncertainty, following Möller (2012)
  • 19. © 2018 International Business Machines Corporation IBM Research AI Safety ▪ Safety is the reduction or minimization of risk and epistemic uncertainty of harmful events ▪ Costs have to be sufficiently high in some human sense for events to be harmful ▪ Safety involves reducing both the probability of expected harms and the possibility of unexpected harms
  • 20. © 2018 International Business Machines Corporation IBM Research AI Risk Minimization in Statistical Machine Learning ▪ Risk minimization is the basis of statistical machine learning theory and practice – Features and labels with probability density – Function mapping – Loss function – Find to minimize risk ▪ Given i.i.d. training samples, not – Empirical risk minimization – converges to uniformly for all as goes to infinity (Glivenko-Cantelli) ▪ When is small, minimizing may not yield an with small – Restrict complexity of based on some inductive bias (Vapnik, 1992)
  • 21. © 2018 International Business Machines Corporation IBM Research AI Epistemic Uncertainty in Machine Learning ▪ Risk minimization has many strengths but does not capture uncertainty ▪ Not always the case that training samples are drawn from true underlying probability distribution of – The distribution the samples come from cannot always be known Precludes covariate shift and domain adaptation techniques – Training on a data set from a different distribution can cause much harm ▪ Even when drawn from true distribution, training samples may be absent from large parts of due to small probability density there
  • 22. © 2018 International Business Machines Corporation IBM Research AI Epistemic Uncertainty in Machine Learning ▪ Statistical learning theory analysis utilizes laws of large numbers to study the effect of finite training data and the convergence of to ▪ In considering safety, should also be cognizant that in deployment, a machine learning system only encounters a finite number of test samples – The actual operational risk is an empirical quantity of the test set – This operational risk may be much larger than the true risk for small cardinality test sets, even if is risk-optimal – This uncertainty caused by the instantiation of the test set can have large safety implications on individual test samples
  • 23. © 2018 International Business Machines Corporation IBM Research AI Loss Functions in Machine Learning ▪ In risk minimization, the domain of the loss function is and the range is an abstract quantity representing prediction error ▪ In real-world applications, the value of the loss function may be endowed with some human cost – That human cost may imply a loss function that also includes in the domain ▪ The cost may be severe enough to be harmful and thus a safety issue in some parts of the domain and not in others
  • 24. © 2018 International Business Machines Corporation IBM Research AI Value Alignment ▪ Obtaining is non-trivial and context-dependent ▪ Eliciting the values of society ▪ Encapsulation of morals
  • 25. © 2018 International Business Machines Corporation IBM Research AI Agenda ▪ Formal definition of safety in machine learning ▪ Strategies for achieving safety ▪ Details on interpretable machine learning ▪ Details on algorithmic fairness – AI Fairness 360 ▪ Transparency via supplier’s declarations of conformity (SDoCs) ▪ Conclusion
  • 26. © 2018 International Business Machines Corporation IBM Research AI Four Categories of Strategies for Achieving Safety ▪ Möller and Hansson (2008) have identified four main categories of approaches to achieve safety 1. Inherently safe design – Exclusion of a potential hazard from the system, e.g. excluding hydrogen from buoyant material of blimp 2. Safe fail – System remains safe when it fails in its intended operation, e.g. dead man’s switches 3. Safety reserves – Safety factors and safety margins, e.g. vessel wall is designed thicker than the largest pressure it will experience by some ratio 4. Procedural safeguards – Audits, training, posted warnings
  • 27. © 2018 International Business Machines Corporation IBM Research AI Inherently Safe Design in Machine Learning ▪ Want robustness against the uncertainty of the training set not being sampled from the test distribution – The training set may have various quirks or biases that are unknown to the user and that will not be present during the test phase – Highly complex modeling techniques used today, including extreme gradient boosting and deep neural networks, may pick up on those data vagaries in the learned models they produce to achieve high accuracy, but might fail due to an unknown shift in the data domain – The models are so complex that it is very difficult to understand how they will react to such shifts and whether they will produce harmful outcomes as a result ▪ Interpretability of models and causality of predictors – Extra constraints on ▪ May be performance loss in accuracy by doing so when measuring accuracy with a common training and testing data probability distribution, but the reduction in epistemic uncertainty by doing so increases safety
  • 28. © 2018 International Business Machines Corporation IBM Research AI Safe Fail in Machine Learning ▪ When predictions cannot be given confidently, a reject option is used – A human operator intervenes and provides a manual prediction ▪ Models are reported to be least confident and in the reject regime near the decision boundary in classification problems – Implicitly assumed that distance from the boundary is inversely related to confidence – Reasonable in parts of with high probability density and many training samples ▪ In parts of with low density and no training samples, decision boundary may be completely based on inductive bias with much epistemic uncertainty – Distance from boundary as a trigger for reject option is meaningless ▪ For a rare combination of features in a test sample, always elect manual examination (Attenberg et al., 2015)
  • 29. © 2018 International Business Machines Corporation IBM Research AI Safety Reserves in Machine Learning (Uncertainty) ▪ Parametrize all uncertainty using – Uncertainty in training set being from true distribution Uncertain class priors, label noise, etc. – Uncertainty in instantiation of test set Do not care as much about average test error as maximum test error in certain applications ▪ Let the risk of the risk-optimal model if were known be ▪ Robust formulations find while constraining or minimizing either:
  • 30. © 2018 International Business Machines Corporation IBM Research AI Safety Reserves in Machine Learning (Loss Function) ▪ Fairness – The risk of harm for members of protected groups should not be much worse than the risk of harm for others – Feature space partition contains unprivileged group members (e.g. defined by caste or gender) – Feature space partition contains priviliged group members ▪ Disparate impact constraint (Feldman et al., 2015)
  • 31. © 2018 International Business Machines Corporation IBM Research AI Procedural Safeguards in Machine Learning ▪ Defining the training set and setting up evaluation procedures have certain subtleties that can cause harm if done incorrectly – User experience design can be used to guide and warn users and thereby increase safety ▪ Open source software and open data allows for the possibility of public audit – Safety hazards can be discovered through public examination
  • 32. © 2018 International Business Machines Corporation IBM Research AI Agenda ▪ Formal definition of safety in machine learning ▪ Strategies for achieving safety ▪ Details on interpretable machine learning ▪ Details on algorithmic fairness – AI Fairness 360 ▪ Transparency via supplier’s declarations of conformity (SDoCs) ▪ Conclusion
  • 33. © 2018 International Business Machines Corporation IBM Research AI Three Dimensions of Explainability
 One explanation does not fit all: There are many ways to explain things directly interpretable vs. post hoc interpretation global (model-level) vs. local (instance-level) static vs. interactive (visual analytics) The oldest AI formats, such as decision rule sets, decision trees, and decision tables are simple enough for people to understand. Supervised learning of these models is directly interpretable. Start with a black box model and probe into it with a companion model to create interpretations. The black box model continues to provide the actual prediction while interpretation improve human interactions. Show the entire predictive model to the user to help them understand it (e.g. a small decision tree, whether obtained directly or in a post hoc manner). Only show the explanations associated with individual predictions (i.e. what was it about the features of this particular person that made her loan denied). The interpretation is simply presented to the user. The user can interact with interpretation.
  • 34. © 2018 International Business Machines Corporation IBM Research AI D. M. Malioutov and K. R. Varshney, “Exact Rule Learning via Boolean Compressed Sensing,” Int. Conf. Machine Learning, p. 765-773, Jun. 2013.
  • 35. © 2018 International Business Machines Corporation IBM Research AI Directly Interpretable Supervised Binary Classification ▪ Binary classification: special case of supervised learning when ▪ As discussed, different inductive biases lead to different ℋ ▪ The ℋ containing only sparse rule-based classifiers is interpretable (Freitas, 2014) ▪ Interpretability implies inherently safe design
  • 36. © 2018 International Business Machines Corporation IBM Research AI Rule-Based Classifiers ▪ Single rule is an AND clause of a few Boolean terms ▪ Predict Federer to defeat Murray if he: – Wins more than 59% of 4 to 9 shot rallies AND – Wins more than 78% of points when serving at 30-30 or Deuce AND – Serves less than 20% of serves into the body ▪ Rule set is a collection of single rules
  • 37. © 2018 International Business Machines Corporation IBM Research AI Challenges of Rule Learning ▪ Finding compact decision rules involving few Boolean terms that best approximate a given data set is an NP hard combinatorial optimization problem ▪ Most existing approaches for finding maximize criteria such as information gain, support, confidence, lift, Gini impurity, etc. – Decision trees, decision lists, RIPPER, SLIPPER, etc. – Greedy heuristics with ad hoc pruning ▪ Renewed interest in rule learning driven by optimizing a principled objective, but which retains interpretability (Rudin, 2015)
  • 38. © 2018 International Business Machines Corporation IBM Research AI Group Testing Problem ▪ Discover a sparse subset of faulty items in a large set of mostly good items using a few pooled tests – Blood screening of large groups of army recruits – Computational biology – Fault discovery in computer networks ▪ Mix together the blood of several recruits – If test is negative, none of the recruits are diseased – If test is positive, at least one of the recruits is diseased – Logical OR operation ▪ Construct the pools in an intelligent way to require a small number of tests with perfect recovery of diseased individuals
  • 39. © 2018 International Business Machines Corporation IBM Research AI Rule Learning as Group Testing ▪ Standard supervised binary classification problem – with features and Boolean labels ▪ Construct individual Boolean clauses from features , for – Months since promoted > 13 – Compensation plan == quota-based – For continuous dimensions of X, make comparisons to set of thresholds ▪ Calculate the truth value of each Boolean term for each training sample to construct an truth table matrix with entries
  • 40. © 2018 International Business Machines Corporation IBM Research AI Rule Learning as Group Testing (continued) ▪ The positive training samples are now equivalent to diseased pools of army recruits ▪ Determine an Boolean coefficient vector that specifies which Boolean terms to OR together in a decision rule to recover the positive samples ▪ Learn so that , where Boolean notation means: ▪ AND rules are more natural than OR rules for people – DeMorgan’s laws
  • 41. © 2018 International Business Machines Corporation IBM Research AI Discussion ▪ Efficient optimization by relaxation with provable guarantees ▪ Achieve competitive accuracy and high interpretability
  • 42. © 2018 International Business Machines Corporation IBM Research AI Selected
 Recent Explainability Innovations from IBM Research AI Global, Directly Interpretable Boolean Decision Rules via Column Generation NIPS 2018 (accepted). Dash, Gunluk, Wei. Global, Post-Hoc Improving Simple Models with Confidence Profiles NIPS 2018 (accepted). Dhurandhar, Shanmugam, Luss, Olsen. Local, Directly Interpretable Teaching Meaningful Explanations AAAI 2019 (submitted). Codella, Hind, Ramamurthy, Campbell, Dhurandhar, Varshney, Wei, Mojsilovic. Local, Post-Hoc Explanations Based on the Missing: Towards Contrastive Explanations with Pertinent Negatives NIPS 2018 (accepted). Dhurandhar, Chen, Luss, Tu, Ting, Shanmugam, Das. Interactive Model Visualization Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models IEEE VAST 2018 (accepted). Strobelt, Gehrmann, Behrisch, Perer, Pfister, Rush.
  • 43. © 2018 International Business Machines Corporation IBM Research AI Global, Post-Hoc Improving Simple Models with Confidence Profiles NIPS 2018. Dhurandhar, Shanmugam, Luss, Olsen. Domain experts preference: SMEs many times want to use a model they are comfortable with and trust. Small Data settings: Small client data where complex neural networks might overfit so simple models are preferred. However, possible to improve performance if we have a pretrained network on a large public or private corpora belonging to the same feature space. Resource Constrained Settings: In memory and power constrained settings only small models can be deployed. Here improving small neural networks with the help of larger neural network can be useful. Challenge: Transfer information from complex models with higher accuracy to the decision trees in a manner that enhances performance and adds valuable insight. Method: Find area under curve (AUC) based on confidence scores at each probe for each example based on complex neural network. Weight each training example by corresponding AUC (or optimal weighting procedure) and retrain simple model such as a decision tree.
  • 44. © 2018 International Business Machines Corporation IBM Research AI Local, Directly Interpretable
 
 
 Teaching Meaningful Explanations AAAI 2019 (submitted). Codella, Hind, Ramamurthy, Campbell, Dhurandhar, Varshney, Wei, Mojsilovic. • Ask SME to provide explanations (e) with their training data (x, y, e) • Train ML algorithm on features to jointly predict label (y) and explanations (e) • Apply classifier on new features inputs to produce label (y) and explanation (e)
  • 45. © 2018 International Business Machines Corporation IBM Research AI Local, Post-Hoc Explanations Based on the Missing: Towards Contrastive Explanations with Pertinent Negatives NIPS 2018. Dhurandhar, Chen, Luss, Tu, Ting, Shanmugam, Das. 3 • Current methods focus mostly on what is present based on positive and negative relevance. • Our method highlights what should be minimally sufficiently present and necessarily absent.
  • 46. © 2018 International Business Machines Corporation IBM Research AI Interactive Model Visualization
 
 
 Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models IEEE VAST 2018. Strobelt, Gehrmann, Behrisch, Perer, Pfister, Rush.
  • 47. © 2018 International Business Machines Corporation IBM Research AI Agenda ▪ Formal definition of safety in machine learning ▪ Strategies for achieving safety ▪ Details on interpretable machine learning ▪ Details on algorithmic fairness – AI Fairness 360 ▪ Transparency via supplier’s declarations of conformity (SDoCs) ▪ Conclusion
  • 48. © 2018 International Business Machines Corporation IBM Research AI Unwanted bias and algorithmic fairness
 Machine learning, by its very nature, is always a form of statistical discrimination Discrimination becomes objectionable when it places certain privileged groups at systematic advantage and certain unprivileged groups at systematic disadvantage Illegal in certain contexts
  • 49. © 2018 International Business Machines Corporation IBM Research AI Unwanted bias and algorithmic fairness
 Machine learning, by its very nature, is always a form of statistical discrimination Unwanted bias in training data yields models with unwanted bias that scale out Prejudice in labels Undersampling or oversampling
  • 50. © 2018 International Business Machines Corporation IBM Research AI Fairness in building and deploying models (d’Alessandro et al., 2017)
  • 51. © 2018 International Business Machines Corporation IBM Research AI Metrics, Algorithms dataset metric pre- processing algorithm in-processing algorithm post- processing algorithm classifier metric
  • 52. © 2018 International Business Machines Corporation IBM Research AI Metrics, Algorithms, and Explainers dataset metric pre- processing algorithm in-processing algorithm post- processing algorithm classifier metric classifier metric explainer dataset metric explainer
  • 53. © 2018 International Business Machines Corporation IBM Research AI 21 (or more) definitions of fairness
 Group and individual fairness Disparate impact, statistical parity, equality of odds, equality of opportunity There is no one definition of fairness applicable in all contexts Some definitions even conflict
  • 54. © 2018 International Business Machines Corporation IBM Research AI Bias mitigation is not easy
 Cannot simply drop protected attributes because features are correlated with them

  • 55. © 2018 International Business Machines Corporation IBM Research AI (Hardt, 2017) Research
 Algorithmic fairness is one of the hottest topics in the ML/AI research community

  • 56. © 2018 International Business Machines Corporation IBM Research AI F. P. Calmon, D. Wei, B. Vinzamuri, K. N. Ramamurthy, and K. R. Varshney, “Optimized Data Pre-Processing for Discrimination Prevention,” Advances in Neural Information Processing Systems, Dec. 2017.
  • 57. © 2018 International Business Machines Corporation Pre-Processing for Flexible Discrimination Prevention (Biased) Data Model Decision Pre-processing In-processing Post-processing Published Data
  • 58. © 2018 International Business Machines Corporation IBM Research AI
  • 59. © 2018 International Business Machines Corporation Our Approach ▪ Given dataset containing • – discriminatory (protected) variables • – other (non-protected) variables • – outcome ▪ Learn optimized (randomized) mapping ▪ Map both train and test data before model Race Age Degree Priors Recid Black >45 Misd. 0 N White 25-45 Felony 0 N Black <25 Felony 1-3 Y White >45 Misd. 0 N Race Age Degree Priors Recid Black >45 Felony 0 N White >45 Felony 0 N Black <25 Felony 1-3 N White >45 Misd. 0 N
  • 60. © 2018 International Business Machines Corporation Optimization Criteria 1. Group discrimination Control dependence of transformed outcome on 3. Utility preservation Retain joint distribution so model can still learn task 2. Individual distortion Avoid large changes in individual features 𝑥, 𝑦 ^𝑥, ^𝑦 𝛿 𝑑1 𝑑2
  • 61. © 2018 International Business Machines Corporation Original COMPAS Dataset re-commit crime race red = misdemeanor
 blue = felony
  • 62. © 2018 International Business Machines Corporation Pre-Processed COMPAS Dataset race red = misdemeanor
 blue = felony re-commit crime
  • 63. © 2018 International Business Machines Corporation Interpretable Classifier Learned from Original Dataset prior crimes race > 0 predicted not to 
 re-offend = 0 prior crimes age Caucasian African-American predicted not to 
 re-offend predicted to 
 re-offend > 3 ≤ 3 predicted not to 
 re-offend predicted to 
 re-offend < 45 ≥ 45
  • 64. © 2018 International Business Machines Corporation Interpretable Classifier Learned from Pre-Processed Dataset prior crimes predicted to 
 re-offend > 3 predicted not to 
 re-offend ≤ 3
  • 65. © 2018 International Business Machines Corporation IBM Research AI Balanced Accuracy vs. Fairness For a classifier trained with original training data, at the best classification rate, this measure is quite high. This implies unfairness. abs(1-disparate impact) must be close to zero (typically < 0.2) for classifier predictions to be fair. For a classifier trained with transformed training data, at the best classification rate, the measure drops significantly, implying bias mitigation.
  • 66. © 2018 International Business Machines Corporation IBM Research AI P. Sattigeri, S. C. Hoffman, V. Chenthamarakshan, and K. R. Varshney, “Fairness GAN,” AAAI 2019, submitted.
  • 67. © 2018 International Business Machines Corporation Our Approach An auxiliary classifier generative adversarial network (AC-GAN) with one main twist: try to predict the protected attribute as poorly as possible from the features
  • 68. © 2018 International Business Machines Corporation 1. Joint Conditional Generative Model Learn a generative model that matches the true distribution . 3. GAN Formulation Posed as an adversarial learning problem and solved using novel GAN formulation 2. Label and Sensitive Attribute Independence Maximization Make outcome and sensitive variable () independent in the learned distribution. Discriminator Training: Generator Training: Fairness GAN
  • 69. © 2018 International Business Machines Corporation IBM Research AI Fairness GAN
  • 70. © 2018 International Business Machines Corporation IBM Research AI Fairness GAN
  • 71. © 2018 International Business Machines Corporation IBM Research AI Agenda ▪ Formal definition of safety in machine learning ▪ Strategies for achieving safety ▪ Details on interpretable machine learning ▪ Details on algorithmic fairness – AI Fairness 360 ▪ Transparency via supplier’s declarations of conformity (SDoCs) ▪ Conclusion
  • 72. 05/03/18 Facebook says it has a tool to detect bias in its artificial intelligence Quartz 05/25/18 Microsoft is creating an oracle for catching biased AI algorithms MIT Technology Review 05/31/18 Pymetrics open-sources Audit AI, an algorithm bias detection tool VentureBeat 06/07/18 Google Education Guide to Responsible AI Practices – Fairness Google 06/09/18 Accenture wants to beat unfair AI with a professional toolkit TechCrunch
  • 73. Fairness Measures Framework to test given algorithm on variety of datasets and fairness metrics https://github.com/megantosh/ fairness_measures_code Fairness Comparison Extensible test-bed to facilitate direct comparisons of algorithms with respect to fairness measures. Includes raw & preprocessed datasets https://github.com/algofairness/ fairness-comparison Themis-ML Python library built on scikit-learn that implements fairness-aware machine learning algorithms https://github.com/cosmicBboy/themis- ml FairML Looks at significance of model inputs to quantify prediction dependence on inputs https://github.com/adebayoj/fairml Aequitas Web audit tool as well as python lib. Generates bias report for given model and dataset https://github.com/dssg/aequitas Fairtest Tests for associations between algorithm outputs and protected populations https://github.com/columbia/fairtest Themis Takes a black-box decision-making procedure and designs test cases automatically to explore where the procedure might be exhibiting group-based or causal discrimination https://github.com/LASER-UMASS/ Themis Audit-AI Python library built on top of scikit-learn with various statistical tests for classification and regression tasks https://github.com/pymetrics/audit-ai
  • 74. © 2018 International Business Machines Corporation IBM Research AI AI Fairness 360
 Differentiation Datasets Toolbox Fairness metrics (30+) Fairness metric explanations Bias mitigation algorithms (9+) Guidance Industry-specific tutorials Comprehensive bias mitigation toolbox (including unique algorithms from IBM Research) Several metrics and algorithms that have no available implementations elsewhere Extensible Designed to translate new research from the lab to industry practitioners
 (e.g. scikit-learn’s fit/predict paradigm)
  • 75. © 2018 International Business Machines Corporation IBM Research AI
  • 76. © 2018 International Business Machines Corporation IBM Research AI
  • 77. © 2018 International Business Machines Corporation IBM Research AI Agenda ▪ Formal definition of safety in machine learning ▪ Strategies for achieving safety ▪ Details on interpretable machine learning ▪ Details on algorithmic fairness – AI Fairness 360 ▪ Transparency via supplier’s declarations of conformity (SDoCs) ▪ Conclusion
  • 78. © 2018 International Business Machines Corporation IBM Research AI Supplier’s Declarations of Conformity ▪ A common factsheet voluntarily released with products in many different industries ▪ Intended use, provenance, etc. ▪ Reliability, safety, and security test results ▪ Transparency for the final step of achieving trust ▪ Information asymmetry / market for lemons
  • 79. © 2018 International Business Machines Corporation IBM Research AI Supplier’s Declarations of Conformity ▪ A common factsheet voluntarily released with products in many different industries ▪ Intended use, provenance, etc. ▪ Reliability, safety, and security test results ▪ Transparency for the final step of achieving trust ▪ Information asymmetry / market for lemons
  • 80. © 2018 International Business Machines Corporation IBM Research AI Supplier’s Declarations of Conformity ▪ What is the intended use of the service output? ▪ What algorithms or techniques does this service implement? ▪ Which datasets was the service tested on? (Provide links to datasets that were used for testing, along with corresponding datasheets.) ▪ Describe the testing methodology. ▪ Describe the test results. ▪ Are you aware of possible examples of bias, ethical issues, or other safety risks as a result of using the service? ▪ Are the service outputs explainable and/or interpretable? ▪ For each dataset used by the service: Was the dataset checked for bias? What efforts were made to ensure that it is fair and representative? ▪ Does the service implement and perform any bias detection and remediation? ▪ What is the expected performance on unseen data or data with different distributions? ▪ Was the service checked for robustness against adversarial attacks? ▪ When were the models last updated?
  • 81. © 2018 International Business Machines Corporation IBM Research AI Agenda ▪ Formal definition of safety in machine learning ▪ Strategies for achieving safety ▪ Details on interpretable machine learning ▪ Details on algorithmic fairness – AI Fairness 360 ▪ Transparency via supplier’s declarations of conformity (SDoCs) ▪ Conclusion
  • 82. © 2018 International Business Machines Corporation IBM Research AI Summary and Conclusion ▪ Considerations beyond accuracy in enterprise AI ▪ Very basic definition of safety in terms of harm, risk, uncertainty ▪ Minimization of epistemic uncertainty is missing from standard modes of machine learning developed around risk minimization ▪ Research agenda motivated by trust – Risk minimization close to the limits of what is achievable – Epistemic uncertainty minimization has new and exciting problems – Some examples of strategies for increasing safety in machine learning, but far from complete – Interpretability, fairness, SDoCs – Open source toolbox: AI Fairness 360 (aif360)
  • 83. © 2018 International Business Machines Corporation http://aif360.mybluemix.net
 https://github.com/ibm/aif360
 https://pypi.org/project/aif360