• Aucun résultat trouvé

Explainable Artificial Intelligence

N/A
N/A
Protected

Academic year: 2022

Partager "Explainable Artificial Intelligence"

Copied!
25
0
0

Texte intégral

(1)

Explainable Artificial Intelligence

Student: Nedeljko Radulović

Supervisors: Mr. Albert Bifet and Mr. Fabian Suchanek

(2)

Introduction

(3)

Research avenues

● Explainability

● Integration of first-order logic and Deep Learning

● Detecting vandalism in Knowledge Bases based on correction history

(4)

Context

● Machine Learning and Deep Learning models sometimes exceed the human performance in decision making

● Major drawback is lack of transparency and interpretability

● Bringing transparency to the ML models is a crucial step towards the Explainable Artificial Intelligence and its use in very sensitive fields

(5)

State of the art

● Exlplainable Artificial Intelligence is the topic of great interest in research in recent years

● Interpretability:

Using visualization techniques (mostly used in image and text classification)

● Explainability:

Computing influence from inputs to outputs

Approximating complex model with a simpler model locally (LIME)

(6)

State of the art

● Attempts to combine Machine Learning and knowledge from Knowledge Bases

Reasoning over knowledge base embeddings to provide explainable recommendations

(7)

Explainability

(8)

Explainability

(9)

Explainability

(10)

LIME 1 - Explaining the predictions of any classifier

1: https://arxiv.org/abs/1602.04938

(11)

Explaining predictions in streaming setting

● Idea behind LIME is to use simple models to explain predictions

● Use already interpretable models - Decision trees

● Build Decision tree in the neighbourhood of the example

● Use the paths to leaves to generate explanations

● Use Hoeffding Adaptive Tree in streaming setting and explain how predictions evolve based on changes in the tree

(12)

Integration of First-order logic

and Deep Learning

(13)

Integration of FOL and Deep Learning

● Ultimate goal of Artificial Intelligence: enable machines to think as humans

● Humans posses some knowledge and are able to reason on top of it

Knowledge Reasoning

KBs ML

Deep Learning

SVM Random forest Logistic regression

(14)

Integration of FOL and Deep Learning

● There are several questions that we want to answer through this research:

How can KBs be used to inject meaning into complex and uninterpretable models, especially deep neural networks?

How can KBs be used more effectively as (additional) input for deep learning models?

How we can adjust all these improvements for streaming setting?

(15)

Main Idea

● Explore symbiosis of crisp knowledge in Knowledge Bases and sub-symbolic knowledge in Deep Neural Networks

● Approaches that combined crisp logic and soft reasoning:

Fuzzy logic

Markov logic

Probabilistic soft logic

(16)

Fuzzy logic - Fuzzy set

(17)

S L

Fuzzy logic - Fuzzy relation and Fuzzy graph

close to Chicago Sydney

New York 0.9 0.1

London 0.5 0.3

Beijing 0.2 0.7

N

B

C

0.9 0.1 0.5

0.2 0.7

0.3

(18)

Markov Logic and Probabilistic Soft Logic

● First-order logic as template language

● Example:

Predicates: friend, spouse, votesFor

Rules:

friend(Bob, Ann) votesFor(Ann,P) → votesFor(Bob, P) spouse(Bob, Ann) votesFor(Ann,P) → votesFor(Bob, P)

(19)

Markov Logic

● Add weights to first-order logic rules:

Interpretation: Every atom (friend(Bob, Ann), votesFor(Ann,P), votesFor(Bob, P), spouse(Bob, Ann)) is considered as random variable which can be: True or False

● To calculate probability of an interpretation:

friend(Bob, Ann) votesFor(Ann,P) → votesFor(Bob, P) : [3]

spouse(Bob, Ann) votesFor(Ann,P) → votesFor(Bob, P) : [8]

(20)

Probabilistic Soft Logic

● Add weights to first-order logic rules:

Interpretation: Every atom (friend(Bob, Ann), votesFor(Ann,P), votesFor(Bob, P), spouse(Bob, Ann)) is mapped to soft truth values in range [0, 1]

● For every rule we compute distance to satisfaction:

Probability density function over I:

friend(Bob, Ann) votesFor(Ann,P) → votesFor(Bob, P) : [3]

spouse(Bob, Ann) votesFor(Ann,P) → votesFor(Bob, P) : [8]

dr(I) = max{0, I(rbody) - I(rhead)}

(21)

Detecting vandalism in

Knowledge bases based on

correction history

(22)

Detecting vandalism in KBs based on correction history

● Collaboration with Thomas Pellissier Tanon

● Based on a paper: “Learning How to Correct a Knowledge Base from Edit History”

● Wikidata project

● Wikidata is a collaborative KB with more than 18000 active contributors

● Huge edit history: over 700 millions edits

● Method uses previous users corrections to infer possible new ones

(23)

Detecting vandalism in KBs based on correction history

● Prospective work in this project:

Release history querying system for external use

Try to use external knowledge (Wikipedia articles) to learn to fix more constraints violations

Use Machine Learning to suggest new updates

Use data stream mining techniques

(24)

Thank you!

Questions, ideas… ?

(25)

Research avenues

● Explainability

● Integration of first-order logic and Deep Learning

● Detecting vandalism in Knowledge Bases based on correction history

Références

Documents relatifs

Authors consider Wildlife Monitoring as a Cooperative Target Observation problem and propose a Multi Criteria Decision Analysis based algorithm to maximize the

salperwy/index.php?id=moa-extension.. • HTNB: Hoeffding Tree using a na¨ıve Bayes - This version is identical to the previous one but possesses a na¨ıve Bayes classifier in the

The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.. L’archive ouverte pluridisciplinaire HAL, est

As a matter of fact, a restricted class of sheaves automata [13] has been used to prove complexity properties of the static fragment of ambient logic, which corresponds to a kind

The question therefore arises as to what happens to familiar, decidable fragments of first-order logic under the assumption that a specified collection of binary predicates

Thus, explana- tions for AI models are less likely to be off in case of the data missingness is because of MAR or MCAR but imputation for NMAR data could occasionally lead

The results of this analysis are summarized in Figure 6, indicating that students want to know why during the first half of their interaction (i.e. first and second

We establish the upper bound by character- izing the satisfiability of a formula in GFU via mosaics, where a mosaic is a special (finite) collection of types that can be used to build