Interpretability and Explainability in Machine Learning and their Application to Nonlinear Dimensionality Reduction
Défense de thèse d'Adrien Bibal à distance via teams
Date : 16/11/2020 13:00 - 16/11/2020 15:00
Lieu : https://teams.microsoft.com/l/meetup-join/19%3ameeting_MzAzMGEyMDMtMzVlZC00MzMxLTlmZTYtM2ZjZWEwMTIyM2Y4%40thread.v2/0?context=%7b%22Tid%22%3a%225f31c5b4-f2e8-4772-8dd6-f268037b1eca%22%2c%22Oid%22%3a%22972bbe5b-17b9-4b75-bbbb-a7623671984b%22%7d
Orateur(s) : Adrien Bibal
Organisateur(s) : Isabelle Daelman
Machine learning (ML) techniques are more and more frequently
used today because of their high performance in many contexts.
However, the rise in performance comes at the cost of a lack of
control over the model that is learned. Indeed, while modelling
was mainly done by experts in the past, the surge of data makes it
possible to automatically derive models. Unfortunately, this
automatization can result in the production of non-understandable
models. This concept of model understandability is referred to as
interpretability in the literature. Furthermore, when models are
not interpretable, it is their ability to be explained (their
explainability) that is exploited.
This thesis explores interpretability and explainability in ML.
Several aspects of these concepts are studied. First, the problem
of defining interpretability and explainability, as well as the
vocabulary used in the literature, is presented. Second, the
requirements of the law for these concept are studied. Then, the
way interpretability and explainability involve users in their
evaluation is discussed and guidelines from the human-computer
interaction community are presented.
This thesis also applies the concepts of interpretability and
explainability to the problem of nonlinear dimensionality
reduction (NLDR). While the subjects of interpretability and
explainability in NLDR have barely been touched in the literature,
this thesis provides a conceptualization of interpretability and
explainability in the context of NLDR, as well as new techniques
to deal with them. In particular, two questions are central in
this thesis ``how can interpretability can be measured in NLDR?''
and ``how can non-interpretable NLDR mappings be explained?''.
For measuring interpretability in NLDR, we analyze how existing
metrics from different communities can be combined to predict user
understanding of NLDR embeddings. In particular, ML quality
metrics are used to assess how low-dimensional (LD) embeddings are
faithful to the high-dimensional (HD) data, and information
visualization quality metrics are used to assess how
understandable visualizations are. In the context of NLDR mappings
that are considered to be non-interpretable, IXVC was developed to
explain the mapping between visual clusters in a NLDR embedding
and HD data through an interactive pipeline. Another approach for
explaining NLDR mappings through the embedding dimensions was
developed in our two techniques BIR and BIOT. Even though previous
work has tried to develop more explicit, parametric, mappings, to
the best of our knowledge, our works in this thesis are the first
to elaborate on the term ``interpretability'' in the field of
NLDR.
Contact :
Isabelle Daelman
-
isabelle.daelman@unamur.be
Télecharger :
vCal