Utilize este identificador para referenciar este registo:
https://hdl.handle.net/1822/79434
Título: | A hybrid post hoc interpretability approach for deep neural networks |
Autor(es): | Santos, Flávio Arthur Oliveira Zanchettin, Cleber Silva, José Vitor Santos Matos, Leonardo Nogueira Novais, Paulo |
Palavras-chave: | Deep learning Optimization Interpretability Fairness |
Data: | 2021 |
Editora: | Springer, Cham |
Revista: | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
Citação: | Santos, F.A.O., Zanchettin, C., Silva, J.V.S., Matos, L.N., Novais, P. (2021). A Hybrid Post Hoc Interpretability Approach for Deep Neural Networks. In: Sanjurjo González, H., Pastor López, I., García Bringas, P., Quintián, H., Corchado, E. (eds) Hybrid Artificial Intelligent Systems. HAIS 2021. Lecture Notes in Computer Science(), vol 12886. Springer, Cham. https://doi.org/10.1007/978-3-030-86271-8_50 |
Resumo(s): | Every day researchers publish works with state-of-the-art results using deep learning models, however as these models become common even in production, ensuring fairness is a main concern of the deep learning models. One way to analyze the model fairness is based on the model interpretability, obtaining the essential features to the model decision. There are many interpretability methods to produce the deep learning model interpretation, such as Saliency, GradCam, Integrated Gradients, Layer-wise relevance propagation, and others. Although those methods make the feature importance map, different methods have different interpretations, and their evaluation relies on qualitative analysis. In this work, we propose the Iterative post hoc attribution approach, which consists of seeing the interpretability problem as an optimization view guided by two objective definitions of what our solution considers important. We solve the optimization problem with a hybrid approach considering the optimization algorithm and the deep neural network model. The obtained results show that our approach can select the features essential to the model prediction more accurately than the traditional interpretability methods. |
Tipo: | Artigo em ata de conferência |
URI: | https://hdl.handle.net/1822/79434 |
ISBN: | 978-3-030-86270-1 |
e-ISBN: | 978-3-030-86271-8 |
DOI: | 10.1007/978-3-030-86271-8_50 |
ISSN: | 0302-9743 |
Arbitragem científica: | yes |
Acesso: | Acesso restrito UMinho |
Aparece nas coleções: |
Ficheiros deste registo:
Ficheiro | Descrição | Tamanho | Formato | |
---|---|---|---|---|
HAIS54.pdf Acesso restrito! | 765,02 kB | Adobe PDF | Ver/Abrir |