An explainable ensemble feedforward method with Gaussian convolutional filter

Jingchen Li, Haobin Shi, Kao Shing Hwang

科研成果: 期刊稿件文章同行评审

29 引用 (Scopus)

摘要

The emerging deep learning technologies are leading to a new wave of artificial intelligence, but in some critical applications such as medical image processing, deep learning is inapplicable due to the lack of interpretation, which is essential for a critical application. This work develops an explainable feedforward model with Gaussian kernels, in which the Gaussian mixture model is leveraged to extract representative features. To make the error within the allowable range, we calculate the lower bound of the number of samples through the Chebyshev inequality. In the training processing, we discuss both the deterministic and stochastic feature representations, and investigate the performance of them and the ensemble model. Additionally, we use Shapely additive explanations to analyze the experiment results. The proposed method is interpretable, so it can replace the deep neural network by working with shallow machine learning technologies, such as the Support Vector Machine and Random Forest. We compare our method with baseline methods on Brain Tumor and Mitosis dataset. The experimental results show our method outperforms the RAM (Recurrent Attention Model), VGG19 (Visual Geometry Group 19), LeNET-5, and Explainable Prediction Framework while having strong interpretability.

源语言英语
文章编号107103
期刊Knowledge-Based Systems
225
DOI
出版状态已出版 - 5 8月 2021

指纹

探究 'An explainable ensemble feedforward method with Gaussian convolutional filter' 的科研主题。它们共同构成独一无二的指纹。

引用此