计算机科学 ›› 2024, Vol. 51 ›› Issue (2): 245-251.doi: 10.11896/jsjkx.230300028
周晟昊, 袁伟伟, 关东海
ZHOU Shenghao, YUAN Weiwei, GUAN Donghai
摘要: 深度学习模型的广泛使用,在更大程度上使人们意识到模型的决策是亟需解决的问题,复杂难以解释的黑盒模型阻碍了算法在实际场景中部署。LIME作为最流行的局部解释方法,生成的扰动数据却具有不稳定性,导致最终的解释产生偏差。针对上述问题,提出了一种基于主动学习和二次有理核的模型无关局部解释方法ActiveLIME,使得局部解释模型更加忠于原始分类器。ActiveLIME生成扰动数据后,通过主动学习的查询策略对扰动数据进行采样,筛选不确定性高的扰动集训练,使用迭代过程中准确度最高的局部模型对感兴趣实例生成解释。并且,针对容易陷入局部过拟合的高维稀疏样本,在模型损失函数中引入了二次有理核来减少过拟合。实验结果表明,所提出的ActiveLIME方法引比传统局部解释方法具有更高的局部保真度和解释质量。
中图分类号:
[1]BAI X,WANG X,LIU X,et al.Explainable deep learning for efficient and robust pattern recognition:A survey of recent deve-lopments[J].Pattern Recognition,2021,120:108102. [2]BREIMAN L.Classification and regression trees[M].Rout-ledge,2017. [3]LINARDATOS P,PAPASTEFANOPOULOS V,KOTSIAN-TIS S.Explainable ai:A review of machine learning interpre-tability methods[J].Entropy,2020,23(1):18. [4]DU M,LIU N,HU X.Techniques for interpretable machine learning[J].Communications of the ACM,2019,63(1):68-77. [5]WANG F,KAUSHAL R,KHULLAR D.Should health care demand interpretable artificial intelligence or accept “black box” medicine?[J].Annals of Internal Medicine,2020,172(1):59-60. [6]SHANKARANARAYANA S M,RUNJE D.ALIME:Autoencoder based approach for local interpretability[C]//Intelligent Data Engineering and Automated Learning-IDEAL 2019:20th International Conference,Manchester,UK,November 14-16,2019,Proceedings,Part I 20.Springer International Publishing,2019:454-463. [7]RIBEIRO M T,SINGH S,GUESTRIN C.“Why should I trustyou?” Explaining the predictions of any classifier[C]//Procee-dings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.2016:1135-1144. [8]LAKKARAJU H,BACH S H,LESKOVEC J.Interpretable decision sets:A joint framework for description and prediction[C]//Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.2016:1675-1684. [9]SETTLES B.Active learning literature survey[R].ComputerSciences Technical Report 1648,University of Wisconsin-Madison,2009. [10]MUSLEA I,MINTON S,KNOBLOCK C A.Active learningwith multiple views[J].Journal of Artificial Intelligence Research,2006,27:203-233. [11]SEUNG H S,OPPER M,SOMPOLINSKY H.Query by committee[C]//Proceedings of the fifth Annual Workshop on Computational Learning Theory.1992:287-294. [12]ZAFAR M R,KHAN N.Deterministic local interpretable mo-del-agnostic explanations for stable explainability[J].Machine Learning and Knowledge Extraction,2021,3(3):525-541. [13]RANJBAR N,SAFABAKHSH R.Using decision tree as local interpretable model in autoencoder-based lime[C]//2022 27th International Computer Conference.Computer Society of Iran(CSICC),IEEE,2022:1-7. [14]LAUGEL T,RENARD X,LESOT M J,et al.Defining locality for surrogates in post-hoc interpretablity[J].arXiv:1806.07498,2018. [15]RIBEIRO M T,SINGH S,GUESTRIN C.Anchors:High-precision model-agnostic explanations[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2018. [16]RIBEIRO M T,SINGH S,GUESTRIN C.Nothing else mat-ters:Model-agnostic explanations by identifying prediction invariance[J].arXiv:1611.05817,2016. [17]ZHAO X,HUANG W,HUANG X,et al.Baylime:Bayesian local interpretable model-agnostic explanations[C]//Uncertainty in Artificial Intelligence(PMLR).2021:887-896. [18]ADLER P,FALK C,FRIEDLER S A,et al.Auditing black-box models for indirect influence[J].Knowledge and Information Systems,2018,54:95-122. [19]BRAMHALL S,HORN H,TIEU M,et al.Qlime-a quadratic local interpretable model-agnostic explanation approach[J].SMU Data Science Review,2020,3(1):4. [20]HOFMANN T,SCHÖLKOPF B,SMOLA A J.Kernel methods in machine learning[J].The Annals of Statistics,2008,36(3):1171. [21]JANOCHA K,CZARNECKI W M.On loss functions for deep neural networks in classification[J].arXiv:1702.05659,2017. [22]GUO G,WANG H,BELL D,et al.KNN model-based approach in classification[C]//OTM Confederated International Confe-rences on The Move to Meaningful Internet Systems 2003:CoopIS,DOA,and ODBASE.2003:986-996. [23]NIELSEN F,NIELSEN F.Hierarchical clustering[J/OL].Introduction to HPC with MPI for Data Science,2016:195-211.https://link.springer.com/chapter/10.1007/978-3-319-21903-5_8. [24]VILONE G,LONGO L.Notions of explainability and evaluation approaches for explainable artificial intelligence[J].Information Fusion,2021,76:89-106. [25]GRANDINI M,BAGLI E,VISANI G.Metrics for multi-class classification:an overview[J].arXiv:2008.05756,2020. [26]JIA Y,BAILEY J,RAMAMOHANARAO K,et al.Improving the quality of explanations with local embedding perturbations[C]//Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining.2019:875-884. [27]KEERTHI S S,LIN C J.Asymptotic behaviors of support vector machines with Gaussian kernel[J].Neural Computation,2003,15(7):1667-1689. |
|