Computer Science ›› 2026, Vol. 53 ›› Issue (4): 66-77.doi: 10.11896/jsjkx.251000012

• Interdisciplinary Integration of Artificial Intelligence and Theoretical Computer Science • Previous Articles     Next Articles

Causal Disentangled Representation Learning with Integrated Sparse Coding

HUANG Beibei, LIU Jinfeng   

  1. School of Information Engineering, Ningxia University, Yinchuan 750021, China
  • Received:2025-10-09 Revised:2026-01-19 Online:2026-04-15 Published:2026-04-08
  • About author:HUANG Beibei,born in 2001,master candidate,is a student member of CCF(No.A06163G).Her main research interests include computer vision and representation learning.
    LIU Jinfeng,born in 1971,Ph.D,professor,is a professional member of CCF(No.75718M).His main research interests include deep learning andhete-rogeneous computing.
  • Supported by:
    Natural Science Foundation of Ningxia(2025AAC030154).

Abstract: Deep learning models often lack of interpretability in their feature representations due to their “black-box” nature.Although existing disentangled representation learning methods can enhance interpretability to some extent by identifying independent factors within the data,they usually neglect complex correlations and potential causal structures,which limits their applicability in critical domains such as autonomous driving and medical diagnosis,especially in scenarios that require understanding and intervention of causal relationships.To address the insufficient causal modeling in current disentangled representation learning,a disentanglement framework integrating sparse coding with causal inference is constructed.Under appropriate supervision,this framework leverages a causal inference mechanism to precisely model causal relationships within the data,thereby not only generating high-quality and structured representations but also enabling the modeling and intervention of potential causal mechanisms,which significantly improves the model’s adaptability and robustness in causal tasks.Meanwhile,the embedded convolutional sparse coding layer imposes sparsity constraints to effectively filter key representations highly relevant to causal structures,further enhancing the model’s sensitivity and expressive capacity for higher-order causal relationships.Experimental results demonstrate that the proposed framework performs excellently on both the Pendulum and CelebA datasets,achieving a sample efficiency of 98.65% on the Pendulum dataset and 99.55% on the CelebA dataset.Moreover,it outperforms existing methods in terms of causal intervention effectiveness and distribution robustness,confirming its superiority in complex causal scenarios.

Key words: Sparse coding, Causal inference, Disentangled representation learning, Sample efficiency, Distribution robustness

CLC Number: 

  • TP181
[1]CHENG K Y,MENG C Y,WANG W S,et al.Research Progress in Disentangled Representation Learning[J].Journal of Computer Applications,2021,41(12):3409-3418.
[2]HIGGINS I,AMOS D,PFAU D,et al.Towards a definition ofdisentangled representations[J].arXiv:1812.02230,2018.
[3]KINGMA D P,WELLING M.Auto-encoding variational Bayes[J].arXiv:1312.6114,2013.
[4]GOODFELLOW I J,POUGET-ABADIE J,MIRZA M,et al.Generative adversarial nets[C]//Proceedings of the 28th International Conference on Neural Information Processing Systems.Cambridge,MA:MIT Press,2014:2672-2680.
[5]SIKKA H D.A Deeper Look at the Unsupervised Learning of Disentangled Representations in β-VAE from the perspective of Core Object Recognition[M].Harvard University,2020.
[6]PEARL J.Models,reasoning and inference[J].Cambridge,UK:Cambridge University Press,2000,19(2):3.
[7]YANG T S.Disentangled Representation Learning Based onCausal Inference[D].Qingdao:Qingdao University of Science and Technology,2023.
[8]SCHÖLKOPF B,LOCATELLO F,BAUER S,et al.Towardcausal representation learning[J].Proceedings of the IEEE,2021,109(5):612-634.
[9]WEN Z D,WANG J R,WANG X X,et al.A Survey on Disentangled Representation Learning[J].Acta Automatica Sinica,2022,48(2):351-374.
[10]HIGGINS I,MATTHEY L,PAL A,et al.beta-vae:Learningbasic visual concepts with a constrained variational framework[C]//International Conference on Learning Representations.2017.
[11]CHEN X,DUAN Y,HOUTHOOFT R,et al.Infogan:Inter-pretable representation learning by information maximizing generative adversarial nets[C]//Advances in Neural Information Processing Systems.2016.
[12]CHEN R T Q,LI X,GROSSE R,et al.Isolating sources of disentanglement in VAEs[C]//Proceedings of the 32nd International Conference on Neural Information Processing Systems.2018:2615-2625.
[13]REZAABAD A L,VISHWANATH S.Learning representations by maximizing mutual information in variational autoencoders[C]//2020 IEEE International Symposium on Information Theory(ISIT).IEEE,2020:2729-2734.
[14]LOCATELLO F,BAUER S,LUCIC M,et al.Challenging common assumptions in the unsupervised learning of disentangled representations[C]//International Conference on Machine Learning.PMLR,2019:4114-4124.
[15]SUTER R,MILADINOVIC D,SCHÖLKOPF B,et al.Robustly disentangled causal mechanisms:Validating deep representations for interventional robustness[C]//International Conference on Machine Learning.PMLR,2019:6056-6065.
[16]REDDY A G,BALASUBRAMANIAN V N.On causally disentangled representations[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2022:8089-8097.
[17]YANG M,LIU F,CHEN Z,et al.Causalvae:Disentangled representation learning via neural structural causal models[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition.2021:9593-9602.
[18]SHEN X,LIU F,DONG H,et al.Weakly supervised disentangled generative causal representation learning[J].Journal of Machine Learning Research,2022,23(241):1-55.
[19]KOMANDURI A,WU Y,CHEN F,et al.Learning CausallyDisentangled Representations via the Principle of Independent Causal Mechanisms[J].arXiv:2306.01213,2023.
[20]YOU D,LI Z,SHEN J,et al.Disentangled representation learning with causal effect transmission in variational autoencoder[J].Pattern Recognition,2026,170:112018.
[21]NASR-ESFAHANY A,ALIZADEH M,SHAH D.Counterfactual identifiability of bijective causal models[C]//International Conference on Machine Learning.PMLR,2023:25733-25754.
[22]LI M,ZHAI P,TONG S,et al.Revisiting sparse convolutional model for visual recognition[J].Advances in Neural Information Processing Systems,2022,35:10492-10504.
[23]PAPAMAKARIOS G,NALISNICK E,REZENDE D J,et al.Normalizing flows for probabilistic modeling and inference[J].Journal of Machine Learning Research,2021,22(57):1-64.
[24]KHEMAKHEM I,MONTI R,LEECH R,et al.Causal autoregressive flows[C]//International Conference on Artificial Intelligence and Statistics.PMLR,2021:3520-3528.
[25]LIU Z,LUO P,WANG X,et al.Deep learning face attributes in the wild[C]//Proceedings of the IEEE International Conference on Computer Vision.2015:3730-3738.
[26]GONDAL M W,WUTHRICH M,MILADINOVIC D,et al.On the transfer of inductive bias from simulation to the real world:a new disentanglement dataset[C]//Advances in Neural Information Processing Systems.2019.
[27]KINNEY J B,ATWAL G S.Equitability,mutual information,and the maximal information coefficient[C]//Proceedings of the National Academy of Sciences.2014:3354-3359.
[1] LIU Yuting, GU Jingjing, ZHOU Qiang. Urban Flow Prediction Method Based on Structural Causal Model [J]. Computer Science, 2025, 52(10): 70-78.
[2] XU Bei, LIU Tong. Semi-supervised Emotional Music Generation Method Based on Improved Gaussian Mixture Variational Autoencoders [J]. Computer Science, 2024, 51(8): 281-296.
[3] LYU Jiahao, LIU Jinfeng. Domain Generalization and Long-tailed Learning Based on Causal Relationships [J]. Computer Science, 2024, 51(11A): 240300041-8.
[4] HUANG Lu, NI Lyu, JIN Cheqing. Rectifying Dual Bias for Recommendation [J]. Computer Science, 2023, 50(9): 152-159.
[5] ZHANG Tao, CHENG Yifei, SUN Xinxu. Graph Attention Networks Based on Causal Inference [J]. Computer Science, 2023, 50(6A): 220600230-9.
[6] LIANG Meiyan, ZHANG Yu, LIANG Jianan, CHEN Qinghui, WANG Ru, WANG Lin. Pathological Image Super-resolution Reconstruction Based on Sparse Coding Non-local AttentionDual Network [J]. Computer Science, 2023, 50(6A): 220700016-8.
[7] YIN Haitao, WANG Tianyou. Image Denoising Algorithm Based on Deep Multi-scale Convolution Sparse Coding [J]. Computer Science, 2023, 50(4): 133-140.
[8] YUAN Zhen, LIU Jinfeng. Novel Method for Trash Classification Based on Causal Inference [J]. Computer Science, 2023, 50(11A): 220800218-6.
[9] LIU Nan, ZHANG Fengli, YIN Jiaqi, CHEN Xueqin, WANG Ruijin. Rumor Detection Model on Social Media Based on Contrastive Learning with Edge-inferenceAugmentation [J]. Computer Science, 2023, 50(11): 49-54.
[10] HAO Jingyu, WEN Jingxuan, LIU Huafeng, JING Liping, YU Jian. Deep Disentangled Collaborative Filtering with Graph Global Information [J]. Computer Science, 2023, 50(1): 41-51.
[11] YUAN De-sen, LIU Xiu-jing, WU Qing-bo, LI Hong-liang, MENG Fan-man, NGAN King-ngi, XU Lin-feng. Visual Question Answering Method Based on Counterfactual Thinking [J]. Computer Science, 2022, 49(12): 229-235.
[12] XU Jia-qing, WAN Wen, LV Qi. Classification of Hyperspectral Remote Sensing Imagery Based on Second Order Moment Sparse Coding [J]. Computer Science, 2018, 45(9): 288-293.
[13] XU Yu-ming, SONG Jia-wei and XIAO Xian-jian. Super Resolution Algorithm Based on Sub-pixel Block Matching and Dictionary Learning [J]. Computer Science, 2016, 43(8): 304-308.
[14] ZHANG Xu, JIANG Jian-guo, HONG Ri-chang and DU Yue. Image Classification Algorithm Based on Low Rank and Sparse Decomposition and Collaborative Representation [J]. Computer Science, 2016, 43(7): 83-88.
[15] SHI Jing-lan, CHANG Kan, ZHANG Zhi-yong and QIN Tuan-fa. Coefficient-similarity-based Dictionary Learning Algorithm for Face Recognition [J]. Computer Science, 2016, 43(6): 298-302.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!