Computer Science ›› 2025, Vol. 52 ›› Issue (11A): 250100059-8.doi: 10.11896/jsjkx.250100059

• Artificial Intelligence • Previous Articles     Next Articles

ZHA_TGCN:A Topic Classification Method for Low-resource Sawcuengh Language

ZHAO Zhuoyang1, QIN Donghong1,4, BAI Fengbo1,4, LIANG Xianye1, XU Chen1, ZHENG Yuehua1, LIANG Yufeng1, LAN Sheng2,4, ZHOU Guoping3   

  1. 1 College of Artificial Intelligence,Guangxi Minzu University,Nanning 530006,China
    2 College of Chinese Language and Literature,Guangxi Minzu University,Nanning 530006,China
    3 College of Preparatory Education,GuangxiMinzu University,Nanning 530006,China
    4 University Engineering Research Center of Computational Linguistics and Intelligence,Nanning 530006,China
  • Online:2025-11-15 Published:2025-11-10
  • Supported by:
    Central Guidance on Local Science and Technology Development Fund of Guangxi Zhuang Autonomous Region(GUIKEZY24212045),Guangxi Science and Technology Base and Talent Project(GUIKEAD23026054) and Guangxi Science and Technology Base and Talent Project(GUIKEAD22035200).

Abstract: Traditional graph convolutional network methods can effectively model graph structures under data-limited conditions.However,due to their reliance on sparse one-hot encoding,they face limitations in capturing the contextual relationships between words.This issue is particularly pronounced in low-resource language environments.Taking the Sawcuengh language text topic classification task as an example,this task faces not only data scarcity but also the challenge of complex linguistic structures.To address these challenges,this paper proposes a Sawcuengh language topic classification method suitable for low-resource settings-ZHA_TGCN.This method leverages the Sawcuengh pre-trained model,ZHA_BERT,to extract textual features,and combines these features with Sawcuengh tone features.These combined features are then input into a BiGRU to learn deep semantic representations.The learned representation vectors are used as node features for the GCN,which propagates labels to learn the feature representations of both the training data and the unlabeled test data.Finally,a Softmax layer is used to output the classification results.Experimental results show that the proposed method achieves an accuracy of 82.12%,precision of 90.08%,recall of 92.46%,and an F1 score of 90.18% in the low-resource Sawcuengh language topic classification task,demonstrating the effectiveness of the method.

Key words: Low-resource language, Sawcuengh language, Subjet classification, Pre-trained model, Graph convolutional network

CLC Number: 

  • TP391
[1]WANG A H.Don’t follow me:Spam detection in twitter[C]//2010 International Conference on Security and Cryptography(SECRYPT).IEEE,2010:1-10.
[2]PANG B,LEE L.Opinion mining and sentiment analysis[J].Foundations and Trends© in Information Retrieval,2008,2(1/2):1-135.
[3]KARIM A,AZAM S,SHANMUGAM B,et al.A comprehensive survey for intelligent spam email detection[J].IEEE Access,2019,7:168261-168295.
[4]LI J Y.From Ancient Zhuang Characters to Zhuang Script:The Zhuang People Now Have Their Own Writing System [J].Contemporary Guangxi,2019(Z1):86.
[5]CHURCH K W.Word2Vec[J].Natural Language Engineering,2017,23(1):155-162.
[6]PENNINGTON J,SOCHER R,MANNING C D.Glove:Global vectors for word representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Proces-sing(EMNLP).2014:1532-1543.
[7]LI Z,LIU F,YANG W,et al.A survey of convolutional neural networks:analysis,applications,and prospects [J].IEEE Transactions on Neural Networks and Learning Systems,2021,33(12):6999-7019.
[8]MIKOLOV T,KARAFIÁT M,BURGET L,et al.Recurrentneural network based language model[C]//Interspeech.2010:1045-1048.
[9]ZHANG S,ZHENG D,HU X,et al.Bidirectional long short-term memory networks for relation classification[C]//Procee-dings of the 29th Pacific Asia Conference on Language,Information and Computation.2015:73-78.
[10]ZULQARNAIN M,GHAZALI R,GHOUSE M G,et al.Efficient processing of GRU based on word embedding for text classification[J].International Journal on Informatics Visualization,2019,3(4):377-383.
[11]DEVLIN J,CHANG M W,LEE K,et al.Bert:Pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.2019:4171-4186.
[12]WU L,CHEN Y,SHEN K,et al.Graph neural networks fornatural language processing:A survey [J].Foundations and Trends© in Machine Learning,2023,16(2):119-328.
[13]SCARSELLI F,GORI M,TSOI A C,et al.The graph neural network model [J].IEEE Transactions on Neural Networks,2008,20(1):61-80.
[14]KIPF T N,WELLING M.Semi-Supervised Classification withGraph Convolutional Networks[C]//International Conference on Learning Representations.2017.
[15]YAO L,MAO C,LUO Y.Graph convolutional networks fortext classification[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2019:7370-7377.
[16]LIU X,YOU X,ZHANG X,et al.Tensor graph convolutional networks for text classification[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2020:8409-8416.
[17]LI X,LI Z,SHENG J,et al.Low-resource text classification via cross-lingual language model fine-tuning[C]//China National Conference on Chinese Computational Linguistics.Cham:Springer,2020:231-246.
[18]SAZZED S.Cross-lingual sentiment analysis in bengali utilizing a new benchmark corpus[C]//Proceedings of the 2020 EMNLP Workshop W-NUT:The Sixth Workshop on Noisy User-gene-rate.2020:50-60.
[19]YAO H,WU Y,AL-SHEDIVAT M,et al.Knowledge-Aware Meta-learning for Low-Resource Text Classification[C]//Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing.2021:1814-1821.
[20]FESSEHA A,XIONG S,EMIRU E D,et al.Text classification based on convolutional neural networks and word embedding for low-resource languages:Tigrinya [J].Information,2021,12(2):52.
[21]AN B,ZHAO W N,LONG C J.Low-resource Tibetan TextClassification Based on Prompt Learning[J].Journal of Chinese Information Processing,2024,38(2):70-78.
[22]WEN Z,FANG Y.Augmenting low-resource text classification with graph-grounded pre-training and prompting[C]//Procee-dings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval.2023:506-516.
[23]LIN Y,MENG Y,SUN X,et al.BertGCN:Transductive TextClassification by Combining GNN and BERT[C]//Findings of the Association for Computational Linguistics:ACL-IJCNLP 2021.2021:1456-1462.
[24]YUAN Y,LV S,BAO Z,et al.A Joint Model for Text Classification with BERT-BiLSTM and GCN[C]//Proceedings of the 2022 5th International Conference on Artificial Intelligence and Pattern Recognition.2022:180-186.
[25]HUANG L,MA D,LI S,et al.Text Level Graph Neural Net-work for Text Classification[C]//Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing(EMNLP-IJCNLP).2019:3444-3450.
[26]JOULIN A,GRAVE É,BOJANOWSKI P,et al.Bag of Tricks for Efficient Text Classification[C]//Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics.2017:427-431.
[27]LIU P,QIU X,HUANG X.Recurrent neural network for text classification with multi-task learning[C]//Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence.2016:2873-2879.
[28]JOHNSON R,ZHANG T.Deep pyramid convolutional neural networks for text categorization[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics.2017:562-570.
[29]GUO B,ZHANG C,LIU J,et al.Improving text classification with weighted word embeddings via a multi-channel TextCNN model[J].Neurocomputing,2019,363:366-374.
[1] HU Hailong, XU Xiangwei, LI Yaqian. Drug Combination Recommendation Model Based on Dynamic Disease Modeling [J]. Computer Science, 2025, 52(9): 96-105.
[2] ZHONG Boyang, RUAN Tong, ZHANG Weiyan, LIU Jingping. Collaboration of Large and Small Language Models with Iterative Reflection Framework for Clinical Note Summarization [J]. Computer Science, 2025, 52(9): 294-302.
[3] GAO Long, LI Yang, WANG Suge. Sentiment Classification Method Based on Stepwise Cooperative Fusion Representation [J]. Computer Science, 2025, 52(9): 313-319.
[4] ZHOU Tao, DU Yongping, XIE Runfeng, HAN Honggui. Vulnerability Detection Method Based on Deep Fusion of Multi-dimensional Features from Heterogeneous Contract Graphs [J]. Computer Science, 2025, 52(9): 368-375.
[5] LI Mengxi, GAO Xindan, LI Xue. Two-way Feature Augmentation Graph Convolution Networks Algorithm [J]. Computer Science, 2025, 52(7): 127-134.
[6] SHI Enyi, CHANG Shuyu, CHEN Kejia, ZHANG Yang, HUANG Haiping. BiGCN-TL:Bipartite Graph Convolutional Neural Network Transformer Localization Model for Software Bug Partial Localization Scenarios [J]. Computer Science, 2025, 52(6A): 250200086-11.
[7] BIAN Hui, MENG Changqian, LI Zihan, CHEN Zihaoand XIE Xuelei. Continuous Sign Language Recognition Based on Graph Convolutional Network and CTC/Attention [J]. Computer Science, 2025, 52(6A): 240400098-9.
[8] TANG Lijun , YANG Zheng, ZHAO Nan, ZHAI Suwei. FLIP-based Joint Similarity Preserving Hashing for Cross-modal Retrieval [J]. Computer Science, 2025, 52(6A): 240400151-10.
[9] YE Jiale, PU Yuanyuan, ZHAO Zhengpeng, FENG Jue, ZHOU Lianmin, GU Jinjing. Multi-view CLIP and Hybrid Contrastive Learning for Multimodal Image-Text Sentiment Analysis [J]. Computer Science, 2025, 52(6A): 240700060-7.
[10] FANG Rui, CUI Liangzhong, FANG Yuanjing. Equipment Event Extraction Method Based on Semantic Enhancement [J]. Computer Science, 2025, 52(6A): 240900096-9.
[11] TAN Qiyin, YU Jiong, CHEN Zixin. Outlier Detection Method Based on Adaptive Graph Autoencoder [J]. Computer Science, 2025, 52(6): 129-138.
[12] ZHANG Jiaxiang, PAN Min, ZHANG Rui. Study on EEG Emotion Recognition Method Based on Self-supervised Graph Network [J]. Computer Science, 2025, 52(5): 122-127.
[13] HUANG Qian, SU Xinkai, LI Chang, WU Yirui. Hypergraph Convolutional Network with Multi-perspective Topology Refinement forSkeleton-based Action Recognition [J]. Computer Science, 2025, 52(5): 220-226.
[14] ZHAO Hongyi, LI Zhiyuan, BU Fanliang. Multi-language Embedding Graph Convolutional Network for Hate Speech Detection [J]. Computer Science, 2025, 52(11A): 241200023-8.
[15] HU Jintao, XIAN Guangming. Self-attention-based Graph Contrastive Learning for Recommendation [J]. Computer Science, 2025, 52(11): 82-89.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!