Computer Science ›› 2026, Vol. 53 ›› Issue (4): 245-251.doi: 10.11896/jsjkx.250700069

• Database & Big Data & Data Science • Previous Articles     Next Articles

Phase-preserved MinMax Framework for Graph Augmentation in Frequency Domain

HUA Yu, ZHOU Xiaocheng, SHEN Xiangjun, LIU Zhifeng, ZHOU Conghua   

  1. School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, Jiangsu 212013, China
  • Received:2025-07-14 Revised:2025-09-05 Online:2026-04-15 Published:2026-04-08
  • About author:HUA Yu,born in 2002,postgraduate.His main research interests include graph neural networks and deep learning.
    SHEN Xiangjun,born in 1977,Ph.D,professor,is a member of CCF(No.E200029891M).His main research interests include cross multimedia analysis,computer vision,pattern recognition and statistical machine learning.
  • Supported by:
    National Natural Science Foundation of China(62376108).

Abstract: Graph data augmentation enhances the generalization and robustness of graph neural networks(GNNs) by performing local or global transformations on graph structures or node features.While existing studies have shown that graph augmentation techniques can effectively leverage low-frequency information to capture the global topology of graphs,they often fail to preserve high-frequency components that encode fine-grained structural details.This shortcoming may result in information loss or feature distortion when learning local representations.To address this challenge,this paper proposes a phase-preserving frequency-domain MinMax framework for graph augmentation.The proposed method integrates frequency-domain analysis with the MinMax optimization paradigm,decomposing graph signals into low and high-frequency components.The low-frequency part captures global topological patterns,whereas the high-frequency part represents rich local structural information.By applying the MinMax strategy in the frequency domain,the proposed framework simultaneously preserves global structure and enhances high-frequency details,leading to more expressive multi-scale graph representations.In addition,it adopts an adaptive augmentation strategy that dynamically adjusts the perturbation amplitude based on the characteristics of different frequency components,thereby improving training efficiency.The phase information,which encodes intrinsic structural relations between graph nodes,is explicitly preserved to further enrich the expressive capacity of node representations.Through this frequency-aware design,the proposed method maintains essential topological structures while effectively enhancing node-level features,improving the GNN’s ability to capture both global and local semantics.Extensive experiments on multiple benchmark datasets demonstrate that the proposed method achieves over a 2 percentage points accuracy gain on node classification tasks compared to existing approaches.Moreover,it deli-vers superior computational efficiency,validating its effectiveness and scalability for large-scale graph learning scenarios.

Key words: Graph augmentation, Phase preservation, Frequency domain MinMax framework, Topological optimization, Amplitude adjustment

CLC Number: 

  • TP181
[1]ZHANG Y,XIONG S,WANG Z,et al.Local augmented graph neural network for multi-omics cancer prognosis prediction and analysis[J].Methods,2023,213:1-9.
[2]ZHOU J,XIE C,GONG S,et al.Data augmentation on graphs:a technical survey[J].arXiv:2212.09970,2022.
[3]WEI Z,XIAO X,ZHANG B,et al.Graph Data Augmentationfor Node Classification[C]//Proceedings of the 2022 26th International Conference on Pattern Recognition(ICPR).2022.
[4]ALI A,LI J.Features based adaptive augmentation for graph contrastive learning[J].Digital Signal Processing,2024,145:104312.
[5]XU Y,WANG J,GUANG M,et al.Graph contrastive learningwith min-max mutual information[J].Information Sciences,2024,665:120378.
[6]ZHOU X C,SHEN X J.Graph Augmentation Method Based on Fourier Spectrum Perception[J].Journal of Chinese Computer Systems,2025,46(11):2667-2673.
[7]LIU Y,DU B.Frequency Domain-Oriented Complex GraphNeural Networks for Graph Classification[J].IEEE Transactions on Neural Networks and Learning Systems,2025,36(2):2733-2746.
[8]SUN Y,DUAN Y,MA H,et al.High-frequency and low-frequency dual-channel graph attention network[J].Pattern Re-cognition,2024,156:110795.
[9]LIANG Z,GONG R,TAN G,et al.A Frequency Domain Kernel Function-Based Manifold Dimensionality Reduction and Its Application for Graph-Based Semi-Supervised Classification[J].Applied Sciences,2024,14(12):5342.
[10]TANG X,YAN J.A CSI Amplitude-phase Information Based Graph Construction for GNN Localization[C]//Proceedings of the 2023 International Conference on Computer,Information and Telecommunication Systems(CITS).2023.
[11]ZHOU Z,HU Y,ZHANG Y,et al.Multiview Deep Graph Infomax to Achieve Unsupervised Graph Embedding[J].IEEE Transactions on Cybernetics,2023,53(10):6329-6339.
[12]XIE Y,LUO L,CAO T,et al.Contrastive Learning Network for Unsupervised Graph Matching[J].IEEE Transactions on Circuits and Systems for Video Technology,2025,35(1):643-656.
[13]REN Y,BAI J,ZHANG J.Label Contrastive Coding basedGraph Neural Network for Graph Classification[C]//Procee-dings of the International Conference on Database Systems for Advanced Applications.2021.
[14]AN D,PAN Z,ZHAO Q,et al.Unsupervised Graph Structure Learning Based on Optimal Graph Topology Modeling and Adaptive Data Augmentation[J].Mathematics,2024,12(13):1991.
[15]XU X,ZHAO H.Universal adaptive data augmentation[J].arXiv:2207.06658,2022.
[16]ZHUO W,TAN G.Graph contrastive learning with adaptiveproximity-based graph augmentation[J].IEEE Transactions on Neural Networks and Learning Systems,2024,35(10):14.
[17]LI X,WANG Y,WANG Y,et al.Graph contrastive learning for recommendation with generative data augmentation[J].Multimedia Systems,2024,30(4):170.
[18]XIA J,WU L,WANG G,et al.Progcl:Rethinking hard negative mining in graph contrastive learning[J].arXiv:2110.02027,2021.
[19]BAGHERI BARDI A,YAZDANPANAH T,DAKOVIC M,et al.Graph Fourier Transform Enhancement through Envelope Extensions[J].arXiv:2407.19934,2024.
[20]ORTEGA Y R,GUERREIRO I M,HUI D,et al.Supervisedlearning and graph signal processing strategies for beam trac-king in highly directional mobile communications[J].Transactions on Emerging Telecommunications Technologies,2019,30(9):e3687.
[21]NEUMEIER M,TOLLKÜHN A,BOTSCH M,et al.A multidimensional graph fourier transformation neural network for ve-hicle trajectory prediction[C]//Proceedings of the 2022 IEEE 25th International Conference on Intelligent Transportation Systems(ITSC).IEEE,2022.
[22]WANG P,WEN Z.A spatio-temporal graph wavelet neural network(ST-GWNN) for association mining in timely social media data[J].Scientific Reports,2024,14(1):31155.
[23]DEB S,RAHMAN S,RAHMAN S.GA-GWNN:GeneralizedAdaptive Graph Wavelet Neural Network[J].Pattern Recognition Letters,2024,177:128-34.
[24]CHAN V,GAN Q,BAYEN A.A Graph Convolutional Network with Signal Phasing Information for Arterial Traffic Prediction[J].arXiv:2012.13479,2020.
[25]XU Z,ZHANG B,LI G,et al.Analysis of phase preservationand interferometric offset test in sparse SAR imaging[J].Science China Information Sciences,2024,67(2):122303.
[26]HUANG H,KANG H,LIU S,et al.Paddles:Phase-amplitudespectrum disentangled early stopping for learning with noisy labels[C]//Proceedings of the IEEE/CVF International Confe-rence on Computer Vision.2023.
[27]JIAO Z,ZHANG H,LI X.Deep Graph Multi-View Representation Learning With Self-Augmented View Fusion[J].IEEE Transactions on Neural Networks and Learning Systems,2025,36(8):14119-14130.
[1] CHEN Hongxiu, ZENG Xia, LIU Zhiming, ZHAO Hengjun. Automatic Theorem Proving Based on Pre-trained Language Models and Unification [J]. Computer Science, 2026, 53(4): 40-47.
[2] WU Yansheng, CAO Xinyi, FAN Weibei. Research on Efficient Construction of Plateaued Functions Based on DQN-enhanced Genetic Algorithm [J]. Computer Science, 2026, 53(4): 57-65.
[3] HUANG Beibei, LIU Jinfeng. Causal Disentangled Representation Learning with Integrated Sparse Coding [J]. Computer Science, 2026, 53(4): 66-77.
[4] ZHANG Xueqin, WANG Zhineng, LI Jinsheng, LU Yisong, LUO Fei. Key Node Identification in Temporal Social Networks Based on Deep Learning and Multi-feature Fusion [J]. Computer Science, 2026, 53(4): 143-154.
[5] QIN Haiqi, MI Jusheng. Concept-cognitive Learning and Incremental Learning in Complex Networks [J]. Computer Science, 2026, 53(4): 208-214.
[6] LIU Jiaqi, WANG Yujie, XIANG Guodu, YU Kui, CAO Fuyuan. Long-term Causal Effect Estimation Based on Deep Reinforcement Learning [J]. Computer Science, 2026, 53(4): 235-244.
[7] PAN Jiahao, FENG Xiang, YU Huiqun. SM-PHT:Robust,Scalable,and Efficient Method for Multi-task Reinforcement Learning [J]. Computer Science, 2026, 53(4): 366-376.
[8] GE Zeqing, HUANG Shengjun. Semi-supervised Learning Method for Multi-label Tabular Data [J]. Computer Science, 2026, 53(3): 151-157.
[9] WANG Yiming, JIAO Min, ZHAO Suyun, CHEN Hong, LI Cuiping. Prompt-conditioned Representation Learning with Diffusion Models for Semi-supervised Clustering [J]. Computer Science, 2026, 53(3): 158-165.
[10] ZHAO Binbei, ZHU Li, ZHAO Hongli, LI Yutong. Computer Vision Applications in Rail Transit Systems [J]. Computer Science, 2026, 53(3): 214-224.
[11] JIA Shuheng, FU Huimin. Optimizing Probabilistic Choice for Solving SAT Problems [J]. Computer Science, 2026, 53(3): 366-374.
[12] HUANG Miaomiao, WANG Huiying, WANG Meixia, WANG Yejiang , ZHAO Yuhai. Review of Graph Embedding Learning Research:From Simple Graph to Complex Graph [J]. Computer Science, 2026, 53(1): 58-76.
[13] WANG Haoyan, LI Chongshou, LI Tianrui. Reinforcement Learning Method for Solving Flexible Job Shop Scheduling Problem Based onDouble Layer Attention Network [J]. Computer Science, 2026, 53(1): 231-240.
[14] DUAN Pengting, WEN Chao, WANG Baoping, WANG Zhenni. Collaborative Semantics Fusion for Multi-agent Behavior Decision-making [J]. Computer Science, 2026, 53(1): 252-261.
[15] ZENG Dan, HE Xingxing, LI Yingfang, LI Tianrui. Structures of Multi-line Standard Contradictions in First-order Logic [J]. Computer Science, 2025, 52(12): 200-208.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!