MarkTechPost@AI 2024年12月26日
CLDG: A Simple Machine Learning Framework that Sets New Benchmarks in Unsupervised Learning on Dynamic Graphs
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文介绍了西安交通大学提出的CLDG框架,一种针对动态图的无监督对比学习方法。该框架通过时间跨度视图采样从动态图中提取信息,利用共享编码器学习节点和邻域的特征表示,并设计了局部和全局对比损失以保持时间平移不变性。CLDG在多个真实动态图数据集上表现优异,超越了现有无监督方法,并能与半监督方法相媲美,同时显著降低了模型参数和训练时间。该框架具有轻量级和可扩展性强的特点,为动态图的无监督学习提供了新的思路。

⏱️ CLDG框架通过时间跨度视图采样方法,从连续动态图中生成多个视图,提取时间上持续存在的信号,为动态图的对比学习奠定基础。

💡 该框架利用共享权重的编码器、读出函数和投影头,学习节点和邻域的特征表示,这种设计提高了模型的效率和可扩展性。

⚖️ CLDG设计了局部和全局两种对比损失,以保持时间平移不变性。局部损失使同一节点在不同时间跨度的表示更接近,而全局损失则将不同节点区分开,从而优化模型性能。

📊 CLDG在七个真实世界动态图数据集上进行了验证,并与十二个基线方法进行了比较。结果表明,该方法优于八个无监督的基线方法,并与四个半监督方法相当,同时显著降低了模型参数和训练时间。

Graph Neural Networks have emerged as a transformative force in many real-life applications, from corporate finance risk management to local traffic prediction. Thereby, there is no gainsaying that much research has been centered around GNNs for a long time. A significant limitation of the current study, however, is its data dependency—with a focus on supervised and semi-supervised paradigms, the investigation’s potential depends on the availability of ground truth, a requirement often unmet. Another reason for the sparsity of actual labels is the inherent nature of GNNs themselves. Since a graph is an abstraction of the real world, it is not as straightforward as video, image, or text, requiring expert knowledge and experience.

With the prevailing challenges and increasing expenses to solve supervised graph paradigms, researchers have begun a pivot toward unsupervised contrastive learning. It works based on mutual information between different augmented graph views generated by perturbing its nodes, edges, and features. Although this approach is promising and eliminates the necessity of labels, it is not always possible to confirm if the labels and semantics remain unchanged post-augmentation, significantly undermining the graphs’ performance. To understand the detrimental effects of augmentation, let’s take the example of a node. One could add or delete a node in the existing graph, which either adds noise or removes information, both detrimental. Therefore, existing static graph contrastive learning methods may not be optimal for dynamic graphs. This article discusses the latest research that claims to generalize contrastive learning to dynamic graphs.

Researchers from Xi’an Jiaotong University, China, presented CLDG, an efficient unsupervised Contrastive Learning framework on the Dynamic Graph, which performs representation learning on discrete and continuous-time dynamic graphs. It solves the dilemma of selecting periods as contrastive pairs while applying contrastive learning to dynamic graphs. CLDG is a light and highly scalable algorithm, credit due to its simplicity. Users get lower time and space complexity and the opportunity to choose from a pool of encoders.

The proposed framework consists of five major components:

    timespan view sampling layerbase encoderreadout functionprojection headcontrastive loss function

The research team first generated multiple views from continuous dynamic graphs via a timespan view sampling method. Here, the view sampling layer extracts the temporally persistent signals. They then learned the feature representations of nodes and neighborhoods through a weight-shared encoder, a readout function, and a weight-shared projection head. The authors used statistical-based methods such as average, maximum, and summation for the readout function layer.

An important insight to discuss at this point is temporal translation invariance. Under this, it is observed that regardless of the encoder used for training, the prediction labels of the same node tend to be similar in different time spans. The paper presented two separate local-level and global-level contrastive losses to maintain temporal translation invariance at both levels. In local-level temporal translation invariance, semantics were treated as positive pairs for one node across time spans, which pulled the same node representations closer and different nodes apart. Conversely, loss for global invariance pulled different nodes together and the same representation away. Following the above, the authors designed four different timespan view sampling strategies to explore the optimal view interval distance selection for contrastive pairs. These strategies differed in the physical and temporal overlap rate and thereby had different semantic contexts.

The paper validated CLDG on seven real-world dynamic graph datasets and across twelve baselines. The proposed method outperformed eight unsupervised state-of-the-art baselines and was on par with the remaining four semi-supervised methods. Furthermore, compared to existing graph methods, CLDG reduced model parameters by an average of 2000 times and the training time by 130.

Conclusion: CLDG is a practical, lightweight framework that generalizes contrastive learning to dynamic graphs. It uses additional temporal information and achieves state-of-the-art performance in unsupervised dynamic graph techniques while competing with semi-supervised methods.


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post CLDG: A Simple Machine Learning Framework that Sets New Benchmarks in Unsupervised Learning on Dynamic Graphs appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

图神经网络 对比学习 动态图 无监督学习 CLDG
相关文章