Skip to content

Open Source Code for GAugLLM: Improving Graph Contrastive Learning for Text-Attributed Graphs with Large Language Models (KDD'24)

License

Notifications You must be signed in to change notification settings

NYUSHCS/GAugLLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

46 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GAugLLM

Official code for "GAugLLM: Improving Graph Contrastive Learning for Text-Attributed Graphs with Large Language Models". GAugLLM is a novel framework for augmenting TAGs, leveraging advanced large language models like Mistral to enhance self-supervised graph learning.

Pipeline of the GAugLLM

The learning paradigm of GAugLLM vs. traditional GCL methods on TAGs. While standard GCL methodologies rely on text attributes primarily to generate numerical node features via shallow embedding models, such as word2vec, our GAugLLM endeavors to advance contrastive learning on graphs through advanced LLMs. This includes the direct perturbation of raw text attributes for feature augmentation, facilitated by a novel mixture-of-prompt experts technique. Additionally, GAugLLM harnesses both structural and textual commonalities to effectively perturb edges deemed most spurious or likely to be connected, thereby enhancing structure augmentation.

img

🚀Quick Start

For Mix-of-Experts-Prompt part, please check LLMs folder. First you should follow GIANT:Node Feature Extraction by Self-Supervised Multi-scale Neighborhood Prediction (https://arxiv.org/abs/2102.06514) to set up and update our modifications. architecture

For Collaborative Edge Modifier part, we add some modification on the original GCL frameworks of BGRL, GBT, GraphMAE and S2GAE. For GraphCL we used its loss function design in GBT framework.

Mix-of-Expert-Prompt

You can use run.sh in LLM folder to run Mix-of-Expert-Prompt over GIANT framework.

Collaborative Edge Modifier

To run original and GAugLLM on GCLs respectively, please use sripts with original or GAug in their names respectively.

Citation

If you find this repo useful, please star the repo and cite:

@article{fang2024gaugl,
  title={GAugLLM: Improving Graph Contrastive Learning for Text-Attributed Graphs with Large Language Models},
  author={Yi Fang and Dongzhe Fan and Daochen Zha and Qiaoyu Tan},
  booktitle={Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery \& Data Mining},
  year={2024},
  url={https://arxiv.org/abs/2406.11945}
}

About

Open Source Code for GAugLLM: Improving Graph Contrastive Learning for Text-Attributed Graphs with Large Language Models (KDD'24)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published