This folder contains examples of running GraphGPT in local machine with multiple GPUs.
We adopt deepspeed
engine to run the script. In Alibaba Cloud PAI platform,
python -m torch.distributed.launch
is used to run the parallel training.
When new version of code is developed, it has to pass tests to be released. The current procedure is listed below.
- Delete the existing vocab files of the datasets
- Run
ggpt_pretrain.sh
andggpt_supervised.sh
for the datasetsogbn-proteins
ogbl-ppa
,ogbg-molpcba
andPCQM4M-v2
withmini
architecture quickly, make sure no bugs occur. This step does not need to produce the best results. - Repeat step 2, but this time run long enough for each dataset, so that best results can be re-produced.
./examples/toy_examples/ggpt_pretrain.sh
./examples/toy_examples/ggpt_supervised.sh
./examples/graph_lvl/pcqm4m_v2_pretrain.sh
./examples/graph_lvl/pcqm4m_v2_supervised.sh