AutoGL is graph learning framework with automatic machine learning techniques. AutoGL now mainly focus on node classification problems, but it's easy to apply this program to other graph learning problems.
AutoGL is the 6th solution for AutoGraph Challenge@KDD'20, the competition rules can be found here. We achieve 1st, 4th, 1st, 6th and 27th on 5 final phase datasets.
# | Dataset1 | Dataset2 | Dataset3 | Dataset4 | Dataset5 | Avg |
---|---|---|---|---|---|---|
rank | 1 | 4 | 1 | 6 | 27 | 7.8 |
Clone this repository to your machine:
git clone https://github.com/JunweiSUN/AutoGL.git
Download datasets from here. You can also create your own datasets with required format.
When the download process finished, unzip the datasets and move them to the data
folder. Or you can just simple use the demo dataset in data
.
AutoGL could be easily started with docker:
cd path/to/AutoGL/
docker run --gpus=0 -it --rm -v "$(pwd):/app/autograph" -w /app/autograph nehzux/kddcup2020:v2
python run_local_test.py --dataset_dir=./data/demo --code_dir=./code_submission
You can change the argument dataset_dir to other datasets. On the other hand, you can also modify the directory containing your other sample code.
You can also use your own python environment to run this program. In this way, you must install all the necessary packages. So we recommend users to run this program with docker.
We refer to these packages and codes when developing this program:
nni: An open source AutoML toolkit from microsoft
AutoDL (tabular part): Automated Deep Learning without ANY human intervention
pytorch_geometric: Geometric Deep Learning Extension Library for PyTorch
sparsesvd: a fast library for sparse Singular Value Decomposition
DropEdge: a Pytorch implementation of paper: DropEdge: Towards Deep Graph Convolutional Networks on Node Classification
If you have any question or advice, please feel free to contact our team members:
Junwei Sun: [email protected]
Ruifeng Kuang: [email protected]
Wei Huang: [email protected]
Changrui Mu: [email protected]
Jiayan Wang: [email protected]