Skip to content

Latest commit

 

History

History
60 lines (42 loc) · 2.59 KB

README.md

File metadata and controls

60 lines (42 loc) · 2.59 KB

PyTorch implementation of Recurrent Event Network (RE-Net)

Paper: TPmod: A Tendency-Guided Prediction Model for Temporal Knowledge Graph Completion

This repository contains the implementation of the TPmod architectures described in the paper.

Installation

Install PyTorch (>= 1.1.0) following the instuctions on the PyTorch . Our code is written in Python3.

Train and Test

Before running, you should preprocess datasets.

python3 data/DATA_NAME/get_history_new.py

Then, we are ready to train and test. We first train the model.

python3 train.py -dataset DATA_NAME -cuda 0 -dim 256 -lr 1e-4 -epochs 50 -b 1024 -dropout 0.5

We are ready to test!

python3 test.py -dataset DATA_NAME -cuda 0 -dim 256 -lr 1e-4 -epochs 50 -b 1024 -dropout 0.5

Datasets

There are four datasets: two with TGvals: GDELT-5 and ICEWS-250. and two with IGvals: GDELT-5I and ICEWS-250I. Each data folder has 'stat.txt', 'train.txt', 'valid.txt', 'test.txt', 'rel2val.txt' and 'get_history_new.py'.

  • 'get_history_new.py': This is for getting history.
  • 'rel2val.txt' : This is the file mapping relations to TGvals (IGvals)
  • 'stat.txt': First value is the number of entities, and second value is the number of relations.
  • 'train.txt', 'valid.txt', 'test.txt': First column is subject entities, second column is relations, and third column is object entities. The fourth column is time.

Baselines

We use the following public codes for baselines and hyperparameters.

Baselines Code parameters
TransE Link { lr=0.0001, dim=512,b=512}
TTransE link { lr=0.001, dim=512,b=512}
DE-SimplE link { lr=0.001, dim=128,b=512}
TA-DistMult link { lr=0.001, dim=512,b=1024}
RE-Net link { lr=0.001, dim=256,b=1024}

We implemented RESCAL, DistMult refer to [RotatE](: https://github.com/DeepGraphLearning/ KnowledgeGraphEmbedding.). The user can run the baselines by the following command.

cd ./baselines
bash run.sh train MODEL_NAME DATA_NAME 0 0 512 1024 512 200.0 0.0005 10000 8 0

The user can find implementations in the 'baselines' folder.