- Install Conda: https://docs.conda.io/en/latest/miniconda.html
- Found Your Conda environment:
conda create -n LsMbert python=3.8
conda activate LsMbert
pip install -r requirements.txt
The next step is to download the data. To this end, first create a download
folder with mkdir -p download
in the root
of this project. You then need to manually download panx_dataset
(for NER) from here
(note that it will download as AmazonPhotos.zip
) to the download directory. Finally, run the following command to
download the remaining datasets:
bash scripts/download_data.sh
To get the POS-tags and dependency parse of input sentences, we use UDPipe. Go to the
udpipe directory and run the task-specific scripts -
[xnli.sh|pawsx.sh|panx.sh|mtop.sh]
.
Notice:Data Preparation
is same as here
You can download files from here,
and put it to file directory.
Or, you can create it yourself with your own data.
# for PAWS-X
sh run_pawsx.sh
# for XNLI
sh run_xnli.sh
sh run_ner.sh
sh mtop.sh
Take mtop
as an example.
First: Download the relevant data from Baidu link here
Second: Download the model from Baidu link here
export CUDA_VISIBLE_DEVICES=0
Output_dir="./outputs/mtop_paper"
python mtop_paper.py \
--data_dir "./download/mtop_udpipe_processed" \
--model_name_or_path "./outputs/mtop_model" \
--intent_labels "./download/mtop_udpipe_processed/intent_label.txt" \
--slot_labels "./download/mtop_udpipe_processed/slot_label.txt" \
--do_test \
--train_langs "en" \
--output_dir $Output_dir
The test of other task , or the other ideas if you have ,You can try it yourself.