Implementation code of CascadeBERT: Accelerating Inference of Pre-trained Language Models via Calibrated Complete Models Cascade, Findings of EMNLP 2021
We recommend using Anaconda for setting up the environment of experiments:
git clone https://github.com/lancopku/CascadeBERT.git
cd CascadeBERT
conda create -n cascadebert python=3.7
conda activate cascadebert
conda install pytorch torchvision cudatoolkit=11.0 -c pytorch
pip install -r requirements
We provide the training data with associated data difficulty for a 2L BERT-Complete model.
You can download it from Google Drive , and 2L BERT-Complete model can be downloaded from Google Drive
We provide a sample running script for MRPC, unzip the downloaded data and model, modify the PATH in the glue_mrpc.sh
, and
sh glue_mrpc.sh
You can obtain results in the saved_models
path.
If you have any problems, raise a issue or contact Lei Li
If you find this repo helpful, we'd appreciate it a lot if you can cite the corresponding paper:
@inproceedings{li2021cascadebert,
title = "{C}ascade{BERT}: Accelerating Inference of Pre-trained Language Models via Calibrated Complete Models Cascade",
author = "Li, Lei and
Lin, Yankai and
Chen, Deli and
Ren, Shuhuai and
Li, Peng and
Zhou, Jie and
Sun, Xu",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
year = "2021",
url = "https://aclanthology.org/2021.findings-emnlp.43",
pages = "475--486",
}