-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
deepspeed zero3 QLoRA finetuning (#11625)
* deepspeed zero3 QLoRA finetuning * Update convert.py * Update low_bit_linear.py * Update utils.py * Update qlora_finetune_llama2_13b_arch_2_card.sh * Update low_bit_linear.py * Update alpaca_qlora_finetuning.py * Update low_bit_linear.py * Update utils.py * Update convert.py * Update alpaca_qlora_finetuning.py * Update alpaca_qlora_finetuning.py * Update low_bit_linear.py * Update deepspeed_zero3.json * Update qlora_finetune_llama2_13b_arch_2_card.sh * Update low_bit_linear.py * Update low_bit_linear.py * Update utils.py * fix style * fix style * Update alpaca_qlora_finetuning.py * Update qlora_finetune_llama2_13b_arch_2_card.sh * Update convert.py * Update low_bit_linear.py * Update model.py * Update alpaca_qlora_finetuning.py * Update low_bit_linear.py * Update low_bit_linear.py * Update low_bit_linear.py
- Loading branch information
Showing
5 changed files
with
119 additions
and
14 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
15 changes: 15 additions & 0 deletions
15
python/llm/example/GPU/LLM-Finetuning/QLoRA/alpaca-qlora/deepspeed_zero3.json
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,15 @@ | ||
{ | ||
"zero_optimization": { | ||
"stage": 3, | ||
"contiguous_gradients": true, | ||
"overlap_comm": true, | ||
"offload_optimizer": {"device": "cpu"} | ||
}, | ||
"bf16": { | ||
"enabled": true | ||
}, | ||
"world_size": 2, | ||
"train_batch_size": 32, | ||
"train_micro_batch_size_per_gpu": 2, | ||
"gradient_accumulation_steps": 8 | ||
} |
41 changes: 41 additions & 0 deletions
41
...lm/example/GPU/LLM-Finetuning/QLoRA/alpaca-qlora/qlora_finetune_llama2_13b_arch_2_card.sh
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,41 @@ | ||
# | ||
# Copyright 2016 The BigDL Authors. | ||
# | ||
# Licensed under the Apache License, Version 2.0 (the "License"); | ||
# you may not use this file except in compliance with the License. | ||
# You may obtain a copy of the License at | ||
# | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# | ||
# Unless required by applicable law or agreed to in writing, software | ||
# distributed under the License is distributed on an "AS IS" BASIS, | ||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
# See the License for the specific language governing permissions and | ||
# limitations under the License. | ||
# | ||
|
||
export MASTER_ADDR=127.0.0.1 | ||
export MASTER_PORT=29503 | ||
export FI_PROVIDER=tcp | ||
export CCL_ATL_TRANSPORT=ofi | ||
export CCL_ZE_IPC_EXCHANGE=sockets | ||
export UR_L0_IN_ORDER_BARRIER_BY_SIGNAL=0 | ||
basekit_root=/opt/intel/oneapi | ||
source $basekit_root/setvars.sh --force | ||
source $basekit_root/ccl/latest/env/vars.sh --force | ||
|
||
NUM_GPUS=2 # number of used GPU | ||
export USE_XETLA=OFF | ||
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=2 | ||
export TORCH_LLM_ALLREDUCE=0 # Different from PVC | ||
export DS_SKIP_CUDA_CHECK=1 | ||
|
||
mpirun -n $NUM_GPUS \ | ||
python -u ./alpaca_qlora_finetuning.py \ | ||
--base_model "meta-llama/Llama-2-13b-hf" \ | ||
--data_path "yahma/alpaca-cleaned" \ | ||
--output_dir "./ipex-llm-qlora-alpaca" \ | ||
--gradient_checkpointing True \ | ||
--micro_batch_size 2 \ | ||
--batch_size 32 \ | ||
--deepspeed ./deepspeed_zero3.json |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters