From dce3c003ebe10f8daefd92ef044e37f64b672083 Mon Sep 17 00:00:00 2001 From: Yu Chin Fabian Lim Date: Sun, 13 Oct 2024 15:08:18 +0000 Subject: [PATCH] update readme Signed-off-by: Yu Chin Fabian Lim --- plugins/accelerated-peft/README.md | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/plugins/accelerated-peft/README.md b/plugins/accelerated-peft/README.md index fc2cf62..8fb1afb 100644 --- a/plugins/accelerated-peft/README.md +++ b/plugins/accelerated-peft/README.md @@ -6,8 +6,8 @@ Currently only supports LoRA-related techniques, but more are in the pipeline to Plugin | Description | Depends | Loading | Augmentation | Callbacks --|--|--|--|--|-- -[autogptq](./src/fms_acceleration_peft/framework_plugin_autogptq.py) | Loads 4bit GPTQ-LoRA with quantized GPTQ as base | AutoGPTQ | ✅ | ✅ -[bnb](./src/fms_acceleration_peft/framework_plugin_bnb.py) | Loads 4bit QLoRA with quantized bitsandbytes Linear4 | Huggingface
bitsandbytes | ✅ | ✅ +[autogptq](./src/fms_acceleration_peft/framework_plugin_autogptq.py) | Loads 4bit GPTQ-LoRA with quantized GPTQ as base | AutoGPTQ | ✅ | ✅ | ✅ +[bnb](./src/fms_acceleration_peft/framework_plugin_bnb.py) | Loads 4bit QLoRA with quantized bitsandbytes Linear4 | Huggingface
bitsandbytes | ✅ | ✅ | ✅ ### Key Points @@ -43,6 +43,7 @@ GPTQ-LORA depends on an AutoGPTQ backend to run. There are 2 backend options ## Known Issues + - GPTQ-LORA sometimes observed to have `nan` grad norms in the begining of training, but training proceeds well otherwise. -- `low_cpu_mem_usage` temporarily disabled for AutoGPTQ until bug with `make_sure_no_tensor_in_meta_device` is resolved.