From 8882d32998cd7360cf40e5d36498fc5dd5917b18 Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Wed, 31 Jan 2024 10:13:17 -0500 Subject: [PATCH] no need to set the lora_model_dir on resume --- src/axolotl/train.py | 4 ---- 1 file changed, 4 deletions(-) diff --git a/src/axolotl/train.py b/src/axolotl/train.py index d0c7658814..d0f58bca95 100644 --- a/src/axolotl/train.py +++ b/src/axolotl/train.py @@ -72,10 +72,6 @@ def train( ) resume_from_checkpoint = cfg.resume_from_checkpoint - if cfg.adapter and cfg.resume_from_checkpoint and not cfg.lora_model_dir: - LOG.info(f"setting lora_model_dir to use {cfg.resume_from_checkpoint}") - cfg.lora_model_dir = cfg.resume_from_checkpoint - # Load the model and tokenizer msg = "loading model" if cfg.adapter: