Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

load_valuehead_params 没有那个文件value_head.bin #30

Closed
18600709862 opened this issue Jun 13, 2023 · 2 comments
Closed

load_valuehead_params 没有那个文件value_head.bin #30

18600709862 opened this issue Jun 13, 2023 · 2 comments
Labels
solved This problem has been already solved

Comments

@18600709862
Copy link

请问大拿,如下操作报错,有遇到这种情况的吗
LLaMA train
(持续)预培训
CUDA_VISIBLE_DEVICES=0 python src/train_pt.py
--model_name_or_path path_to_llama_model
--do_train
--dataset wiki_demo
--finetuning_type lora
--output_dir path_to_pt_checkpoint
--overwrite_cache
--per_device_train_batch_size 4
--gradient_accumulation_steps 4
--lr_scheduler_type cosine
--logging_steps 10
--save_steps 1000
--learning_rate 5e-5
--num_train_epochs 3.0
--plot_loss
--fp16

目录 path_to_pt_checkpoint 没有文件value_head.bin
当训练rw模型时
CUDA_VISIBLE_DEVICES=0 python src/train_rm.py
--model_name_or_path path_to_llama_model
--do_train
--dataset comparison_gpt4_en
--finetuning_type lora
--checkpoint_dir path_to_pt_checkpoint
--output_dir path_to_rm_checkpoint
--per_device_train_batch_size 4
--gradient_accumulation_steps 4
--lr_scheduler_type cosine
--logging_steps 10
--save_steps 1000
--learning_rate 1e-5
--num_train_epochs 1.0
--plot_loss
--fp16
报错这个目录path_to_pt_checkpoint下面没有这个文件value_head.bin
从来没有见过这个文件啊

@hiyouga
Copy link
Owner

hiyouga commented Jun 13, 2023

该问题已修复,谢谢反馈。

@hiyouga
Copy link
Owner

hiyouga commented Jun 13, 2023

请使用 git pull 更新代码。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
solved This problem has been already solved
Projects
None yet
Development

No branches or pull requests

2 participants