From 848fa04dd65ae55f08fdced7d63a6fbbf0248c5c Mon Sep 17 00:00:00 2001 From: "Keyan (Kyrie) Zhang" <79576162+Zhangky11@users.noreply.github.com> Date: Thu, 28 Mar 2024 22:31:47 -0700 Subject: [PATCH] Fix typo in Baichuan2 example (#10589) --- .../llm/example/GPU/PyTorch-Models/Model/baichuan2/generate.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/llm/example/GPU/PyTorch-Models/Model/baichuan2/generate.py b/python/llm/example/GPU/PyTorch-Models/Model/baichuan2/generate.py index 9ebb56be847..a45926796d0 100644 --- a/python/llm/example/GPU/PyTorch-Models/Model/baichuan2/generate.py +++ b/python/llm/example/GPU/PyTorch-Models/Model/baichuan2/generate.py @@ -24,7 +24,7 @@ # prompt format referred from https://github.com/baichuan-inc/Baichuan2/issues/227 # and https://huggingface.co/baichuan-inc/Baichuan2-7B-Chat/blob/main/generation_utils.py#L7-L49 # For English prompt, you are recommended to change the prompt format. -BAICHUAN_PROMPT_FORMAT = " {prompt} " +BAICHUAN2_PROMPT_FORMAT = " {prompt} " if __name__ == '__main__': parser = argparse.ArgumentParser(description='Predict Tokens using `generate()` API for Baichuan2 model')