Skip to content

Commit

Permalink
Extend max_prompt_length and input text for 128k evaluation
Browse files Browse the repository at this point in the history
  • Loading branch information
HeyangQin committed Apr 15, 2024
1 parent 7778d1c commit 25ae45c
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion benchmarks/inference/mii/src/client.py
Original file line number Diff line number Diff line change
Expand Up @@ -275,7 +275,7 @@ def run_client(args):
tokenizer = AutoTokenizer.from_pretrained(args.model)

# make sure max_prompt_length is longer than the target prompt length
args.max_prompt_length = max(args.max_prompt_length, int(args.mean_prompt_length * (1 + args.prompt_length_var)+1))
args.max_prompt_length = max(args.max_prompt_length, int(args.mean_prompt_length * 3))
# check if the all_text is longer than the max prompt length, if not expand it
global all_text
while len(tokenizer.tokenize(all_text)) < args.max_prompt_length:
Expand Down

0 comments on commit 25ae45c

Please sign in to comment.