diff --git a/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py b/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py index a82242da5..90d7a3211 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py @@ -7,7 +7,7 @@ https://huggingface.co/GAIR/autoj-bilingual-6b ''' -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='autoj-bilingual-6b', path="GAIR/autoj-bilingual-6b", @@ -21,4 +21,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) + )] diff --git a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py index 60bcc50ab..9af5f354a 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py @@ -1,7 +1,7 @@ from opencompass.models import HuggingFaceCausalLM -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='autoj-13b-GPTQ-4bits', path="GAIR/autoj-13b-GPTQ-4bits", @@ -15,4 +15,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) \ No newline at end of file + )] \ No newline at end of file diff --git a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py index 6975bd5bf..73b363bff 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py @@ -6,7 +6,7 @@ https://huggingface.co/GAIR/autoj-13b-GPTQ-4bits ''' -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='autoj-13b', path="GAIR/autoj-13b", @@ -20,4 +20,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) + )] diff --git a/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py b/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py index fd1bfb35e..68d0340ea 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py @@ -1,7 +1,7 @@ from opencompass.models import HuggingFaceCausalLM -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='autoj-scenario-classifier', path="GAIR/autoj-scenario-classifier", @@ -15,4 +15,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) \ No newline at end of file + )] \ No newline at end of file diff --git a/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py b/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py index 7a4f0417a..d8bf27517 100644 --- a/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py +++ b/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py @@ -1,7 +1,7 @@ from opencompass.models import HuggingFaceCausalLM -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='judgelm-13b-v1-hf', path="BAAI/JudgeLM-13B-v1.0", @@ -15,4 +15,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) \ No newline at end of file + )] \ No newline at end of file diff --git a/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py b/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py index 5bf0e12c6..9be5f8733 100644 --- a/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py +++ b/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py @@ -1,7 +1,7 @@ from opencompass.models import HuggingFaceCausalLM -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='judgelm-33b-v1-hf', path="BAAI/JudgeLM-33B-v1.0", @@ -15,4 +15,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) \ No newline at end of file + )] \ No newline at end of file diff --git a/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py b/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py index 7d9295986..b9ee641f9 100644 --- a/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py +++ b/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py @@ -1,7 +1,7 @@ from opencompass.models import HuggingFaceCausalLM -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='judgelm-7b-v1-hf', path="BAAI/JudgeLM-7B-v1.0", @@ -15,4 +15,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) \ No newline at end of file + )] \ No newline at end of file diff --git a/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py b/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py index 837279598..576a99f12 100644 --- a/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py +++ b/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py @@ -1,7 +1,7 @@ from opencompass.models import HuggingFaceCausalLM -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='alpaca-pandalm-7b-v1-hf', path="WeOpenML/PandaLM-Alpaca-7B-v1", @@ -15,4 +15,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) \ No newline at end of file + )] \ No newline at end of file diff --git a/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py b/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py index 3211c2771..ebcddda91 100644 --- a/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py +++ b/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py @@ -1,7 +1,7 @@ from opencompass.models import HuggingFaceCausalLM -models = dict( +models = [dict( type=HuggingFaceCausalLM, abbr='pandalm-7b-v1-hf', path="WeOpenML/PandaLM-7B-v1", @@ -15,4 +15,4 @@ batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - ) \ No newline at end of file + )] \ No newline at end of file