diff --git a/configs/datasets/subjective_alignbench/alignbench_judgeby_judgelm.py b/configs/datasets/subjective_alignbench/alignbench_judgeby_judgelm.py index 048de51b..abdc1b88 100644 --- a/configs/datasets/subjective_alignbench/alignbench_judgeby_judgelm.py +++ b/configs/datasets/subjective_alignbench/alignbench_judgeby_judgelm.py @@ -28,7 +28,7 @@ for _name in subjective_all_sets: ]), ), retriever=dict(type=ZeroRetriever), - inferencer=dict(type=GenInferencer, max_out_len=2048), + inferencer=dict(type=GenInferencer, max_out_len=2048), ) subjective_eval_cfg = dict( diff --git a/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py b/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py index 90d7a321..51a64235 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_bilingual_6b.py @@ -16,8 +16,8 @@ models = [dict( truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), diff --git a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py index 9af5f354..8bed3bb9 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b.py @@ -3,15 +3,15 @@ from opencompass.models import HuggingFaceCausalLM models = [dict( type=HuggingFaceCausalLM, - abbr='autoj-13b-GPTQ-4bits', - path="GAIR/autoj-13b-GPTQ-4bits", - tokenizer_path='GAIR/autoj-13b-GPTQ-4bits', + abbr='autoj-13b', + path="GAIR/autoj-13b", + tokenizer_path='GAIR/autoj-13b', tokenizer_kwargs=dict(padding_side='left', truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), diff --git a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py index 73b363bf..b0aa80b0 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_eng_13b_4bit.py @@ -8,16 +8,16 @@ https://huggingface.co/GAIR/autoj-13b-GPTQ-4bits models = [dict( type=HuggingFaceCausalLM, - abbr='autoj-13b', - path="GAIR/autoj-13b", - tokenizer_path='GAIR/autoj-13b', + abbr='autoj-13b-GPTQ-4bits', + path="GAIR/autoj-13b-GPTQ-4bits", + tokenizer_path='GAIR/autoj-13b-GPTQ-4bits', tokenizer_kwargs=dict(padding_side='left', truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), - )] + )] \ No newline at end of file diff --git a/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py b/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py index 68d0340e..fd7cd3b1 100644 --- a/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py +++ b/configs/models/judge_llm/auto_j/hf_autoj_scen_classifier.py @@ -10,8 +10,8 @@ models = [dict( truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), diff --git a/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py b/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py index d8bf2751..bb1d3e36 100644 --- a/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py +++ b/configs/models/judge_llm/judgelm/hf_judgelm_13b_v1.py @@ -10,8 +10,8 @@ models = [dict( truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), diff --git a/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py b/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py index 9be5f873..c81cfb60 100644 --- a/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py +++ b/configs/models/judge_llm/judgelm/hf_judgelm_33b_v1.py @@ -10,9 +10,9 @@ models = [dict( truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), - run_cfg=dict(num_gpus=1, num_procs=1), + run_cfg=dict(num_gpus=4, num_procs=1), )] \ No newline at end of file diff --git a/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py b/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py index b9ee641f..a4d22b76 100644 --- a/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py +++ b/configs/models/judge_llm/judgelm/hf_judgelm_7b_v1.py @@ -10,8 +10,8 @@ models = [dict( truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), diff --git a/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py b/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py index 576a99f1..510e7adf 100644 --- a/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py +++ b/configs/models/judge_llm/pandalm/hf_alpaca_pandalm_7b_v1.py @@ -10,8 +10,8 @@ models = [dict( truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1), diff --git a/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py b/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py index ebcddda9..f083ee10 100644 --- a/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py +++ b/configs/models/judge_llm/pandalm/hf_pandalm_7b_v1.py @@ -10,8 +10,8 @@ models = [dict( truncation_side='left', trust_remote_code=True, use_fast=False,), - max_out_len=512, - max_seq_len=2048, + max_out_len=1024, + max_seq_len=4096, batch_size=8, model_kwargs=dict(device_map='auto', trust_remote_code=True), run_cfg=dict(num_gpus=1, num_procs=1),