diff --git a/configs/models/qwen/hf_qwen2_beta_7b.py b/configs/models/qwen/hf_qwen1_5_0_5b.py similarity index 82% rename from configs/models/qwen/hf_qwen2_beta_7b.py rename to configs/models/qwen/hf_qwen1_5_0_5b.py index 86e025a7..62a219f0 100644 --- a/configs/models/qwen/hf_qwen2_beta_7b.py +++ b/configs/models/qwen/hf_qwen1_5_0_5b.py @@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-7b-hf', - path="Qwen/Qwen2-beta-7B", - tokenizer_path='Qwen/Qwen2-beta-7B', + abbr='qwen1.5-0.5b-hf', + path="Qwen/Qwen1.5-0.5B", + tokenizer_path='Qwen/Qwen1.5-0.5B', model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_7b_chat.py b/configs/models/qwen/hf_qwen1_5_0_5b_chat.py similarity index 91% rename from configs/models/qwen/hf_qwen2_beta_7b_chat.py rename to configs/models/qwen/hf_qwen1_5_0_5b_chat.py index 2a3f0099..04a6afed 100644 --- a/configs/models/qwen/hf_qwen2_beta_7b_chat.py +++ b/configs/models/qwen/hf_qwen1_5_0_5b_chat.py @@ -11,8 +11,8 @@ _meta_template = dict( models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-7b-chat-hf', - path="Qwen/Qwen2-beta-7B-Chat", + abbr='qwen1.5-0.5b-chat-hf', + path="Qwen/Qwen1.5-0.5B-Chat", model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_4b.py b/configs/models/qwen/hf_qwen1_5_14b.py similarity index 82% rename from configs/models/qwen/hf_qwen2_beta_4b.py rename to configs/models/qwen/hf_qwen1_5_14b.py index 52206a1a..e9d75e4e 100644 --- a/configs/models/qwen/hf_qwen2_beta_4b.py +++ b/configs/models/qwen/hf_qwen1_5_14b.py @@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-4b-hf', - path="Qwen/Qwen2-beta-4B", - tokenizer_path='Qwen/Qwen2-beta-4B', + abbr='qwen1.5-14b-hf', + path="Qwen/Qwen1.5-14B", + tokenizer_path='Qwen/Qwen1.5-14B', model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_4b_chat.py b/configs/models/qwen/hf_qwen1_5_14b_chat.py similarity index 91% rename from configs/models/qwen/hf_qwen2_beta_4b_chat.py rename to configs/models/qwen/hf_qwen1_5_14b_chat.py index 7308a6f2..56c8bc47 100644 --- a/configs/models/qwen/hf_qwen2_beta_4b_chat.py +++ b/configs/models/qwen/hf_qwen1_5_14b_chat.py @@ -11,8 +11,8 @@ _meta_template = dict( models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-4b-chat-hf', - path="Qwen/Qwen2-beta-4B-Chat", + abbr='qwen1.5-14b-chat-hf', + path="Qwen/Qwen1.5-14B-Chat", model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_14b.py b/configs/models/qwen/hf_qwen1_5_1_8b.py similarity index 82% rename from configs/models/qwen/hf_qwen2_beta_14b.py rename to configs/models/qwen/hf_qwen1_5_1_8b.py index 2957b6d1..71492cf8 100644 --- a/configs/models/qwen/hf_qwen2_beta_14b.py +++ b/configs/models/qwen/hf_qwen1_5_1_8b.py @@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-14b-hf', - path="Qwen/Qwen2-beta-14B", - tokenizer_path='Qwen/Qwen2-beta-14B', + abbr='qwen1.5-1.8b-hf', + path="Qwen/Qwen1.5-1.8B", + tokenizer_path='Qwen/Qwen1.5-1.8B', model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_14b_chat.py b/configs/models/qwen/hf_qwen1_5_1_8b_chat.py similarity index 91% rename from configs/models/qwen/hf_qwen2_beta_14b_chat.py rename to configs/models/qwen/hf_qwen1_5_1_8b_chat.py index ee5e4f55..22f425ad 100644 --- a/configs/models/qwen/hf_qwen2_beta_14b_chat.py +++ b/configs/models/qwen/hf_qwen1_5_1_8b_chat.py @@ -11,8 +11,8 @@ _meta_template = dict( models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-14b-chat-hf', - path="Qwen/Qwen2-beta-14B-Chat", + abbr='qwen1.5-1.8b-chat-hf', + path="Qwen/Qwen1.5-1.8B-Chat", model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_1_8b.py b/configs/models/qwen/hf_qwen1_5_4b.py similarity index 82% rename from configs/models/qwen/hf_qwen2_beta_1_8b.py rename to configs/models/qwen/hf_qwen1_5_4b.py index ccba9e32..6aa57263 100644 --- a/configs/models/qwen/hf_qwen2_beta_1_8b.py +++ b/configs/models/qwen/hf_qwen1_5_4b.py @@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-1.8b-hf', - path="Qwen/Qwen2-beta-1_8B", - tokenizer_path='Qwen/Qwen2-beta-1_8B', + abbr='qwen1.5-4b-hf', + path="Qwen/Qwen1.5-4B", + tokenizer_path='Qwen/Qwen1.5-4B', model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_0_5b_chat.py b/configs/models/qwen/hf_qwen1_5_4b_chat.py similarity index 91% rename from configs/models/qwen/hf_qwen2_beta_0_5b_chat.py rename to configs/models/qwen/hf_qwen1_5_4b_chat.py index 72db6a4a..b5ed4f89 100644 --- a/configs/models/qwen/hf_qwen2_beta_0_5b_chat.py +++ b/configs/models/qwen/hf_qwen1_5_4b_chat.py @@ -11,8 +11,8 @@ _meta_template = dict( models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-0.5b-chat-hf', - path="Qwen/Qwen2-beta-0_5B-Chat", + abbr='qwen1.5-4b-chat-hf', + path="Qwen/Qwen1.5-4B-Chat", model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_72b.py b/configs/models/qwen/hf_qwen1_5_72b.py similarity index 82% rename from configs/models/qwen/hf_qwen2_beta_72b.py rename to configs/models/qwen/hf_qwen1_5_72b.py index 99e017cb..3dd6e638 100644 --- a/configs/models/qwen/hf_qwen2_beta_72b.py +++ b/configs/models/qwen/hf_qwen1_5_72b.py @@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-72b-hf', - path="Qwen/Qwen2-beta-72B", - tokenizer_path='Qwen/Qwen2-beta-72B', + abbr='qwen1.5-72b-hf', + path="Qwen/Qwen1.5-72B", + tokenizer_path='Qwen/Qwen1.5-72B', model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen2_beta_72b_chat.py b/configs/models/qwen/hf_qwen1_5_72b_chat.py similarity index 91% rename from configs/models/qwen/hf_qwen2_beta_72b_chat.py rename to configs/models/qwen/hf_qwen1_5_72b_chat.py index 84f6d809..b8b27b4c 100644 --- a/configs/models/qwen/hf_qwen2_beta_72b_chat.py +++ b/configs/models/qwen/hf_qwen1_5_72b_chat.py @@ -11,8 +11,8 @@ _meta_template = dict( models = [ dict( type=HuggingFaceCausalLM, - abbr='qwen2-beta-72b-chat-hf', - path="Qwen/Qwen2-beta-72B-Chat", + abbr='qwen1.5-72b-chat-hf', + path="Qwen/Qwen1.5-72B-Chat", model_kwargs=dict( device_map='auto', trust_remote_code=True diff --git a/configs/models/qwen/hf_qwen1_5_7b.py b/configs/models/qwen/hf_qwen1_5_7b.py new file mode 100644 index 00000000..d9df3031 --- /dev/null +++ b/configs/models/qwen/hf_qwen1_5_7b.py @@ -0,0 +1,25 @@ +from opencompass.models import HuggingFaceCausalLM + +models = [ + dict( + type=HuggingFaceCausalLM, + abbr='qwen1.5-7b-hf', + path="Qwen/Qwen1.5-7B", + tokenizer_path='Qwen/Qwen1.5-7B', + model_kwargs=dict( + device_map='auto', + trust_remote_code=True + ), + tokenizer_kwargs=dict( + padding_side='left', + truncation_side='left', + trust_remote_code=True, + use_fast=False, + ), + pad_token_id=151645, + max_out_len=100, + max_seq_len=2048, + batch_size=8, + run_cfg=dict(num_gpus=1, num_procs=1), + ) +] diff --git a/configs/models/qwen/hf_qwen1_5_7b_chat.py b/configs/models/qwen/hf_qwen1_5_7b_chat.py new file mode 100644 index 00000000..8eb05bee --- /dev/null +++ b/configs/models/qwen/hf_qwen1_5_7b_chat.py @@ -0,0 +1,34 @@ +from opencompass.models import HuggingFaceCausalLM + +_meta_template = dict( + round=[ + dict(role="HUMAN", begin='<|im_start|>user\n', end='<|im_end|>\n'), + dict(role="BOT", begin="<|im_start|>assistant\n", end='<|im_end|>\n', generate=True), + ], + eos_token_id=151645, +) + +models = [ + dict( + type=HuggingFaceCausalLM, + abbr='qwen1.5-7b-chat-hf', + path="Qwen/Qwen1.5-7B-Chat", + model_kwargs=dict( + device_map='auto', + trust_remote_code=True + ), + tokenizer_kwargs=dict( + padding_side='left', + truncation_side='left', + trust_remote_code=True, + use_fast=False, + ), + meta_template=_meta_template, + pad_token_id=151645, + max_out_len=100, + max_seq_len=2048, + batch_size=8, + run_cfg=dict(num_gpus=1, num_procs=1), + end_str='<|im_end|>', + ) +] diff --git a/configs/models/qwen/hf_qwen2_beta_0_5b.py b/configs/models/qwen/hf_qwen2_beta_0_5b.py deleted file mode 100644 index 3c1d9e09..00000000 --- a/configs/models/qwen/hf_qwen2_beta_0_5b.py +++ /dev/null @@ -1,25 +0,0 @@ -from opencompass.models import HuggingFaceCausalLM - -models = [ - dict( - type=HuggingFaceCausalLM, - abbr='qwen2-beta-0.5b-hf', - path="Qwen/Qwen2-beta-0_5B", - tokenizer_path='Qwen/Qwen2-beta-0_5B', - model_kwargs=dict( - device_map='auto', - trust_remote_code=True - ), - tokenizer_kwargs=dict( - padding_side='left', - truncation_side='left', - trust_remote_code=True, - use_fast=False, - ), - pad_token_id=151645, - max_out_len=100, - max_seq_len=2048, - batch_size=8, - run_cfg=dict(num_gpus=1, num_procs=1), - ) -] diff --git a/configs/models/qwen/hf_qwen2_beta_1_8b_chat.py b/configs/models/qwen/hf_qwen2_beta_1_8b_chat.py deleted file mode 100644 index 73f7e5df..00000000 --- a/configs/models/qwen/hf_qwen2_beta_1_8b_chat.py +++ /dev/null @@ -1,34 +0,0 @@ -from opencompass.models import HuggingFaceCausalLM - -_meta_template = dict( - round=[ - dict(role="HUMAN", begin='<|im_start|>user\n', end='<|im_end|>\n'), - dict(role="BOT", begin="<|im_start|>assistant\n", end='<|im_end|>\n', generate=True), - ], - eos_token_id=151645, -) - -models = [ - dict( - type=HuggingFaceCausalLM, - abbr='qwen2-beta-1.8b-chat-hf', - path="Qwen/Qwen2-beta-1_8B-Chat", - model_kwargs=dict( - device_map='auto', - trust_remote_code=True - ), - tokenizer_kwargs=dict( - padding_side='left', - truncation_side='left', - trust_remote_code=True, - use_fast=False, - ), - meta_template=_meta_template, - pad_token_id=151645, - max_out_len=100, - max_seq_len=2048, - batch_size=8, - run_cfg=dict(num_gpus=1, num_procs=1), - end_str='<|im_end|>', - ) -] diff --git a/configs/models/qwen/vllm_qwen2_beta_72b.py b/configs/models/qwen/vllm_qwen1_5_72b.py similarity index 80% rename from configs/models/qwen/vllm_qwen2_beta_72b.py rename to configs/models/qwen/vllm_qwen1_5_72b.py index 3b7c9af7..10ac1569 100644 --- a/configs/models/qwen/vllm_qwen2_beta_72b.py +++ b/configs/models/qwen/vllm_qwen1_5_72b.py @@ -4,8 +4,8 @@ from opencompass.models import VLLM models = [ dict( type=VLLM, - abbr='qwen2-beta-72b-vllm', - path="Qwen/Qwen2-beta-72B", + abbr='qwen1.5-72b-vllm', + path="Qwen/Qwen1.5-72B", model_kwargs=dict(tensor_parallel_size=4), max_out_len=100, max_seq_len=2048, diff --git a/configs/models/qwen/vllm_qwen2_beta_72b_chat.py b/configs/models/qwen/vllm_qwen1_5_72b_chat.py similarity index 88% rename from configs/models/qwen/vllm_qwen2_beta_72b_chat.py rename to configs/models/qwen/vllm_qwen1_5_72b_chat.py index 2e7ca59c..035c7a8a 100644 --- a/configs/models/qwen/vllm_qwen2_beta_72b_chat.py +++ b/configs/models/qwen/vllm_qwen1_5_72b_chat.py @@ -12,8 +12,8 @@ _meta_template = dict( models = [ dict( type=VLLM, - abbr='qwen2-beta-72b-chat-vllm', - path="Qwen/Qwen2-beta-72B-Chat", + abbr='qwen1.5-72b-chat-vllm', + path="Qwen/Qwen1.5-72B-Chat", model_kwargs=dict(tensor_parallel_size=4), meta_template=_meta_template, max_out_len=100,