From 808582d9522af49eee91af254ceb19252dc94848 Mon Sep 17 00:00:00 2001 From: Xu Song Date: Wed, 29 May 2024 10:14:08 +0800 Subject: [PATCH] Fix VLLM argument error (#1207) --- configs/models/mistral/vllm_mistral_7b_instruct_v0_1.py | 2 +- configs/models/mistral/vllm_mistral_7b_instruct_v0_2.py | 2 +- configs/models/mistral/vllm_mixtral_8x7b_instruct_v0_1.py | 2 +- configs/models/others/vllm_orionstar_14b_longchat.py | 2 +- configs/models/qwen/vllm_qwen1_5_14b_chat.py | 2 +- configs/models/qwen/vllm_qwen1_5_72b_chat.py | 2 +- configs/models/qwen/vllm_qwen_14b_chat.py | 2 +- configs/models/qwen/vllm_qwen_72b_chat.py | 2 +- configs/models/vicuna/vllm_vicuna_13b_v15_16k.py | 2 +- configs/models/vicuna/vllm_vicuna_7b_v15_16k.py | 2 +- configs/models/wizardlm/vllm_wizardlm_13b_v1_2.py | 2 +- configs/models/wizardlm/vllm_wizardlm_70b_v1_0.py | 2 +- configs/models/wizardlm/vllm_wizardlm_7b_v1_0.py | 2 +- configs/models/zephyr/vllm_zephyr_7b_beta.py | 2 +- 14 files changed, 14 insertions(+), 14 deletions(-) diff --git a/configs/models/mistral/vllm_mistral_7b_instruct_v0_1.py b/configs/models/mistral/vllm_mistral_7b_instruct_v0_1.py index 5f4eee67..db4094f6 100644 --- a/configs/models/mistral/vllm_mistral_7b_instruct_v0_1.py +++ b/configs/models/mistral/vllm_mistral_7b_instruct_v0_1.py @@ -19,7 +19,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=1, num_procs=1), ) ] diff --git a/configs/models/mistral/vllm_mistral_7b_instruct_v0_2.py b/configs/models/mistral/vllm_mistral_7b_instruct_v0_2.py index ca202d1b..c1dd32ba 100644 --- a/configs/models/mistral/vllm_mistral_7b_instruct_v0_2.py +++ b/configs/models/mistral/vllm_mistral_7b_instruct_v0_2.py @@ -19,7 +19,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=1, num_procs=1), ) ] diff --git a/configs/models/mistral/vllm_mixtral_8x7b_instruct_v0_1.py b/configs/models/mistral/vllm_mixtral_8x7b_instruct_v0_1.py index 03975355..2bbc0f0d 100644 --- a/configs/models/mistral/vllm_mixtral_8x7b_instruct_v0_1.py +++ b/configs/models/mistral/vllm_mixtral_8x7b_instruct_v0_1.py @@ -20,7 +20,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=2, num_procs=1), ) ] diff --git a/configs/models/others/vllm_orionstar_14b_longchat.py b/configs/models/others/vllm_orionstar_14b_longchat.py index 58cdea8d..2c90f7a3 100644 --- a/configs/models/others/vllm_orionstar_14b_longchat.py +++ b/configs/models/others/vllm_orionstar_14b_longchat.py @@ -21,6 +21,6 @@ models = [ max_seq_len=4096, batch_size=32, run_cfg=dict(num_gpus=4, num_procs=1), - end_str='<|endoftext|>', + stop_words=['<|endoftext|>'], ) ] diff --git a/configs/models/qwen/vllm_qwen1_5_14b_chat.py b/configs/models/qwen/vllm_qwen1_5_14b_chat.py index e4736857..b5e97d3d 100644 --- a/configs/models/qwen/vllm_qwen1_5_14b_chat.py +++ b/configs/models/qwen/vllm_qwen1_5_14b_chat.py @@ -19,7 +19,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='<|im_end|>', + stop_words=['<|im_end|>'], run_cfg=dict(num_gpus=2, num_procs=1), ) ] diff --git a/configs/models/qwen/vllm_qwen1_5_72b_chat.py b/configs/models/qwen/vllm_qwen1_5_72b_chat.py index 921e220a..2cb2ca53 100644 --- a/configs/models/qwen/vllm_qwen1_5_72b_chat.py +++ b/configs/models/qwen/vllm_qwen1_5_72b_chat.py @@ -19,7 +19,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='<|im_end|>', + stop_words=['<|im_end|>'], run_cfg=dict(num_gpus=4, num_procs=1), ) ] diff --git a/configs/models/qwen/vllm_qwen_14b_chat.py b/configs/models/qwen/vllm_qwen_14b_chat.py index 830010ef..53e221aa 100644 --- a/configs/models/qwen/vllm_qwen_14b_chat.py +++ b/configs/models/qwen/vllm_qwen_14b_chat.py @@ -19,7 +19,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='<|im_end|>', + stop_words=['<|im_end|>'], run_cfg=dict(num_gpus=4, num_procs=1), ) ] diff --git a/configs/models/qwen/vllm_qwen_72b_chat.py b/configs/models/qwen/vllm_qwen_72b_chat.py index cf2caa2c..fb727913 100644 --- a/configs/models/qwen/vllm_qwen_72b_chat.py +++ b/configs/models/qwen/vllm_qwen_72b_chat.py @@ -19,7 +19,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='<|im_end|>', + stop_words=['<|im_end|>'], run_cfg=dict(num_gpus=4, num_procs=1), ) ] diff --git a/configs/models/vicuna/vllm_vicuna_13b_v15_16k.py b/configs/models/vicuna/vllm_vicuna_13b_v15_16k.py index b15b4921..e26b3ff1 100644 --- a/configs/models/vicuna/vllm_vicuna_13b_v15_16k.py +++ b/configs/models/vicuna/vllm_vicuna_13b_v15_16k.py @@ -17,7 +17,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=2, num_procs=1), ) ] diff --git a/configs/models/vicuna/vllm_vicuna_7b_v15_16k.py b/configs/models/vicuna/vllm_vicuna_7b_v15_16k.py index b45646f0..ca261f13 100644 --- a/configs/models/vicuna/vllm_vicuna_7b_v15_16k.py +++ b/configs/models/vicuna/vllm_vicuna_7b_v15_16k.py @@ -17,7 +17,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=1, num_procs=1), ) ] diff --git a/configs/models/wizardlm/vllm_wizardlm_13b_v1_2.py b/configs/models/wizardlm/vllm_wizardlm_13b_v1_2.py index 5df42752..2e7fafeb 100644 --- a/configs/models/wizardlm/vllm_wizardlm_13b_v1_2.py +++ b/configs/models/wizardlm/vllm_wizardlm_13b_v1_2.py @@ -18,7 +18,7 @@ models = [ max_seq_len=2048, batch_size=1, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=1, num_procs=1), ) ] diff --git a/configs/models/wizardlm/vllm_wizardlm_70b_v1_0.py b/configs/models/wizardlm/vllm_wizardlm_70b_v1_0.py index 8c25d010..a722593a 100644 --- a/configs/models/wizardlm/vllm_wizardlm_70b_v1_0.py +++ b/configs/models/wizardlm/vllm_wizardlm_70b_v1_0.py @@ -19,7 +19,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=4, num_procs=1), ) ] diff --git a/configs/models/wizardlm/vllm_wizardlm_7b_v1_0.py b/configs/models/wizardlm/vllm_wizardlm_7b_v1_0.py index b7fe7f02..60b33c86 100644 --- a/configs/models/wizardlm/vllm_wizardlm_7b_v1_0.py +++ b/configs/models/wizardlm/vllm_wizardlm_7b_v1_0.py @@ -18,7 +18,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=1, num_procs=1), ) ] diff --git a/configs/models/zephyr/vllm_zephyr_7b_beta.py b/configs/models/zephyr/vllm_zephyr_7b_beta.py index 956ac073..499b58b6 100644 --- a/configs/models/zephyr/vllm_zephyr_7b_beta.py +++ b/configs/models/zephyr/vllm_zephyr_7b_beta.py @@ -17,7 +17,7 @@ models = [ max_seq_len=2048, batch_size=32, generation_kwargs=dict(temperature=0), - end_str='', + stop_words=[''], run_cfg=dict(num_gpus=1, num_procs=1), ) ]