OpenCompass/configs/models/qwen/vllm_qwen_14b_chat.py
Mo Li 0a6a03fe1a
[Feature] update needlebench and configs (#986)
* add Needlebench-1000K configs

* add prompt postion args

* add model configs

* Update parallel.py

* fix lint
2024-03-25 18:05:01 +08:00

26 lines
658 B
Python

from opencompass.models import VLLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='\n<|im_start|>user\n', end='<|im_end|>'),
dict(role="BOT", begin="\n<|im_start|>assistant\n", end='<|im_end|>', generate=True),
],
)
models = [
dict(
type=VLLM,
abbr='qwen-14b-chat-vllm',
path="Qwen/Qwen-14B-Chat",
model_kwargs=dict(tensor_parallel_size=4),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=32,
generation_kwargs=dict(temperature=0),
end_str='<|im_end|>',
run_cfg=dict(num_gpus=4, num_procs=1),
)
]