2025-01-03 12:04:29 +08:00
|
|
|
from mmengine.config import read_base
|
|
|
|
|
|
|
|
with read_base():
|
|
|
|
# Models
|
|
|
|
# Datasets
|
2025-01-20 19:17:38 +08:00
|
|
|
from opencompass.configs.datasets.longbenchv2.longbenchv2_gen import \
|
|
|
|
LongBenchv2_datasets as LongBenchv2_datasets
|
|
|
|
from opencompass.configs.models.chatglm.lmdeploy_glm4_9b_chat import \
|
|
|
|
models as lmdeploy_glm4_9b_chat_model
|
|
|
|
from opencompass.configs.models.hf_llama.lmdeploy_llama3_1_8b_instruct import \
|
|
|
|
models as lmdeploy_llama3_1_8b_instruct_model
|
|
|
|
from opencompass.configs.models.qwen2_5.lmdeploy_qwen2_5_7b_instruct import \
|
|
|
|
models as lmdeploy_qwen2_5_7b_instruct_model
|
2025-01-03 12:04:29 +08:00
|
|
|
|
|
|
|
datasets = sum((v for k, v in locals().items() if k.endswith('_datasets')), [])
|
|
|
|
|
|
|
|
models = sum([v for k, v in locals().items() if k.endswith('_model')], [])
|
|
|
|
|
|
|
|
for model in models:
|
|
|
|
model['max_seq_len'] = 128 * 1024
|
|
|
|
model['engine_config']['session_len'] = 128 * 1024
|
|
|
|
model['engine_config']['tp'] = 2
|
|
|
|
model['run_cfg']['num_gpus'] = 2
|
|
|
|
# Drop middle tokens to make input length shorter than session_len, use 128k to keep sync with Longbenchv2 original code
|
|
|
|
# Drop middle now only support LMDeploy models
|
|
|
|
model['drop_middle'] = True
|
|
|
|
|
|
|
|
work_dir = './outputs/longbenchv2'
|