diff --git a/opencompass/configs/models/deepseek/lmdeploy_deepseek_v2_5.py b/opencompass/configs/models/deepseek/lmdeploy_deepseek_v2_5.py index 34574eff..f0864516 100644 --- a/opencompass/configs/models/deepseek/lmdeploy_deepseek_v2_5.py +++ b/opencompass/configs/models/deepseek/lmdeploy_deepseek_v2_5.py @@ -5,6 +5,7 @@ models = [ type=TurboMindModelwithChatTemplate, abbr='deepseek-v2_5-turbomind', path='deepseek-ai/DeepSeek-V2.5', + backend='pytorch', engine_config=dict( session_len=7168, max_batch_size=4, diff --git a/opencompass/configs/models/deepseek/lmdeploy_deepseek_v2_5_1210.py b/opencompass/configs/models/deepseek/lmdeploy_deepseek_v2_5_1210.py new file mode 100644 index 00000000..bbb9f4b2 --- /dev/null +++ b/opencompass/configs/models/deepseek/lmdeploy_deepseek_v2_5_1210.py @@ -0,0 +1,21 @@ +from opencompass.models import TurboMindModelwithChatTemplate + +models = [ + dict( + type=TurboMindModelwithChatTemplate, + abbr='deepseek-v2_5-1210-turbomind', + path='deepseek-ai/DeepSeek-V2.5-1210', + backend='pytorch', + engine_config=dict( + session_len=7168, + max_batch_size=4, + tp=8, + cache_max_entry_count=0.7, + ), + gen_config=dict(top_k=1, temperature=1e-6, top_p=0.9), + max_seq_len=7168, + max_out_len=2048, + batch_size=4, + run_cfg=dict(num_gpus=8), + ) +] diff --git a/requirements/runtime.txt b/requirements/runtime.txt index 16d20d74..fb87ca1a 100644 --- a/requirements/runtime.txt +++ b/requirements/runtime.txt @@ -8,6 +8,7 @@ func_timeout fuzzywuzzy gradio-client h5py +httpx==0.27.2 huggingface_hub<=0.24.7 immutabledict importlib-metadata