This commit is contained in:
Alexander Lam 2025-05-28 11:28:54 +02:00 committed by GitHub
commit 5aeac8e77b
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
8 changed files with 89 additions and 10 deletions

View File

@ -3,7 +3,7 @@ from opencompass.models import TurboMindModelwithChatTemplate
models = [
dict(
type=TurboMindModelwithChatTemplate,
abbr='mixtral-large-instruct-2407-turbomind',
abbr='mistral-large-instruct-2407-turbomind',
path='mistralai/Mistral-Large-Instruct-2407',
engine_config=dict(
session_len=32768,
@ -11,9 +11,7 @@ models = [
tp=4,
cache_max_entry_count=0.7,
),
gen_config=dict(
top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096
),
gen_config=dict(top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096),
max_seq_len=32768,
max_out_len=4096,
batch_size=8,

View File

@ -0,0 +1,20 @@
from opencompass.models import TurboMindModelwithChatTemplate
models = [
dict(
type=TurboMindModelwithChatTemplate,
abbr='mistral-large-instruct-2411-turbomind',
path='mistralai/Mistral-Large-Instruct-2411',
engine_config=dict(
session_len=32768,
max_batch_size=16,
tp=4,
cache_max_entry_count=0.7,
),
gen_config=dict(top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096),
max_seq_len=32768,
max_out_len=4096,
batch_size=8,
run_cfg=dict(num_gpus=4),
)
]

View File

@ -0,0 +1,15 @@
from opencompass.models import TurboMindModelwithChatTemplate
models = [
dict(
type=TurboMindModelwithChatTemplate,
abbr='mistral-nemo-instruct-2407-turbomind',
path='mistralai/Mistral-Nemo-Instruct-2407',
engine_config=dict(session_len=32768, max_batch_size=16, tp=1),
gen_config=dict(top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096),
max_seq_len=32768,
max_out_len=4096,
batch_size=16,
run_cfg=dict(num_gpus=1),
)
]

View File

@ -0,0 +1,15 @@
from opencompass.models import TurboMindModelwithChatTemplate
models = [
dict(
type=TurboMindModelwithChatTemplate,
abbr='mistral-small-instruct-2409-turbomind',
path='mistralai/Mistral-Small-Instruct-2409',
engine_config=dict(session_len=32768, max_batch_size=16, tp=2),
gen_config=dict(top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096),
max_seq_len=32768,
max_out_len=4096,
batch_size=16,
run_cfg=dict(num_gpus=2),
)
]

View File

@ -1,10 +1,9 @@
from opencompass.models import VLLMwithChatTemplate
models = [
dict(
type=VLLMwithChatTemplate,
abbr='mixtral-large-instruct-2407-vllm',
abbr='mistral-large-instruct-2407-vllm',
path='mistralai/Mistral-Large-Instruct-2407',
model_kwargs=dict(tensor_parallel_size=8),
max_out_len=256,

View File

@ -0,0 +1,20 @@
from opencompass.models import TurboMindModelwithChatTemplate
models = [
dict(
type=TurboMindModelwithChatTemplate,
abbr='mistral-large-instruct-2407-turbomind',
path='mistralai/Mistral-Large-Instruct-2407',
engine_config=dict(
session_len=32768,
max_batch_size=16,
tp=4,
cache_max_entry_count=0.7,
),
gen_config=dict(top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096),
max_seq_len=32768,
max_out_len=4096,
batch_size=8,
run_cfg=dict(num_gpus=4),
)
]

View File

@ -3,7 +3,7 @@ from opencompass.models import TurboMindModelwithChatTemplate
models = [
dict(
type=TurboMindModelwithChatTemplate,
abbr='mixtral-large-instruct-2411-turbomind',
abbr='mistral-large-instruct-2411-turbomind',
path='mistralai/Mistral-Large-Instruct-2411',
engine_config=dict(
session_len=32768,
@ -11,9 +11,7 @@ models = [
tp=4,
cache_max_entry_count=0.7,
),
gen_config=dict(
top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096
),
gen_config=dict(top_k=1, temperature=1e-6, top_p=0.9, max_new_tokens=4096),
max_seq_len=32768,
max_out_len=4096,
batch_size=8,

View File

@ -0,0 +1,14 @@
from opencompass.models import VLLMwithChatTemplate
models = [
dict(
type=VLLMwithChatTemplate,
abbr='mistral-large-instruct-2407-vllm',
path='mistralai/Mistral-Large-Instruct-2407',
model_kwargs=dict(tensor_parallel_size=8),
max_out_len=256,
batch_size=16,
generation_kwargs=dict(temperature=0),
run_cfg=dict(num_gpus=8),
)
]