This commit is contained in:
Mor-Li 2025-04-25 18:14:57 +08:00
parent 3914aef997
commit 70e0dc1674
3 changed files with 3 additions and 3 deletions

View File

@ -7,7 +7,7 @@ models = [
path='google/gemma-3-12b-it', path='google/gemma-3-12b-it',
model_kwargs=dict(tensor_parallel_size=4, model_kwargs=dict(tensor_parallel_size=4,
# for long context # for long context
rope_scaling={"factor": 8.0, "rope_type": "linear"}), rope_scaling={'factor': 8.0, 'rope_type': 'linear'}),
max_out_len=4096, max_out_len=4096,
batch_size=1, batch_size=1,
generation_kwargs=dict(temperature=0), generation_kwargs=dict(temperature=0),

View File

@ -7,7 +7,7 @@ models = [
path='google/gemma-3-27b-it', path='google/gemma-3-27b-it',
model_kwargs=dict(tensor_parallel_size=4, model_kwargs=dict(tensor_parallel_size=4,
# for long context # for long context
rope_scaling={"factor": 8.0, "rope_type": "linear"}), rope_scaling={'factor': 8.0, 'rope_type': 'linear'}),
max_out_len=4096, max_out_len=4096,
batch_size=1, batch_size=1,
generation_kwargs=dict(temperature=0), generation_kwargs=dict(temperature=0),

View File

@ -7,7 +7,7 @@ models = [
path='google/gemma-3-4b-it', path='google/gemma-3-4b-it',
model_kwargs=dict(tensor_parallel_size=2, model_kwargs=dict(tensor_parallel_size=2,
# for long context # for long context
rope_scaling={"factor": 8.0, "rope_type": "linear"}), rope_scaling={'factor': 8.0, 'rope_type': 'linear'}),
max_seq_len=140000, max_seq_len=140000,
max_out_len=4096, max_out_len=4096,
batch_size=1, batch_size=1,