diff --git a/opencompass/configs/models/gemma/vllm_gemma_3_12b_it.py b/opencompass/configs/models/gemma/vllm_gemma_3_12b_it.py index f3d2c643..2914640f 100644 --- a/opencompass/configs/models/gemma/vllm_gemma_3_12b_it.py +++ b/opencompass/configs/models/gemma/vllm_gemma_3_12b_it.py @@ -7,7 +7,7 @@ models = [ path='google/gemma-3-12b-it', model_kwargs=dict(tensor_parallel_size=4, # for long context - rope_scaling={"factor": 8.0, "rope_type": "linear"}), + rope_scaling={'factor': 8.0, 'rope_type': 'linear'}), max_out_len=4096, batch_size=1, generation_kwargs=dict(temperature=0), diff --git a/opencompass/configs/models/gemma/vllm_gemma_3_27b_it.py b/opencompass/configs/models/gemma/vllm_gemma_3_27b_it.py index 7eeb6275..b6f4b93b 100644 --- a/opencompass/configs/models/gemma/vllm_gemma_3_27b_it.py +++ b/opencompass/configs/models/gemma/vllm_gemma_3_27b_it.py @@ -7,7 +7,7 @@ models = [ path='google/gemma-3-27b-it', model_kwargs=dict(tensor_parallel_size=4, # for long context - rope_scaling={"factor": 8.0, "rope_type": "linear"}), + rope_scaling={'factor': 8.0, 'rope_type': 'linear'}), max_out_len=4096, batch_size=1, generation_kwargs=dict(temperature=0), diff --git a/opencompass/configs/models/gemma/vllm_gemma_3_4b_it.py b/opencompass/configs/models/gemma/vllm_gemma_3_4b_it.py index c6476d0c..22516ff7 100644 --- a/opencompass/configs/models/gemma/vllm_gemma_3_4b_it.py +++ b/opencompass/configs/models/gemma/vllm_gemma_3_4b_it.py @@ -7,7 +7,7 @@ models = [ path='google/gemma-3-4b-it', model_kwargs=dict(tensor_parallel_size=2, # for long context - rope_scaling={"factor": 8.0, "rope_type": "linear"}), + rope_scaling={'factor': 8.0, 'rope_type': 'linear'}), max_seq_len=140000, max_out_len=4096, batch_size=1,