from opencompass.models import VLLMwithChatTemplate models = [ dict( type=VLLMwithChatTemplate, abbr='gemma-3-12b-it-vllm', path='google/gemma-3-12b-it', model_kwargs=dict(tensor_parallel_size=4, # for long context rope_scaling={"factor": 8.0, "rope_type": "linear"}), max_out_len=4096, batch_size=1, generation_kwargs=dict(temperature=0), run_cfg=dict(num_gpus=4), ) ]