From aca8ec3c6ab5f2558005862ed0183bb4cfb6414d Mon Sep 17 00:00:00 2001 From: bittersweet1999 <148421775+bittersweet1999@users.noreply.github.com> Date: Wed, 13 Nov 2024 10:14:27 +0800 Subject: [PATCH] [Hotfix] Hotfix (#1683) * fix pip version * fix pip version * fix lint * hotfix --- opencompass/models/turbomind_with_tf_above_v4_33.py | 2 +- opencompass/models/vllm.py | 6 +++++- 2 files changed, 6 insertions(+), 2 deletions(-) diff --git a/opencompass/models/turbomind_with_tf_above_v4_33.py b/opencompass/models/turbomind_with_tf_above_v4_33.py index 76bbf194..cf5b880b 100644 --- a/opencompass/models/turbomind_with_tf_above_v4_33.py +++ b/opencompass/models/turbomind_with_tf_above_v4_33.py @@ -123,7 +123,7 @@ class TurboMindModelwithChatTemplate(BaseModel): gen_config = copy.deepcopy(DEFAULT_GEN_CONFIG) gen_config.update(self.gen_config) - if do_sample or self.gen_config['do_sample']: + if do_sample or ('do_sample' in self.gen_config and self.gen_config['do_sample']): gen_config['top_k'] = 40 gen_config['temperature'] = temperature else: diff --git a/opencompass/models/vllm.py b/opencompass/models/vllm.py index fbfaf66e..7e166ca7 100644 --- a/opencompass/models/vllm.py +++ b/opencompass/models/vllm.py @@ -101,7 +101,11 @@ class VLLM(BaseModel): if not self.lora_path: outputs = self.model.generate(inputs, sampling_kwargs) else: - outputs = self.model.generate(inputs, sampling_kwargs, lora_request=LoRARequest("sql_adapter", 1, self.lora_path)) + outputs = self.model.generate(inputs, + sampling_kwargs, + lora_request=LoRARequest( + 'sql_adapter', 1, + self.lora_path)) prompt_list, output_strs = [], [] for output in outputs: