diff --git a/configs/models/hf_internlm_7b.py b/configs/models/hf_internlm_7b.py index aea4208e..31ec8484 100644 --- a/configs/models/hf_internlm_7b.py +++ b/configs/models/hf_internlm_7b.py @@ -5,12 +5,13 @@ models = [ dict( type=HuggingFaceCausalLM, abbr='internlm-7b-hf', - path="internlm-7b", - tokenizer_path='internlm-7b', + path="internlm/internlm-7b", + tokenizer_path='internlm/internlm-7b', tokenizer_kwargs=dict( padding_side='left', truncation_side='left', use_fast=False, + trust_remote_code=True, ), max_out_len=100, max_seq_len=2048, diff --git a/configs/models/hf_internlm_chat_7b.py b/configs/models/hf_internlm_chat_7b.py index 6491cd97..2d526f60 100644 --- a/configs/models/hf_internlm_chat_7b.py +++ b/configs/models/hf_internlm_chat_7b.py @@ -12,12 +12,13 @@ models = [ dict( type=HuggingFaceCausalLM, abbr='internlm-chat-7b-hf', - path="internlm-chat-7b", - tokenizer_path='internlm-chat-7b', + path="internlm/internlm-chat-7b", + tokenizer_path='internlm/internlm-chat-7b', tokenizer_kwargs=dict( padding_side='left', truncation_side='left', use_fast=False, + trust_remote_code=True, ), max_out_len=100, max_seq_len=2048, diff --git a/configs/models/hf_internlm_chat_7b_8k.py b/configs/models/hf_internlm_chat_7b_8k.py new file mode 100644 index 00000000..19b9a757 --- /dev/null +++ b/configs/models/hf_internlm_chat_7b_8k.py @@ -0,0 +1,30 @@ +from opencompass.models import HuggingFaceCausalLM + + +_meta_template = dict( + round=[ + dict(role='HUMAN', begin='<|User|>:', end='\n'), + dict(role='BOT', begin='<|Bot|>:', end='\n', generate=True), + ], +) + +models = [ + dict( + type=HuggingFaceCausalLM, + abbr='internlm-chat-7b-8k-hf', + path="internlm/internlm-chat-7b-8k", + tokenizer_path='internlm/internlm-chat-7b-8k', + tokenizer_kwargs=dict( + padding_side='left', + truncation_side='left', + use_fast=False, + trust_remote_code=True, + ), + max_out_len=100, + max_seq_len=2048, + batch_size=8, + meta_template=_meta_template, + model_kwargs=dict(trust_remote_code=True, device_map='auto'), + run_cfg=dict(num_gpus=1, num_procs=1), + ) +] diff --git a/configs/models/hf_llama_13b.py b/configs/models/hf_llama_13b.py index d1db1574..70fd6fea 100644 --- a/configs/models/hf_llama_13b.py +++ b/configs/models/hf_llama_13b.py @@ -5,6 +5,7 @@ models = [ # LLaMA 13B dict( type=HuggingFaceCausalLM, + abbr='llama-13b-hf', path="decapoda-research/llama-13b-hf", tokenizer_path='decapoda-research/llama-13b-hf', tokenizer_kwargs=dict(padding_side='left', diff --git a/configs/models/hf_vicuna_v1.3_13b.py b/configs/models/hf_vicuna_13b.py similarity index 100% rename from configs/models/hf_vicuna_v1.3_13b.py rename to configs/models/hf_vicuna_13b.py diff --git a/configs/models/hf_vicuna_v1.3_33b.py b/configs/models/hf_vicuna_33b.py similarity index 100% rename from configs/models/hf_vicuna_v1.3_33b.py rename to configs/models/hf_vicuna_33b.py diff --git a/configs/models/hf_vicuna_v1.3_7b.py b/configs/models/hf_vicuna_7b.py similarity index 100% rename from configs/models/hf_vicuna_v1.3_7b.py rename to configs/models/hf_vicuna_7b.py