from opencompass.models import Llama2Chat # Please follow the instruction in the Meta AI website https://github.com/facebookresearch/llama # and download the LLaMA-2-Chat model and tokenizer to the path './models/llama2/llama/'. # # The LLaMA requirement is also needed to be installed. # # git clone https://github.com/facebookresearch/llama.git # cd llama # pip install -e . api_meta_template = dict( round=[ dict(role='HUMAN', api_role='HUMAN'), dict(role='BOT', api_role='BOT', generate=True), ], ) models = [ dict( abbr='llama-2-70b-chat', type=Llama2Chat, path='./models/llama2/llama/llama-2-70b-chat/', tokenizer_path='./models/llama2/llama/tokenizer.model', meta_template=api_meta_template, max_out_len=100, max_seq_len=2048, batch_size=16, run_cfg=dict(num_gpus=8, num_procs=8), ), ]