2024-05-28 23:09:59 +08:00
|
|
|
# TriviaQA
|
|
|
|
|
2024-05-30 00:06:39 +08:00
|
|
|
```bash
|
|
|
|
python3 run.py --models hf_internlm2_7b --datasets triviaqa_wiki_1shot_gen_20a989 --debug
|
|
|
|
python3 run.py --models hf_internlm2_chat_7b --datasets triviaqa_wiki_1shot_gen_eaf81e --debug
|
|
|
|
```
|
|
|
|
|
2024-05-28 23:09:59 +08:00
|
|
|
## Base Models
|
|
|
|
|
|
|
|
| model | triviaqa |
|
|
|
|
|:------------------------:|-----------:|
|
|
|
|
| llama-7b-turbomind | 40.39 |
|
|
|
|
| llama-13b-turbomind | 66.41 |
|
|
|
|
| llama-30b-turbomind | 75.90 |
|
|
|
|
| llama-65b-turbomind | 82.26 |
|
|
|
|
| llama-2-7b-turbomind | 43.21 |
|
|
|
|
| llama-2-13b-turbomind | 71.32 |
|
|
|
|
| llama-2-70b-turbomind | 67.45 |
|
|
|
|
| llama-3-8b-turbomind | 71.24 |
|
|
|
|
| llama-3-70b-turbomind | 88.16 |
|
|
|
|
| internlm2-1.8b-turbomind | 38.42 |
|
|
|
|
| internlm2-7b-turbomind | 69.15 |
|
|
|
|
| internlm2-20b-turbomind | 74.03 |
|
|
|
|
| qwen-1.8b-turbomind | 22.76 |
|
|
|
|
| qwen-7b-turbomind | 53.61 |
|
|
|
|
| qwen-14b-turbomind | 49.72 |
|
|
|
|
| qwen-72b-turbomind | 79.13 |
|
|
|
|
| qwen1.5-0.5b-hf | 21.24 |
|
|
|
|
| qwen1.5-1.8b-hf | 34.32 |
|
|
|
|
| qwen1.5-4b-hf | 44.59 |
|
|
|
|
| qwen1.5-7b-hf | 56.60 |
|
|
|
|
| qwen1.5-14b-hf | 59.96 |
|
|
|
|
| qwen1.5-32b-hf | 56.20 |
|
|
|
|
| qwen1.5-72b-hf | 77.81 |
|
|
|
|
| qwen1.5-moe-a2-7b-hf | 65.49 |
|
|
|
|
| mistral-7b-v0.1-hf | 72.93 |
|
|
|
|
| mistral-7b-v0.2-hf | 70.91 |
|
|
|
|
| mixtral-8x7b-v0.1-hf | 85.05 |
|
|
|
|
| mixtral-8x22b-v0.1-hf | 89.47 |
|
|
|
|
| yi-6b-hf | 23.76 |
|
|
|
|
| yi-34b-hf | 14.73 |
|
|
|
|
| deepseek-7b-base-hf | 59.48 |
|
|
|
|
| deepseek-67b-base-hf | 72.15 |
|
|
|
|
|
|
|
|
## Chat Models
|
|
|
|
|
|
|
|
| model | triviaqa |
|
|
|
|
|:-----------------------------:|-----------:|
|
|
|
|
| qwen1.5-0.5b-chat-hf | 19.84 |
|
|
|
|
| qwen1.5-1.8b-chat-hf | 35.81 |
|
|
|
|
| qwen1.5-4b-chat-hf | 48.93 |
|
|
|
|
| qwen1.5-7b-chat-hf | 53.65 |
|
|
|
|
| qwen1.5-14b-chat-hf | 62.58 |
|
|
|
|
| qwen1.5-32b-chat-hf | 74.72 |
|
|
|
|
| qwen1.5-72b-chat-hf | 83.25 |
|
|
|
|
| qwen1.5-110b-chat-hf | 86.20 |
|
|
|
|
| internlm2-chat-1.8b-hf | 46.69 |
|
|
|
|
| internlm2-chat-1.8b-sft-hf | 46.50 |
|
|
|
|
| internlm2-chat-7b-hf | 69.54 |
|
|
|
|
| internlm2-chat-7b-sft-hf | 70.75 |
|
|
|
|
| internlm2-chat-20b-hf | 75.53 |
|
|
|
|
| internlm2-chat-20b-sft-hf | 75.90 |
|
|
|
|
| llama-3-8b-instruct-hf | 78.99 |
|
|
|
|
| llama-3-70b-instruct-hf | 89.79 |
|
|
|
|
| llama-3-8b-instruct-lmdeploy | 76.77 |
|
|
|
|
| llama-3-70b-instruct-lmdeploy | 89.62 |
|
|
|
|
| mistral-7b-instruct-v0.1-hf | 62.94 |
|
|
|
|
| mistral-7b-instruct-v0.2-hf | 67.72 |
|
|
|
|
| mixtral-8x7b-instruct-v0.1-hf | 79.57 |
|