
* update crb * update crbbench * update crbbench * update crbbench * minor update wildbench * [Fix] Update doc of wildbench, and merge wildbench into subjective * [Fix] Update doc of wildbench, and merge wildbench into subjective, fix crbbench * Update crb.md * Update crb_pair_judge.py * Update crb_single_judge.py * Update subjective_evaluation.md * Update openai_api.py * [Update] update wildbench readme * [Update] update wildbench readme * [Update] update wildbench readme, remove crb * Delete configs/eval_subjective_wildbench_pair.py * Delete configs/eval_subjective_wildbench_single.py * Update __init__.py --------- Co-authored-by: bittersweet1999 <148421775+bittersweet1999@users.noreply.github.com>
1.2 KiB
Wildbench
Prepare the dataset
We support the wildbench dataset, developed by Lin et al. Please refer to their repo for more detail.
You have to download our preprocessed dataset. The format of dir should be like:
wildbench
---wildbench.jsonl
---gpt4
------wildbench.json
---claude
------wildbench.json
---llama2-70b
------wildbench.json
The wildbench.jsonl is the preprocessed dataset, and the other three are the reference, used for score.
Once you download the dataset, you have to modify the path defined in configs/datasets/subjective/wildbench/wildbench_pair_judge.py
and configs/datasets/subjective/wildbench/wildbench_single_judge.py
.
Note that you have to modify the given_preds in line 57-61 of configs/datasets/subjective/wildbench/wildbench_pair_judge.py
too,
Run
We have provide the script for wildbench in configs/eval_subjective_wildbench_pair.py
and configs/eval_subjective_wildbench_single.py
.
Note that if you test the wildbench with other models, please set the max_out_lens to 4096.
Acknowledgement
We greatly appreciate the authors of wildbench dataset. If you find it is useful in your research, please consider cite them.