OpenCompass/opencompass/datasets/lveval/lveval_factrecall_en.py
yuantao2108 bbec7d8733
[Feature] add lveval benchmark (#914)
* add lveval benchmark

* add LVEval readme file

* update LVEval readme file

* Update configs/eval_bluelm_32k_lveval.py

* Update configs/eval_llama2_7b_lveval.py

---------

Co-authored-by: yuantao <yuantao@infini-ai.com>
Co-authored-by: Mo Li <82895469+DseidLi@users.noreply.github.com>
2024-03-04 11:22:03 +08:00

29 lines
875 B
Python

from datasets import Dataset, load_dataset
from opencompass.registry import LOAD_DATASET
from ..base import BaseDataset
@LOAD_DATASET.register_module()
class LVEvalfactrecallenDataset(BaseDataset):
@staticmethod
def load(**kwargs):
dataset = load_dataset(**kwargs)
split = 'test'
raw_data = []
for i in range(len(dataset[split])):
question = dataset[split]['input'][i]
context = dataset[split]['context'][i]
answers = dataset[split]['answers'][i]
confusing_facts = dataset[split]['confusing_facts'][i]
raw_data.append({
'input': question,
'context': context,
'answers': answers,
'confusing_facts': confusing_facts,
})
dataset[split] = Dataset.from_list(raw_data)
return dataset