mirror of
https://github.com/open-compass/opencompass.git
synced 2025-05-30 16:03:24 +08:00

* Update JuderBench * Support O1-style Prompts * Update Code * Update OpenAI * Update BigCodeBench * Update BigCodeBench * Update BigCodeBench * Update BigCodeBench * Update BigCodeBench * Update
53 lines
1.5 KiB
Python
53 lines
1.5 KiB
Python
from opencompass.openicl.icl_prompt_template import PromptTemplate
|
|
from opencompass.openicl.icl_retriever import ZeroRetriever
|
|
from opencompass.openicl.icl_inferencer import GenInferencer
|
|
from opencompass.datasets import (
|
|
BigCodeBenchDataset,
|
|
BigCodeBenchEvaluator
|
|
)
|
|
|
|
|
|
bigcodebench_full_reader_cfg = dict(
|
|
input_columns=['complete_prompt'],
|
|
output_column='test',
|
|
)
|
|
|
|
|
|
bigcodebench_full_infer_cfg = dict(
|
|
prompt_template=dict(
|
|
type=PromptTemplate,
|
|
template=dict(
|
|
begin=[dict(role='system',
|
|
fallback_role='HUMAN',
|
|
prompt='')],
|
|
round=[
|
|
dict(role='HUMAN', prompt='{complete_prompt}'),
|
|
]
|
|
)
|
|
),
|
|
retriever=dict(type=ZeroRetriever),
|
|
inferencer=dict(type=GenInferencer, max_out_len=1024)
|
|
)
|
|
|
|
bigcodebench_full_eval_cfg = dict(
|
|
evaluator=dict(
|
|
type=BigCodeBenchEvaluator,
|
|
release_version='v0.1.2',
|
|
eval_type='complete',
|
|
remote_execute_api="https://bigcode-bigcodebench-evaluator.hf.space/",
|
|
dataset_version='full',
|
|
),
|
|
pred_role='BOT',
|
|
)
|
|
|
|
bigcodebench_full_complete_datasets = [
|
|
dict(
|
|
abbr='bigcodebench_full_complete',
|
|
type=BigCodeBenchDataset,
|
|
path="opencompass/bigcodebench",
|
|
reader_cfg=bigcodebench_full_reader_cfg,
|
|
infer_cfg=bigcodebench_full_infer_cfg,
|
|
eval_cfg=bigcodebench_full_eval_cfg,
|
|
release_version='v0.1.2'
|
|
)
|
|
] |