mirror of
https://github.com/open-compass/opencompass.git
synced 2025-05-30 16:03:24 +08:00
[Feat] refine docs and codes for more user guides (#409)
This commit is contained in:
parent
a11cb45c83
commit
2c15a0c01d
@ -18,8 +18,11 @@ truthfulqa_infer_cfg = dict(
|
||||
# Metrics such as 'truth' and 'info' needs
|
||||
# OPENAI_API_KEY with finetuned models in it.
|
||||
# Please use your own finetuned openai model with keys and refers to
|
||||
# the source code for more details
|
||||
# Metrics such as 'bleurt', 'rouge', 'bleu' are free to test
|
||||
# the source code of `TruthfulQAEvaluator` for more details.
|
||||
#
|
||||
# If you cannot provide available models for 'truth' and 'info',
|
||||
# and want to perform basic metric eval, please set
|
||||
# `metrics=('bleurt', 'rouge', 'bleu')`
|
||||
|
||||
# When key is set to "ENV", the key will be fetched from the environment
|
||||
# variable $OPENAI_API_KEY. Otherwise, set key in here directly.
|
||||
|
@ -20,8 +20,11 @@ truthfulqa_infer_cfg = dict(
|
||||
# Metrics such as 'truth' and 'info' needs
|
||||
# OPENAI_API_KEY with finetuned models in it.
|
||||
# Please use your own finetuned openai model with keys and refers to
|
||||
# the source code for more details
|
||||
# Metrics such as 'bleurt', 'rouge', 'bleu' are free to test
|
||||
# the source code of `TruthfulQAEvaluator` for more details.
|
||||
#
|
||||
# If you cannot provide available models for 'truth' and 'info',
|
||||
# and want to perform basic metric eval, please set
|
||||
# `metrics=('bleurt', 'rouge', 'bleu')`
|
||||
|
||||
# When key is set to "ENV", the key will be fetched from the environment
|
||||
# variable $OPENAI_API_KEY. Otherwise, set key in here directly.
|
||||
|
@ -39,7 +39,9 @@ class TruthfulQAEvaluator(BaseEvaluator):
|
||||
|
||||
Args:
|
||||
truth_model (str): Truth model name. See "notes" for details.
|
||||
Defaults to ''.
|
||||
info_model (str): Informativeness model name. See "notes" for details.
|
||||
Defaults to ''.
|
||||
metrics (tuple): Computing needed metrics for truthfulqa dataset.
|
||||
Supported metrics are `bleurt`, `rouge`, `bleu`, `truth`, `info`.
|
||||
key (str): Corresponding API key. If set to `ENV`, find it in
|
||||
@ -67,10 +69,9 @@ class TruthfulQAEvaluator(BaseEvaluator):
|
||||
'bleu': 'bleu',
|
||||
}
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
truth_model: str, # noqa
|
||||
info_model: str, # noqa
|
||||
def __init__(self,
|
||||
truth_model: str = '',
|
||||
info_model: str = '',
|
||||
metrics=('bleurt', 'rouge', 'bleu', 'truth', 'info'),
|
||||
key='ENV'):
|
||||
self.API_MODEL = {
|
||||
@ -85,6 +86,11 @@ class TruthfulQAEvaluator(BaseEvaluator):
|
||||
if metric in self.SCORE_KEY.keys():
|
||||
self.metrics.append(metric)
|
||||
if metric in self.API_MODEL.keys():
|
||||
assert self.API_MODEL.get(metric), \
|
||||
f'`{metric}_model` should be set to perform API eval.' \
|
||||
'If you want to perform basic metric eval, ' \
|
||||
f'please refer to the docstring of {__file__} ' \
|
||||
'for more details.'
|
||||
self.api_metrics.append(metric)
|
||||
|
||||
if self.api_metrics:
|
||||
|
Loading…
Reference in New Issue
Block a user