Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about the evaluation of OpenBookQA #2610

Open
xumingyu2021 opened this issue Jan 6, 2025 · 0 comments
Open

Question about the evaluation of OpenBookQA #2610

xumingyu2021 opened this issue Jan 6, 2025 · 0 comments
Labels
asking questions For asking for clarification / support on library usage.

Comments

@xumingyu2021
Copy link

It seems that the fact is not provided to give a answer, as a openbook qa, it may be better to use the fact, otherwise it become "closebookqa"
https://github.com/EleutherAI/lm-evaluation-harness/blob/main/lm_eval/tasks/openbookqa/openbookqa.yaml

task: openbookqa
dataset_path: openbookqa
dataset_name: main #here maybe additional
output_type: multiple_choice
training_split: train
validation_split: validation
test_split: test
doc_to_text: question_stem #here may need use fact1
doc_to_target: "{{choices.label.index(answerKey.lstrip())}}"
doc_to_choice: "{{choices.text}}"
should_decontaminate: true
doc_to_decontamination_query: question_stem
metric_list:

  • metric: acc
    aggregation: mean
    higher_is_better: true
  • metric: acc_norm
    aggregation: mean
    higher_is_better: true
    metadata:
    version: 1.0
@baberabb baberabb added the asking questions For asking for clarification / support on library usage. label Jan 7, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
asking questions For asking for clarification / support on library usage.
Projects
None yet
Development

No branches or pull requests

2 participants