facts-grounding-prompts / grounding_accuracy_response_level.yaml
MoritzLaurer's picture
Upload prompt template grounding_accuracy_response_level.yaml
417d7df verified
raw
history blame
1.23 kB
prompt:
template: "Your task is to check if the Response is accurate to the Evidence.\nGenerate 'Accurate' if the Response is accurate
when verified according to the Evidence, or 'Inaccurate' if the Response is inaccurate (contradicts the evidence) or cannot
be verified.\n\n**Query**:\n\n{{user_request}}\n\n**End of Query**\n\n**Evidence**\n\n{{context_document}}\n\n**End of
Evidence**\n\n**Response**:\n\n{{response}}\n\n**End of Response**\n\nLet's think step-by-step."
template_variables:
- user_request
- context_document
- response
metadata:
description: "An evaluation prompt from the paper 'The FACTS Grounding Leaderboard: Benchmarking LLMs’ Ability to Ground
Responses to Long-Form Input' by Google DeepMind.\n The prompt was copied from the evaluation_prompts.csv file from
Kaggle.\n This specific prompt elicits a binary accurate/inaccurate classifier for the entire response."
evaluation_method: response_level
tags:
- fact-checking
version: 1.0.0
author: Google DeepMind
source: https://www.kaggle.com/datasets/deepmind/FACTS-grounding-examples?resource=download&select=evaluation_prompts.csv
client_parameters: {}
custom_data: {}