| |
| """ |
| Custom evaluation tasks for lighteval. Complete this task with your own configuration if you want to use a custom lighteval task. |
| |
| This file generally create just a TASKS_TABLE and TASKS_GROUPS which are then imported by LightEval. |
| |
| Author: |
| """ |
|
|
| from lighteval.tasks.lighteval_task import LightevalTaskConfig |
| from lighteval.tasks.requests import Doc |
|
|
|
|
| |
| |
| |
| task = LightevalTaskConfig( |
| name="myothertask", |
| prompt_function="prompt_fn", |
| suite=["community"], |
| hf_repo="", |
| hf_subset="default", |
| hf_avail_splits=[], |
| evaluation_splits=[], |
| few_shots_split="", |
| few_shots_select="", |
| metric=[""], |
| ) |
|
|
| |
| |
| |
|
|
| |
| SAMPLE_SUBSETS = [] |
| |
|
|
|
|
| class CustomSubsetTask(LightevalTaskConfig): |
| def __init__( |
| self, |
| name, |
| hf_subset, |
| ): |
| super().__init__( |
| name=name, |
| hf_subset=hf_subset, |
| prompt_function="prompt_fn", |
| hf_repo="", |
| metric=[""], |
| hf_avail_splits=[], |
| evaluation_splits=[], |
| few_shots_split="", |
| few_shots_select="", |
| suite=["community"], |
| generation_size=-1, |
| stop_sequence=None, |
| output_regex=None, |
| frozen=False, |
| ) |
|
|
|
|
| |
| |
| def prompt_fn(line, task_name: str = None): |
| """Defines how to go from a dataset line to a doc object. |
| Follow examples in src/lighteval/tasks/tasks_prompt_formatting.py, or get more info |
| about what this function should do in the README. |
| """ |
| return Doc( |
| task_name=task_name, |
| query="", |
| choices=[], |
| gold_index=0, |
| instruction="", |
| ) |
|
|
|
|
| |
| SUBSET_TASKS = [CustomSubsetTask(name=f"mytask:{subset}", hf_subset=subset) for subset in SAMPLE_SUBSETS] |
| _TASKS = SUBSET_TASKS + [task] |
|
|
| |
| |
| |
| TASKS_TABLE = [task.as_dict() for task in _TASKS] |
|
|
| if __name__ == "__main__": |
| print(t["name"] for t in TASKS_TABLE) |
| print(len(TASKS_TABLE)) |
|
|