--- language: - it language_details: it-IT license: cc-by-nc-sa-4.0 task_categories: - question-answering task_ids: - text-classification configs: - config_name: default data_files: - split: test_1 path: "multichoice_v1_test.jsonl" - split: dev_1 path: "multichoice_v1_dev.jsonl" - split: test_2 path: "multichoice_v2_test.jsonl" - split: dev_2 path: "multichoice_v2_dev.jsonl" size_categories: - n<1K --- ### QA4FAQ @ EVALITA 2016 Original dataset information available [here](http://qa4faq.github.io/) ## Data format The data has been converted to be used as a questin answering task. There are two splits, test-1 and test-2, each containing the same data processed in slightly different ways. ### test-1 The data is in jsonl format, where each line is a json object with the following fields: - `id`: a unique identifier for the question - `question`: the question - `A`, `B`, `C`, `D`: the possible answers to the question - `correct_answer`: correct answer to the question ('A', 'B', 'C', 'D') wrong answers are randomly drawn from the other question, answers pairs in the dataset. ### test-2 The data is in jsonl format, where each line is a json object with the following fields: - `id`: a unique identifier for the question - `question`: the question - `A`, `B`, `C`, `D`: the possible question,answers pairs e.g. (question, answer) - `correct_answer`: correct question,answer pair to the question ('A', 'B', 'C', 'D') wrong (q,a) pairs are randomly created by randomy choosing answers from the dataset. ## Publications ``` @inproceedings{agirre-etal-2015-semeval, title = "{S}em{E}val-2015 Task 2: Semantic Textual Similarity, {E}nglish, {S}panish and Pilot on Interpretability", author = "Agirre, Eneko and Banea, Carmen and Cardie, Claire and Cer, Daniel and Diab, Mona and Gonzalez-Agirre, Aitor and Guo, Weiwei and Lopez-Gazpio, I{\~n}igo and Maritxalar, Montse and Mihalcea, Rada and Rigau, German and Uria, Larraitz and Wiebe, Janyce", editor = "Nakov, Preslav and Zesch, Torsten and Cer, Daniel and Jurgens, David", booktitle = "Proceedings of the 9th International Workshop on Semantic Evaluation ({S}em{E}val 2015)", month = jun, year = "2015", address = "Denver, Colorado", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/S15-2045", doi = "10.18653/v1/S15-2045", pages = "252--263", } ``` ``` @inproceedings{nakov-etal-2015-semeval, title = "{S}em{E}val-2015 Task 3: Answer Selection in Community Question Answering", author = "Nakov, Preslav and M{\`a}rquez, Llu{\'\i}s and Magdy, Walid and Moschitti, Alessandro and Glass, Jim and Randeree, Bilal", editor = "Nakov, Preslav and Zesch, Torsten and Cer, Daniel and Jurgens, David", booktitle = "Proceedings of the 9th International Workshop on Semantic Evaluation ({S}em{E}val 2015)", month = jun, year = "2015", address = "Denver, Colorado", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/S15-2047", doi = "10.18653/v1/S15-2047", pages = "269--281", } ```