Question answering through pretrained transformer-based models from Hugging Face.
Question Answering (QA) is the task consisting in generating answers for questions from the passages containing the needed information. Optionally, also the history of previous question.answer turns can be used for producing the answer.
The CoQA dataset has been used: https://stanfordnlp.github.io/coqa/.
In our work, we use a model which consists of two modules: the tokens importances extractor and the encoder-decoder (i.e. seq2seq). The first module computes an importance score in
Two different pre-trained models have been considered, namely DistilRoBERTa and BERTTiny. Different random seeds have been set for generating our experiments. Finally, also whether to use or not the conversation history has been taken into account.
For evaluating these different experiments, the average SQuAD F1 score has been computed, both on the validation and test datasets.
.
├── coqa # It contains the dataset files
├── images # It contains some explanatory images
├── models # It contains the models
├── utils # It contains the python files with useful functions
├── weigths # It contains the models weigths
├── Assignment.ipynb # Task description
├── question answering.ipynb # Task resolution
├── .gitignore
├── LICENSE
├── report.pdf # Report of the assignment
└── README.md
Git is used for versioning.
Name | Surname | Username | |
---|---|---|---|
Samuele | Bortolato | [email protected] |
Sam |
Antonio | Politano | [email protected] |
S1082351 |
Enrico | Pittini | [email protected] |
EnricoPittini |
Riccardo | Spolaor | [email protected] |
RiccardoSpolaor |
This project is licensed under the MIT License - see the LICENSE file for details.