We explored recent studies in Question Answering System. Then tried out 3 different models for the sake of learning. Our steps were:
-
Firstly, we studied recent works on QA. More precisely, we studied Zylich et al.'s
Exploring Automated Question Answering Methods for Teaching Assistance
which is published in AIE conf. in 2020 Link. The summary of the paper is uploaded here. -
After that, we studied about
BERT
, what is the input-output format of it and how it works in case of QA. Then, tried out pretrained & fine-tunedBERT
model. ThisBERT
model is fine-tuned using SQuAD v1.1 dataset. Then viewed our output.
- Our used model was bert-large-uncased-whole-word-masking-finetuned-squad
- BERT paper
- Next, we studied about
DistilBERT
, which is a distiled version ofBERT
. It is more smaller, faster, cheaper and lighter thanBERT
. It doesn't have token ids likeBERT
, as a result is gives 70% more faster output thanBERT
. Gives almost accurate result likeBERT
. Our used model is pretrained and fined-tuned with same dataset asBERT
was. Then we compared the output withBERT
and verify the output.
- DistilBERT paper
- distilbert-base-uncased-distilled-squad
- Huggingface All Pretrained Model List
- All functions of DistilBERT is mentioned here
- Lastly, we wanted to use a
DistilBERT
pretrained model and fine-tuned it with a custom dataset SQuAD v2.0 trained dataset. Then tested our pretrained model with SQuAD v2.0 dev dataset and checked accuracy of the model.