What does BERT learn about questions
Recent research in Question Answering is highly motivated by the introduction of the BERT  model. This model has gained considerable attention since the researcher of Google AI Language has claimed state-of-the-art results over various NLP tasks, including QA. On one side, where the introduction of end-to-end pipeline models consisting of an IR and an RC model has opened the scope of research in two different areas, new BERT representations alone show a significant improvement in the performance of a QA system. In this study, we have covered several pipeline models like R3: Reinforced Ranker-Reader , Re-Ranker Model , and Interactive Retriever-Reader Model  along with the transformer-based QA system i.e., BERT. The motivation of this work is to deeply understand the black-box BERT model and try to identify the BERT’s learning about the question to predict the correct answer for it from a given context. We will discuss all the experiments that we have performed to understand BERT’s behavior from a different perspective. We have performed all the experiments using the SQuAD dataset. We have also used the LRP  technique to get a better understanding and for a better analysis of the experiment results. Along with the study about what the model learns, we have also tried to find what the model does not learn. For this, we have analyzed various examples from the dataset to determine the types of questions for whom the model predicts an incorrect answer. Finally, we have presented the overall findings of the BERT model in the conclusion section.
- M Tech Dissertations