Wenping Wang
Adv. Artif. Intell. Mach. Learn., 3 (3):1325-1339
Wenping Wang : Individual Researcher
DOI: 10.54364/AAIML.2023.1178
Article History: Received on: 01-Jun-23, Accepted on: 12-Aug-23, Published on: 19-Aug-23
Corresponding Author: Wenping Wang
Email: wenpingw@alumni.cmu.edu
Citation: Longxiang Zhang, Wenping Wang, Keyi Yu, Jingxian Huang, Qi Lyu, Haoru Xue, Congrui Hetang (2023). Sliding-BERT: Striding Towards Conversational Machine Comprehension in Long Context. Adv. Artif. Intell. Mach. Learn., 3 (3 ):1325-1339
Pre-trained contextual embeddings like BERT have shown substantial improvement across a wide range of natural language processing tasks.
% Advent of BERT revolutionized the domain of NLP by beating state-of-the-art results in various tasks.
We proposed Sliding-BERT, which incorporates BERT with state-of-the-art conversational machine comprehension (MC) model, \textsc{FlowQA}, and supersedes its standing performance on the \daffy challenge. We designed a striding filter to overcome the sequence length limit of BERT model in the long conversation context. We also applied various aggregation methods to handle the incompatible tokenization between BERT and \textsc{FlowQA} models. Given the long conversation context, we used gradient accumulation to simulate batched training scenarios without extra memory cost during training. We also found that pretraining our Sliding-BERT on CoQA dataset helps improve its performance on \daffy dataset. Detailed analysis of the model performance considering the types of questions, lengths of questions and other metrics of QA datasets indicates that our Sliding-BERT exceeds \textsc{FlowQA} model in terms of F1, HEQ-Q, and HEQ-D scores by a significant margin.