Conversational machine reading comprehension (CMRC) requires models to effectively combine dialogue historyandanswer current questions. Previous works have shortcomings in handling historical information as they did not consider the role of historical questions in the learning process. Moreover, in the reasoning process, parallel input of multiplerounds of dialogue does not conform to human reasoning habits. Therefore, to address these limitations, this paperproposes the HistoryintoFlow model. In our model, we incorporate historical questions into the encoding layer, whichenables the model to extract complete historical information. In the reasoning layer of the model, we designa flowmodule that integrates intermediate representations generated from past conversations and performs reasoninginaccordance with the order of conversations. The final results show that the HistoryintoFlow model achieves an accuracyrate of 67.1% on the QuAC. Compared with some publicly available models, our model has improved in F1, HEQ-Q, and HEQ-D.
|