Zephyrnet Logo

Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-based Question Answering. (arXiv:2004.03561v1 [cs.CL])

Date:

(Submitted on 7 Apr 2020)

Abstract: We introduce a novel approach to transformers that learns hierarchical
representations in multiparty dialogue. First, three language modeling tasks
are used to pre-train the transformers, token- and utterance-level language
modeling and utterance order prediction, that learn both token and utterance
embeddings for better understanding in dialogue contexts. Then, multi-task
learning between the utterance prediction and the token span prediction is
applied to fine-tune for span-based question answering (QA). Our approach is
evaluated on the FriendsQA dataset and shows improvements of 3.8% and 1.4% over
the two state-of-the-art transformer models, BERT and RoBERTa, respectively.

Submission history

From: Changmao Li [view email]
[v1]
Tue, 7 Apr 2020 17:36:33 UTC (377 KB)

Source: http://arxiv.org/abs/2004.03561

spot_img

Latest Intelligence

spot_img

Chat with us

Hi there! How can I help you?