Multi-Level Head-Wise Match and Aggregation in Transformer for Textual Sequence Matching
2020 ◽
Vol 34
(05)
◽
pp. 9209-9216
Keyword(s):
Transformer has been successfully applied to many natural language processing tasks. However, for textual sequence matching, simple matching between the representation of a pair of sequences might bring in unnecessary noise. In this paper, we propose a new approach to sequence pair matching with Transformer, by learning head-wise matching representations on multiple levels. Experiments show that our proposed approach can achieve new state-of-the-art performance on multiple tasks that rely only on pre-computed sequence-vector-representation, such as SNLI, MNLI-match, MNLI-mismatch, QQP, and SQuAD-binary.
2017 ◽
Vol 5
◽
pp. 135-146
◽
2019 ◽
Vol 33
◽
pp. 9259-9266
◽
Keyword(s):
2009 ◽
Vol 19
(1)
◽
pp. 85-96
◽
Keyword(s):
2020 ◽
Vol 34
(05)
◽
pp. 9250-9257
2020 ◽
Vol 34
(07)
◽
pp. 11418-11425
◽