Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence 2020
DOI: 10.24963/ijcai.2020/540
|View full text |Cite
|
Sign up to set email alerts
|

Multi-hop Reading Comprehension across Documents with Path-based Graph Convolutional Network

Abstract: Multi-hop reading comprehension across multiple documents attracts much attentions recently. In this paper, we propose a novel approach to tackle this multi-hop reading comprehension problem. Inspired by the human reasoning processing, we introduce a path-based graph with reasoning paths which extracted from supporting documents. The path-based graph can combine both the idea of the graph-based and path-based approaches, so it is better for multi-hop reasoning. Meanwhile, we propose Gated-GCN to accum… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(12 citation statements)
references
References 2 publications
2
10
0
Order By: Relevance
“…In Table II we present the performances of ClueReader in the development and test sets of WIKIHOP and MEDHOP, and compare them with previously published models mainly basing on GNNs. Our model has improved the accuracy of the HDE based on the heterogeneous GCNs in test set from 70.9% to 72.0% and Path-based GCN (with GloVe word embedding setting) in dev set from 64.5% to 66.9%, while Path-based GCN using GloVe and ELMo surpassed our model by 0.5% in the test set, which confirms the conclusion that the initial representations of nodes is extremely critical [25]. However, limited by the architecture and computing resources, we have not used the powerful contextual word embedding like the ELMo and the BERT in our model which can be further addressed.…”
Section: Results and Analysessupporting
confidence: 78%
See 2 more Smart Citations
“…In Table II we present the performances of ClueReader in the development and test sets of WIKIHOP and MEDHOP, and compare them with previously published models mainly basing on GNNs. Our model has improved the accuracy of the HDE based on the heterogeneous GCNs in test set from 70.9% to 72.0% and Path-based GCN (with GloVe word embedding setting) in dev set from 64.5% to 66.9%, while Path-based GCN using GloVe and ELMo surpassed our model by 0.5% in the test set, which confirms the conclusion that the initial representations of nodes is extremely critical [25]. However, limited by the architecture and computing resources, we have not used the powerful contextual word embedding like the ELMo and the BERT in our model which can be further addressed.…”
Section: Results and Analysessupporting
confidence: 78%
“…And on account of full usage of the question's contextual information, it applied the bi-directional attention mechanism, both node2query and query2node, which aimed to obtain the query-aware nodes' representations in the reasoning graph for better predictions. And Path-based GCN [25] introduced related entities in graph more than the nodes merely match to the candidates to enhance the performance of the model. Furthermore, the HDE [26] introduced the heterogeneous nodes into GCNs, which contains different granularity levels of information.…”
Section: B Graph Neural Network For Multi-hop Mrcmentioning
confidence: 99%
See 1 more Smart Citation
“…We also compare our model with two state-ofthe-art GNN models (i.e., SAE and HGN), shown in Table 2. Both of them need to set the number of GNN layers manually while BFR-Graph can pass through all the connected nodes automatically with Model Accuracy HDE (Tu et al, 2019) 68.1 DynSAN (Zhuang and Wang, 2019) 70.1 Path-based GCN (Tang et al, 2020) 70.8 ChainEx 72.2 Longformer * 73.8 Longformer+BFR 74.4 an extremely low risk of over-smoothing (Kipf and Welling, 2017). SAE and HGN set a fixed types of edges, which is still not fine-grained enough, while BFR-Graph define different weights (can up to ∞ different weights depends on the dataset) to distinguish nodes in a finer granularity.…”
Section: Resultsmentioning
confidence: 99%
“…Gated Graph Neural Network (GGNNs) [43] was designed to encode the node feature with gated recurrent units. Tang et al [44] employed the Gated Relational Graph Convolution Network (Gated-RGCN) to aggregate messages on the path-based reasoning graph, where the attention and gate mechanisms were employed to adjust the usefulness of information propagating across the graph.…”
Section: B Graph-based Reasoningmentioning
confidence: 99%