Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017) 2017
DOI: 10.18653/v1/s17-2009
|View full text |Cite
|
Sign up to set email alerts
|

IIT-UHH at SemEval-2017 Task 3: Exploring Multiple Features for Community Question Answering and Implicit Dialogue Identification

Abstract: In this paper we present the system for Answer Selection and Ranking in Community Question Answering, which we build as part of our participation in SemEval-2017 Task 3. We develop a Support Vector Machine (SVM) based system that makes use of textual, domain-specific, wordembedding and topic-modeling features. In addition, we propose a novel method for dialogue chain identification in comment threads. Our primary submission won subtask C, outperforming other systems in all the primary evaluation metrics. We pe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(10 citation statements)
references
References 12 publications
0
10
0
Order By: Relevance
“…Extracting Keyphrases Keyphrases are extracted from clarification questions using the RAKE algorithm [16], which is an efficient way to find noun phrases. This algorithm has been used in a similar setting where CQA comments should be matched to related questions [12]. We tokenize the keyphrases and consider each token individually.…”
Section: System Componentsmentioning
confidence: 99%
“…Extracting Keyphrases Keyphrases are extracted from clarification questions using the RAKE algorithm [16], which is an efficient way to find noun phrases. This algorithm has been used in a similar setting where CQA comments should be matched to related questions [12]. We tokenize the keyphrases and consider each token individually.…”
Section: System Componentsmentioning
confidence: 99%
“…Table 1 summarizes the results of different methods on SemEval 2017 dataset. For our methods, the term "single" denotes that we only consider word-to-word matches as in Equation 14, while "multi" means that we consider Method MAP MRR Baseline (IR) 9.18 10.11 Baseline (random) 5.77 7.69 (Tian et al 2017) 10.64 11.09 (Zhang et al 2017a) 13.23 14.27 (Xie et al 2017) 13.48 16.04 (Filice, Da Martino, and Moschitti 2017) 14.35 16.07 (Koreeda et al 2017) 14.71 16.48 (Nandi et al 2017) 15 (Filice, Da Martino, and Moschitti 2017;Xie et al 2017;Nandi et al 2017) and neural networks (Tian et al 2017;Zhang et al 2017a;Koreeda et al 2017). For singlescale model, the MAP is increased from 14.67 to 17.25, while for multi-scale model, the number is increased from 14.80 to 17.91.…”
Section: Training Hyper-parametersmentioning
confidence: 99%
“…For classification tasks like question similarity across community QA forums, machine learning classification algorithms like Support Vector Machines (SVMs) have been used (Šaina et al, 2017;Nandi et al, 2017;Xie et al, 2017;Mihaylova et al, 2016;Wang and Poupart, 2016;. Recently, advances in deep neural network architectures have also led to the use of Convolutional Neural Networks (CNNs) (Šaina et al, 2017;Mohtarami et al, 2016) which perform reasonably well for selection of the correct answer amongst cQA formus.…”
Section: Related Workmentioning
confidence: 99%
“…Other works in the space include use of Random Forests (Wang and Poupart, 2016); topic models to match the questions at both the term level and topic level (Zhang et al, 2014). There have also been works on translation based retrieval models (Jeon et al, 2005;Zhou et al, 2011); Xg-Boost (Feng et al, 2017) and Feedforward Neural Networks (NN) (Wang and Poupart, 2016 (Wang and Poupart, 2016;Mohtarami et al, 2016;Nandi et al, 2017); and Metadata-based features (Mohtarami et al, 2016;Mihaylova et al, 2016;Xie et al, 2017).…”
Section: Related Workmentioning
confidence: 99%