Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017) 2017
DOI: 10.18653/v1/s17-2047
|View full text |Cite
|
Sign up to set email alerts
|

EICA Team at SemEval-2017 Task 3: Semantic and Metadata-based Features for Community Question Answering

Abstract: We describe our system for participating in SemEval-2017 Task 3 on Community Question Answering. Our approach relies on combining a rich set of various types of features: semantic and metadata. The most important types turned out to be the metadata feature and the semantic vectors trained on QatarLiving data. In the main Subtask C, our primary submission was ranked fourth, with a MAP of 13.48 and accuracy of 97.08. In Subtask A, our primary submission get into the top 50%.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 19 publications
0
8
0
Order By: Relevance
“…Table 1 summarizes the results of different methods on SemEval 2017 dataset. For our methods, the term "single" denotes that we only consider word-to-word matches as in Equation 14, while "multi" means that we consider Method MAP MRR Baseline (IR) 9.18 10.11 Baseline (random) 5.77 7.69 (Tian et al 2017) 10.64 11.09 (Zhang et al 2017a) 13.23 14.27 (Xie et al 2017) 13.48 16.04 (Filice, Da Martino, and Moschitti 2017) 14.35 16.07 (Koreeda et al 2017) 14.71 16.48 (Nandi et al 2017) 15 (Filice, Da Martino, and Moschitti 2017;Xie et al 2017;Nandi et al 2017) and neural networks (Tian et al 2017;Zhang et al 2017a;Koreeda et al 2017). For singlescale model, the MAP is increased from 14.67 to 17.25, while for multi-scale model, the number is increased from 14.80 to 17.91.…”
Section: Training Hyper-parametersmentioning
confidence: 99%
“…Table 1 summarizes the results of different methods on SemEval 2017 dataset. For our methods, the term "single" denotes that we only consider word-to-word matches as in Equation 14, while "multi" means that we consider Method MAP MRR Baseline (IR) 9.18 10.11 Baseline (random) 5.77 7.69 (Tian et al 2017) 10.64 11.09 (Zhang et al 2017a) 13.23 14.27 (Xie et al 2017) 13.48 16.04 (Filice, Da Martino, and Moschitti 2017) 14.35 16.07 (Koreeda et al 2017) 14.71 16.48 (Nandi et al 2017) 15 (Filice, Da Martino, and Moschitti 2017;Xie et al 2017;Nandi et al 2017) and neural networks (Tian et al 2017;Zhang et al 2017a;Koreeda et al 2017). For singlescale model, the MAP is increased from 14.67 to 17.25, while for multi-scale model, the number is increased from 14.80 to 17.91.…”
Section: Training Hyper-parametersmentioning
confidence: 99%
“…For classification tasks like question similarity across community QA forums, machine learning classification algorithms like Support Vector Machines (SVMs) have been used (Šaina et al, 2017;Nandi et al, 2017;Xie et al, 2017;Mihaylova et al, 2016;Wang and Poupart, 2016;. Recently, advances in deep neural network architectures have also led to the use of Convolutional Neural Networks (CNNs) (Šaina et al, 2017;Mohtarami et al, 2016) which perform reasonably well for selection of the correct answer amongst cQA formus.…”
Section: Related Workmentioning
confidence: 99%
“…Other works in the space include use of Random Forests (Wang and Poupart, 2016); topic models to match the questions at both the term level and topic level (Zhang et al, 2014). There have also been works on translation based retrieval models (Jeon et al, 2005;Zhou et al, 2011); Xg-Boost (Feng et al, 2017) and Feedforward Neural Networks (NN) (Wang and Poupart, 2016 (Wang and Poupart, 2016;Mohtarami et al, 2016;Nandi et al, 2017); and Metadata-based features (Mohtarami et al, 2016;Mihaylova et al, 2016;Xie et al, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…Some of the earlier works on cQA include the use of classification models -Support Vector Machines(SVMs) (Šaina et al, 2017;Nandi et al, 2017;Xie et al, 2017;Mihaylova et al, 2016;Wang and Poupart, 2016; for Similarity tasks; Convolutional Neural Networks (CNNs) for Similarity tasks (Šaina et al, 2017;Mohtarami et al, 2016) and for answer selection (Zhang et al, 2017); Long-Short Term Memory (LSTM) model for answer selection (Zhang et al, 2017;Feng et al, 2017;Mohtarami et al, 2016); Random Forests (Wang and Poupart, 2016); LDA topic language model to match the questions at both the term level and topic level (Zhang et al, 2014); translation based retrieval models (Jeon et al, 2005;Zhou et al, 2011); XgBoost (Feng et al, 2017) and Feedforward Neural Network (NN) (Wang and Poupart, 2016 (Mikolov et al, 2013), GloVe 6 (Pennington et al, 2014) etc.) (Wang and Poupart, 2016;Mohtarami et al, 2016;Nandi et al, 2017); Metadata-based features (like user information, answer length, question length, question marks in answer, question to comment length etc.)…”
Section: Related Workmentioning
confidence: 99%