2011
DOI: 10.1007/978-3-642-19437-5_19
|View full text |Cite
|
Sign up to set email alerts
|

Comparing Manual Text Patterns and Machine Learning for Classification of E-Mails for Automatic Answering by a Government Agency

Abstract: Abstract. E-mails to government institutions as well as to large companies may contain a large proportion of queries that can be answered in a uniform way. We analysed and manually annotated 4,404 e-mails from citizens to the Swedish Social Insurance Agency, and compared two methods for detecting answerable e-mails: manually-created text patterns (rule-based) and machine learning-based methods. We found that the text pattern-based method gave much higher precision at 89 percent than the machine learning-based … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2012
2012
2020
2020

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 6 publications
0
5
0
Order By: Relevance
“…An alternative to SVM could be Naïve Bayes, which is also implemented in WEKA. In a previously conducted test A‐B for SVM and Naïve Bayes (Dalianis et al, ), both techniques had similar performance whereas SVM had a slightly better precision; we prioritize precision. In some older experiments with answer‐specific email categorization (Busemann, Schmeier, & Arens, ; Scheffer, ), SVM outperformed Naïve Bayes.…”
Section: Evaluation Methodsmentioning
confidence: 98%
See 3 more Smart Citations
“…An alternative to SVM could be Naïve Bayes, which is also implemented in WEKA. In a previously conducted test A‐B for SVM and Naïve Bayes (Dalianis et al, ), both techniques had similar performance whereas SVM had a slightly better precision; we prioritize precision. In some older experiments with answer‐specific email categorization (Busemann, Schmeier, & Arens, ; Scheffer, ), SVM outperformed Naïve Bayes.…”
Section: Evaluation Methodsmentioning
confidence: 98%
“…Finally, we manually labelled each message according to its text category. Preprocessing of the messages is covered in Dalianis et al ().…”
Section: Evaluation Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Neural network based approaches are strong alternatives but usually less interpretable because those black box models cannot be logically explained [15]. In addition, those black box models cannot be quickly modified except retraining models [16]. To address those difficult issues discussed above, some related work has been done by using regular expressions for classification tasks, and some autogenerated regular expressions can be effectively used to solve the classification problems as an interpretable way.…”
Section: Related Workmentioning
confidence: 99%