2014
DOI: 10.1016/j.artmed.2014.01.004
|View full text |Cite
|
Sign up to set email alerts
|

Adaptation of machine translation for multilingual information retrieval in the medical domain

Abstract: Objective. We investigate machine translation (MT) of user search queries in the context of cross-lingual information retrieval (IR) in the medical domain. The main focus is on techniques to adapt MT to increase translation quality; however, we also explore MT adaptation to improve effectiveness of cross-lingual IR.Methods and Data. Our MT system is Moses, a state-of-the-art phrase-based statistical machine translation system. The IR system is based on the BM25 retrieval model implemented in the Lucene searc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
46
0
1

Year Published

2014
2014
2019
2019

Publication Types

Select...
4
3
2

Relationship

2
7

Authors

Journals

citations
Cited by 26 publications
(48 citation statements)
references
References 76 publications
1
46
0
1
Order By: Relevance
“…We used the CLEF 2013 eHealth Task 3 test collection containing about 1 million web pages (in English), 50 test queries (originally in English and translated to Czech, German, and French), and their relevance assessments. Some of the participants of the WMT Medical Task (three teams with five submissions in total) submitted translations of the queries (from Czech, German, and French) into English and these translations were used to query the CLEF 2013 eHealth Task 3 test collection using a state-of-the-art system based on a BM25 model, described in Pecina et al (2014). Originally, we asked for 10 best translations for each query, but only the best one were used for the evaluation.…”
Section: Resultsmentioning
confidence: 99%
“…We used the CLEF 2013 eHealth Task 3 test collection containing about 1 million web pages (in English), 50 test queries (originally in English and translated to Czech, German, and French), and their relevance assessments. Some of the participants of the WMT Medical Task (three teams with five submissions in total) submitted translations of the queries (from Czech, German, and French) into English and these translations were used to query the CLEF 2013 eHealth Task 3 test collection using a state-of-the-art system based on a BM25 model, described in Pecina et al (2014). Originally, we asked for 10 best translations for each query, but only the best one were used for the evaluation.…”
Section: Resultsmentioning
confidence: 99%
“…proposed methods would in fact significantly improve efficiency over a baseline approach, this increase in effectiveness might not be noted due to missing relevance information in the benchmark data. In particular, it was argued before that experiments with blind relevance feedback or query expansion did not show a significant improvement in performance due to incomplete or missing relevance information [45]. Our analysis shows that for all three strategies investigated, this is clearly not the case (cf.…”
Section: Documents and Relevance Assessmentmentioning
confidence: 52%
“…In the same vein, data translation also arises as a complex challenge for researchers. In addition to the obvious sense (translating data between multiple languages [ 46 ]), there is the data translation from a low-level free text data to structured information [ 47 , 48 ]. Clinicians' reports traditionally include their notes in free text.…”
Section: Resultsmentioning
confidence: 99%