2020
DOI: 10.1162/coli_a_00377
|View full text |Cite
|
Sign up to set email alerts
|

A Systematic Study of Inner-Attention-Based Sentence Representations in Multilingual Neural Machine Translation

Abstract: Neural machine translation has considerably improved the quality of automatic translations by learning good representations of input sentences. In this article, we explore a multilingual translation model capable of producing fixed-size sentence representations by incorporating an intermediate crosslingual shared layer, which we refer to as attention bridge. This layer exploits the semantics from each language and develops into a language-agnostic meaning representation that can be efficiently used fo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 13 publications
(11 citation statements)
references
References 52 publications
0
11
0
Order By: Relevance
“…HY (Vázquez et al, 2020) participated with 14 http://sox.sourceforge.net/ both cascade and end-to-end systems. For the end-to-end system, they used a multimodal approach (with audio and text as the two modalities treated as different languages) trained in a multitask fashion, which maps the internal representations of different encoders into a shared space before decoding.…”
Section: Submissionsmentioning
confidence: 99%
See 3 more Smart Citations
“…HY (Vázquez et al, 2020) participated with 14 http://sox.sourceforge.net/ both cascade and end-to-end systems. For the end-to-end system, they used a multimodal approach (with audio and text as the two modalities treated as different languages) trained in a multitask fashion, which maps the internal representations of different encoders into a shared space before decoding.…”
Section: Submissionsmentioning
confidence: 99%
“…For the end-to-end system, they used a multimodal approach (with audio and text as the two modalities treated as different languages) trained in a multitask fashion, which maps the internal representations of different encoders into a shared space before decoding. To this aim, they incorporated the inner-attention based architecture proposed by (Vázquez et al, 2020) within Transformer-based encoders (inspired by (Tu et al, 2019;Di Gangi et al, 2019c)) and decoders. For the cascade approach, they used a pipeline of three stages: (1) ASR (trained with S-Transformer (Di Gangi et al, 2019c)), (2) re-punctuation and letter case restoration (based on Marian's implementation of Transformer), and ( 3) MT (also based on Marian).…”
Section: Submissionsmentioning
confidence: 99%
See 2 more Smart Citations
“…We propose to explore the impact of context variation on word representations. We specifically address representations generated by the BERT model (Devlin et al, 2019), trained using a language modeling objective, and translation models involving one or more language pairs (Artetxe and Schwenk, 2019;Vázquez et al, 2020).…”
Section: Introductionmentioning
confidence: 99%