ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8683367
|View full text |Cite
|
Sign up to set email alerts
|

Investigating the Effects of Word Substitution Errors on Sentence Embeddings

Abstract: A key initial step in several natural language processing (NLP) tasks involves embedding phrases of text to vectors of real numbers that preserve semantic meaning. To that end, several methods have been recently proposed with impressive results on semantic similarity tasks. However, all of these approaches assume that perfect transcripts are available when generating the embeddings. While this is a reasonable assumption for analysis of written text, it is limiting for analysis of transcribed text. In this pape… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 18 publications
0
7
0
Order By: Relevance
“…Performing a pairwise comparison on the slopes of the similarity values, between our sensitivity analysis for Hungarian and a similar one for English [29], our results are quite satisfactory. The BOW (regardless of the usage of the stop words), SIF and uSIF are more robust (around rel.…”
Section: Results and Discussion 41 Sts-related Resultsmentioning
confidence: 65%
“…Performing a pairwise comparison on the slopes of the similarity values, between our sensitivity analysis for Hungarian and a similar one for English [29], our results are quite satisfactory. The BOW (regardless of the usage of the stop words), SIF and uSIF are more robust (around rel.…”
Section: Results and Discussion 41 Sts-related Resultsmentioning
confidence: 65%
“…When using the AMI ASR (A1) transcripts instead of the manual transcripts, the decrease in ROUGE is around 2-3%. This relatively small drop is likely because even at 30% WER, the sentence/utterance embedding similarity between a manual source and an ASR source is about 0.70-0.85% [33,34]. This system achieves higher all ROUGE measures than extractive method CoreRank [6], and when compared to abstractive method Top-icSeg (without visual signals) [12] our system achieves higher ROUGE-2 and ROUGE-L although lower ROUGE-1.…”
Section: Modelmentioning
confidence: 80%
“…The advantages of using the R-TLM structure are as follows. First, during the rescoring stage at test-time, there usually exists word errors in the transcriptions of past utterances, and the LSTM layer in the R-TLM structure provides LSTM hidden state representations which are more robust against such errors [31]. Second, the R-TLM provides complementary history representations from both the LSTM and Transformer-XL to the attention-based representation, and increases the network capability.…”
Section: R-tlm Structurementioning
confidence: 99%