Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 1 2017
DOI: 10.18653/v1/e17-1098
|View full text |Cite
|
Sign up to set email alerts
|

Decoding with Finite-State Transducers on GPUs

Abstract: Weighted finite automata and transducers (including hidden Markov models and conditional random fields) are widely used in natural language processing (NLP) to perform tasks such as morphological analysis, part-of-speech tagging, chunking, named entity recognition, speech recognition, and others. Parallelizing finite state algorithms on graphics processing units (GPUs) would benefit many areas of NLP. Although researchers have implemented GPU versions of basic graph algorithms, limited previous work, to our kn… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 21 publications
0
7
0
Order By: Relevance
“…If tokens reach the same state in the same frame, only the token with the smallest weight remains. This mechanism is called token recombina- [17,20,21], the atomicMin 1 operation is applied to an array whose size is the number of states or arcs for the token recombination. Using this procedure, the array consumes a large amount of GPU memory when a WFST has a large number of states and arcs.…”
Section: Parallel Viterbi Search On Gpumentioning
confidence: 99%
“…If tokens reach the same state in the same frame, only the token with the smallest weight remains. This mechanism is called token recombina- [17,20,21], the atomicMin 1 operation is applied to an array whose size is the number of states or arcs for the token recombination. Using this procedure, the array consumes a large amount of GPU memory when a WFST has a large number of states and arcs.…”
Section: Parallel Viterbi Search On Gpumentioning
confidence: 99%
“…Our GPU implementation stores FST transition functions in a format similar to compressed sparse row (CSR) format, as introduced by our previous work Argueta and Chiang (2017). For the composition task we use a slightly different representation.…”
Section: Transducer Representationmentioning
confidence: 99%
“…In our previous work (Argueta and Chiang, 2017), we created transducers for a toy translation task. We trained a bigram language model (as in Figure 3a) and a one-state translation model (as in Figure 3 Figure 3: The transducers used for testing were obtained by pre-composing: (a) a language model and (b) a translation model.…”
Section: Setupmentioning
confidence: 99%
See 1 more Smart Citation
“…Moreover, it also requires less data transfer compared with [22]. Finally, to the best of our understanding, [24] is the only open-source project in this field, but it only implemented the basic Viterbi decoding without combining AM posteriors and beam [9], which cannot be applied to ASR.…”
Section: Introductionmentioning
confidence: 99%