2020
DOI: 10.48550/arxiv.2012.02558
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pre-trained language models as knowledge bases for Automotive Complaint Analysis

Abstract: Recently it has been shown that large pre-trained language models like BERT (Devlin et al., 2018) are able to store commonsense factual knowledge captured in its pre-training corpus (Petroni et al., 2019). In our work we further evaluate this ability with respect to an application from industry creating a set of probes specifically designed to reveal technical quality issues captured as described incidents out of unstructured customer feedback in the automotive industry. After probing the out-ofthe-box version… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
references
References 17 publications
0
0
0
Order By: Relevance