“…Czech Electra model (Kocián et al, 2021), two multilingual models mBERT (Devlin et al, 2019), XLM-R (Conneau et al, 2020) and the original monolingual English BERT model (Devlin et al, 2019), see We fine-tune all the models for the binary classification task, i.e., subjective vs. objective sentence detection. For all models based on the original BERT model, we use the hidden vector h ∈ R H of the classification token [CLS] that represents the entire input sequence, where H is the hidden size of the model.…”