model decisions by specifying the parts of the input which are most salient in the model's decision process [6,18,48]. In natural language processing (NLP), this refers to which words, phrases or sentences in the input contributed most to the model prediction [10,36]. While much research exists on developing and verifying such explanations [1,4,32,35,50,51],less is known about the information that human explainees actually understand from them [2,12,19,39].In the explainable NLP literature, it is generally (implicitly) assumed that the explainee interprets the information "correctly", as it is communicated [4,17,20]: e.g., when one word is explained to be influential in the model's decision process, or more influential than another word, it is assumed that the explainee understands this relationship [28].We question this assumption: research in the social sciences describes modes in which the human explainee may be biased-via some cognitive habit-in their interpretation of processes [15,37,39,52]. Additional research shows this effect manifests in practice in AI settings [11,14,22,25,40]. This means, for example, that the explainee may underestimate the influence of a punctuation token, even if the explanation reports that this token is highly significant (Figure 1), because the explainee is attempting to understand how the model reasons by analogy to the explainee's own mind which is an instance of anthropomorphic bias [8,29,61] and belief bias [16,22].We identify three different such biases which may influence the explainee's interpretation: (i) anthropomorphic bias and belief bias: influence by the explainee's self projection onto the model; (ii) visual perception bias: influence by the explainee's visual affordances for comprehending information; (iii) learning effects: observable temporal changes in the explainee's interpretation as a result of interacting with the explanation over multiple instances.We thus address the following question in this paper: When a human explainee observes feature-attribution explanations, does their comprehended information differ from what the explanation "objectively" attempts to communicate? If so, how?We propose a methodology to investigate whether explainees exhibit biases when interpreting feature-attribution explanations in NLP, which effectively distort the objective attribution into a subjective interpretation of it (Section 4).We conduct user studies in which we show an input sentence and a feature-attribution explanation (i.e., saliency map) to explainees, ask them to report their subjective interpretation, and analyze their responses for statistical significance across multiple factors, such as word length, total input length, or dependency relation, using GAMMs (Section 5).We find that word length, sentence length, the position of the sentence in the temporal course of the experiment, the saliency rank, capitalization, dependency relation, word position, word frequency as well as sentiment can significantly affect user perception. In addition to whether a factor has...