Taylor's law describes the fluctuation characteristics underlying a complex system in which the variance of an event within a time span grows by a power law with respect to the mean. The previous paper, Taylor's Law for Linguistic Sequences and Random Walk Models (Tanaka-Ishii and Kobayashi 2018), appeared in Journal of Physics Communications and described a new way to apply Taylor analysis to texts. The method was applied to over 1100 texts across 14 languages. The results showed how the Taylor exponents of natural-language written texts were consistently around 0.58, thus being universal.Experimentally, the Taylor exponent α is known to take a value within the range of 0.5α1.0 across a wide variety of domains, including finance, meteorology, agriculture, and biology. The previous paper shows how this is the case for language.The Taylor exponent is analytically proven to be 0.5 for an independent and identically distributed (i.i.d.) process. The paper also shows a case when 1.0 is reached. This Addendum provides two additional cases of rare word alignment for α=0.5 and α=1.0. These cases provide an understanding to interpret the value of the exponent of a real text.Consider dividing a text of length N into Q segments of length Δt, i.e., N=QΔt. Suppose that Q is sufficiently large.First of all, if a word only appears once in the entire text, then μ 1 and σ 1 are calculated as follows.
AbstractTaylor's law describes the fluctuation characteristics underlying a complex system in which the variance of an event within a time span grows by a power law with respect to the mean. Although Taylor's law has been applied in many natural and social systems, its application for language has been scarce. This article describes a new, natural way to apply Taylor analysis to texts. The method was applied to over 1100 texts across 14 languages and showed how the Taylor exponents of natural language written texts were consistently around 0.58, thus being universal. The exponents were also evaluated for other languagerelated data, such as speech corpora (0.63 for adult speech, 0.68 for child-directed speech), programming language sources (0.79), and music (0.79). The results show how the Taylor exponent serves to quantify the fundamental structural complexity underlying linguistic time series. To explain the nature of natural language sequences possessing such different degrees of fluctuation, we investigated various mathematical models that could produce a Taylor exponent similar to that of real data. While the majority of previous probabilistic sequential models could not produce a Taylor exponent larger than 0.50, the same as in the independent and identically distributed (i.i.d.) case, random walk sequences on complex networks could produce fluctuation. We show that among various possibilities, random walks on a Barabási-Albert (BA) graph with small mean degree could fulfill the scaling properties of Zipf's law and the long-range correlation, in addition to having a Taylor's law exponent larger than 0.5, thus giving a...