Given two time series X and Y, their mutual information, I͑X , Y͒ = I͑Y , X͒, is the average number of bits of X that can be predicted by measuring Y and vice versa. In the analysis of observational data, calculation of mutual information occurs in three contexts: identification of nonlinear correlation, determination of an optimal sampling interval, particularly when embedding data, and in the investigation of causal relationships with directed mutual information. In this contribution a minimum description length argument is used to determine the optimal number of elements to use when characterizing the distributions of X and Y. However, even when using partitions of the X and Y axis indicated by minimum description length, mutual information calculations performed with a uniform partition of the XY plane can give misleading results. This motivated the construction of an algorithm for calculating mutual information that uses an adaptive partition. This algorithm also incorporates an explicit test of the statistical independence of X and Y in a calculation that returns an assessment of the corresponding null hypothesis. The previously published Fraser-Swinney algorithm for calculating mutual information includes a sophisticated procedure for local adaptive control of the partitioning process. When the Fraser and Swinney algorithm and the algorithm constructed here are compared, they give very similar numerical results ͑less than 4% difference in a typical application͒. Detailed comparisons are possible when X and Y are correlated jointly Gaussian distributed because an analytic expression for I͑X , Y͒ can be derived for that case. Based on these tests, three conclusions can be drawn. First, the algorithm constructed here has an advantage over the Fraser-Swinney algorithm in providing an explicit calculation of the probability of the null hypothesis that X and Y are independent. Second, the Fraser-Swinney algorithm is marginally the more accurate of the two algorithms when large data sets are used. With smaller data sets, however, the Fraser-Swinney algorithm reports structures that disappear when more data are available. Third, the algorithm constructed here requires about 0.5% of the computation time required by the Fraser-Swinney algorithm.
Embedding experimental data is a common first step in many forms of dynamical analysis. The choice of appropriate embedding parameters ͑dimension and lag͒ is crucial to the success of the subsequent analysis. We argue here that the optimal embedding of a time series cannot be determined by criteria based solely on the time series itself. Therefore we base our analysis on an examination of systems that have explicit analytic representations. A comparison of analytically obtained results with those obtained by an examination of the corresponding time series provides a means of assessing the comparative success of different embedding criteria. The assessment also includes measures of robustness to noise. The limitations of this study are explicitly delineated. While bearing these limitations in mind, we conclude that for the examples considered here, the best identification of the embedding dimension was achieved with a global false nearest neighbors argument, and the best value of lag was identified by the mutual information function.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.