Abstract
We propose a methodology of estimating the scaling exponent for a long-range correlation in a nonstationary time series from the perspective of the regression analysis. By an adaptive degree determination of a regression polynomial, the proposed methodology is designed to properly remove various types of trends embedded in the nonstationary signal so that the scaling exponent can be estimated without artificial crossovers. To show the validity of the proposed methodology, we applied it to the detrended fluctuation analysis and tested it out against correlated data superimposed by various types of trends. It turned out that, unlike the conventional technique, our approach was capable of eliminating artificial crossovers. We also discuss the statistical characteristics of the proposed method with regard to the estimation of the scaling exponent.
- Received 27 March 2012
DOI:https://doi.org/10.1103/PhysRevE.86.011135
©2012 American Physical Society