Stationarity of Time Series: Tests

In summary, the conversation discusses the use of statistical tests, such as coefficient of error and R-squared/Adjusted R-squared, in determining the stationarity of mean and standard deviation/variance in time series models. The possibility of using a modified version of the coefficient of variation, known as the coefficient of error, is also raised. Questions are posed regarding the most stationary SMA in a time series and the use of regression models in evaluating riskiness. Ultimately, it is noted that a structural model, such as the CAPM, provides a more concrete statistical basis for these measures.
  • #1
kimberley
14
0
This post may seem a bit meandering, but it does well to fully communicate my thoughts and ultimate questions.

Very little of the literature on Time Series Models makes reference to what is sparsely referred to elsewhere as the "Coefficient of Error", or even R-squared/Adjusted R-squared, in testing n-periods in a data series for stationarity of Mean and/or Standard Deviation/Variance. Viscerally, it seems to me that these two statistical tests would be the most parsimonious in accomplishing this end, but I suspect that there are various arguments that militate against this conclusion since these tests seem to be rarely, if ever, used. I'd be interested in your views.

Along these lines, consider a basic, Non-Linear, Simple Moving Average ("SMA") Time Series where your X axis is simply consecutive numbers representing Time (i.e., 1=Day 1, 2=Day 2, 3=Day 3 etc.), and your Y axis are your residuals for those days. The three primary goals in constructing an SMA Model are stationarity of Mean, stationarity of Standard Deviation/Variance, and Normality.

Now, follow me through this tangential reasoning in framing my ultimate questions. We know that the "Coefficient of Variation" is used to measure the relative dispersion of data sets by taking the standard deviation of each data set and dividing each by its Mean. The simplest and most common example given for the use of the "Coefficient of Variation" seems to be with regard to stock prices and risk. That is, if you have "Stock A" with a Mean price of $50 over the last 30 days, and a Standard Deviation of $5 over the last 30 days, its Coefficient of Variation is .1 over the last 30 days, and if you have "Stock B" with a Mean Price of $50 and Standard Deviation of $2 over the last 30 days, its Coefficient of Variation would be .04 over the last 30 days. We conclude, therefore, that the relative dispersion of Stock B is less than Stock A over the last 30 days. In other words, Stock B has been less risky over the last 30 days because the ratio of its standard deviation to Mean is smaller than that of Stock A. NOTE: COEFFICIENT OF VARIATION IS TYPICALLY USED IN THE LITERATURE WHERE YOU ARE COMPARING DISPERSION BETWEEN TWO DATA SETS (i.e., stocks in the above case) OVER TIME PERIODS OF EQUAL LENGTH.

With the above example in mind, couldn't we use a modified version of the Coefficient of Variation, the so-called "Coefficient of Error", to determine the relative stationarity of nested Moving Averages, of unequal length, in the same overall Time Series? (i.e., comparing the stationarity of the 89 day Simple Moving Average to the stationarity of the 58 day Simple Moving Average, as of today). Toward this end, as I understand it, the Coefficient of Error is really just a combination of the Standard Error of the Mean and Coefficient of Variation (i.e., the same formula used for calculating a confidence interval). To wit, the Coefficient of Error of a data set is calculated by multiplying the standard deviation of the data set by the square root of 1/n and then dividing that product by the Mean. Again, in other words, it's really just the same formula for calculating a confidence interval, but it uses the standard deviation of the sample rather than the whole population.

QUESTIONS:

1. At the end of each day, when analyzing a Simple Moving Average ("SMA") Time Series, wouldn't the most stationary SMA be the n-period with the narrowest confidence interval (i.e. smallest Coefficient of Error)?

2. Alternatively, in a SMA Time Series, isn't the n-period with the lowest R-squared/Adjusted R-squared statistic the most stationary (no linear trend)? For instance, if over the last 98 days the R-squared statistic is 0, and is the lowest R-squared reading of any n-period, wouldn't that be an indication of stationarity as well?
 
Physics news on Phys.org
  • #2
With respect to Q.2, what is the regression model you have in mind? (See below.)

With respect to Q.1: on the face of it, CoE appears as a neat way to introduce the sample size into the formula, thus enabling the researcher to compare models with different sample sizes (which the CoV does not). However, even a visual comparison of two or more CoEs is not a statistical test and cannot be used for "statistical inference" in a technical sense.

Measures of risk (e.g. financial) usually assume some kind of structural model, even a naive one. The CAPM regression model used in financial theory is a simple and commonly accepted model of evaluating riskiness of stocks (relative to the market). Using a simple CAPM regression model, it is possible to test the relative riskiness of two stocks in a way that a visual comparison of two or more coefficients does not allow.

Also, a structural model such as the CAPM provides a concrete statistical basis to define terms such as "coefficient of determination" (commonly referred to as the R-squared statistic).
 
Last edited:
  • #3



Thank you for sharing your thoughts on stationarity of time series and the use of statistical tests. I agree that the coefficient of error and R-squared/Adjusted R-squared can be useful in testing for stationarity, but there are other factors that also need to be considered.

Firstly, it is important to note that the coefficient of error and R-squared/Adjusted R-squared are not commonly used in testing for stationarity because they are not specifically designed for this purpose. These tests are more commonly used for assessing the fit of a model and the strength of linear relationships, respectively. Therefore, while they may provide some indication of stationarity, they are not the most reliable tests for this purpose.

Additionally, stationarity of mean and variance can be assessed more accurately using specialized tests such as the Augmented Dickey-Fuller (ADF) test or the Kwiatkowski-Phillips-Schmidt-Shin (KPSS) test. These tests take into account the autocorrelation and trend in the data, which can affect the results of simpler tests like the ones you mentioned.

Furthermore, the coefficient of error and R-squared/Adjusted R-squared may not be suitable for comparing the stationarity of different moving averages of unequal length. This is because they do not account for the fact that the moving averages have different sample sizes and may have different underlying trends or patterns. In this case, it would be more appropriate to use a specialized test that can compare the stationarity of different moving averages.

To address your questions, I would say that the most stationary SMA at the end of each day would not necessarily be the one with the narrowest confidence interval or the lowest R-squared/Adjusted R-squared statistic. This is because these measures do not take into account the other factors that can affect stationarity. It is important to use specialized tests that are designed for this purpose and to consider all relevant factors in assessing stationarity.

In conclusion, while the coefficient of error and R-squared/Adjusted R-squared can provide some indication of stationarity, they are not the most reliable tests for this purpose. It is important to use specialized tests and consider all relevant factors when assessing stationarity in time series data.
 

1. What is stationarity of time series?

Stationarity of time series refers to the property of a time series data where the statistical properties such as mean, variance, and autocorrelation remain constant over time. In simpler terms, it means that the data is not affected by factors such as trend, seasonality, and other external influences.

2. Why is it important to test for stationarity of time series?

It is important to test for stationarity of time series because most time series forecasting models assume that the data is stationary. If the data is not stationary, the forecasting models may not provide accurate predictions. Additionally, stationarity allows for the use of various statistical tools and techniques to analyze and forecast the data.

3. What are the common tests for stationarity of time series?

The common tests for stationarity of time series include the Augmented Dickey-Fuller (ADF) test, Kwiatkowski-Phillips-Schmidt-Shin (KPSS) test, and Phillips-Perron (PP) test. These tests analyze the data for the presence of a unit root, which indicates non-stationarity.

4. How do you interpret the results of a stationarity test?

The results of a stationarity test typically include a test statistic and a p-value. If the test statistic is lower than the critical value and the p-value is less than 0.05, we can reject the null hypothesis and conclude that the data is stationary. On the other hand, if the test statistic is higher than the critical value and the p-value is greater than 0.05, we fail to reject the null hypothesis and conclude that the data is non-stationary.

5. Can non-stationary data be made stationary?

Yes, non-stationary data can be made stationary through a process called differencing. Differencing involves taking the difference between consecutive data points. This can be done multiple times until the data becomes stationary. However, it is important to note that the interpretation of the data may change after differencing, and it is recommended to use stationary data for modeling and forecasting.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
10
Views
948
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
921
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
713
  • Set Theory, Logic, Probability, Statistics
Replies
28
Views
3K
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
471
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
793
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
9
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
8
Views
3K
Back
Top