This article has Open Peer Review reports available.

# Autoregression as a means of assessing the strength of seasonality in a time series

- Rahim Moineddin
^{1, 3}, - Ross EG Upshur
^{1, 2, 3, 4}Email author, - Eric Crighton
^{2}and - Muhammad Mamdani
^{4, 5, 6}

**1**:10

https://doi.org/10.1186/1478-7954-1-10

© Moineddin et al; licensee BioMed Central Ltd. 2003

**Received: **27 August 2003

**Accepted: **15 December 2003

**Published: **15 December 2003

## Abstract

### Background

The study of the seasonal variation of disease is receiving increasing attention from health researchers. Available statistical tests for seasonality typically indicate the presence or absence of statistically significant seasonality but do not provide a meaningful measure of its strength.

### Methods

We propose the coefficient of determination of the autoregressive regression model fitted to the data ( ) as a measure for quantifying the strength of the seasonality. The performance of the proposed statistic is assessed through a simulation study and using two data sets known to demonstrate statistically significant seasonality: atrial fibrillation and asthma hospitalizations in Ontario, Canada.

### Results

The simulation results showed the power of the in adequately quantifying the strength of the seasonality of the simulated observations for all models. In the atrial fibrillation and asthma datasets, while the statistical tests such as Bartlett's Kolmogorov-Smirnov (BKS) and Fisher's Kappa support statistical evidence of seasonality for both, the quantifies the strength of that seasonality. Corroborating the visual evidence that asthma is more conspicuously seasonal than atrial fibrillation, the calculated for atrial fibrillation indicates a weak to moderate seasonality ( = 0.44, 0.28 and 0.45 for both genders, males and females respectively), whereas for asthma, it indicates a strong seasonality ( = 0.82, 0.78 and 0.82 for both genders, male and female respectively).

### Conclusions

For the purposes of health services research, evidence of the statistical presence of seasonality is insufficient to determine the etiologic, clinical and policy relevance of findings. Measurement of the strength of the seasonal effect, as can be determined using the technique, is also important in order to provide a robust sense of seasonality.

## Background

Seasonality is an important component of disease manifestation. The presence of predictable seasonality is a clue to the possible etiology of disease, be it from microbial, environmental or social factors. Understanding seasonality is also essential for setting rational policy, particularly with respect to the planning for seasonal demands for health services.

For studying seasonality, several statistical methods are available ranging from simple graphical techniques to more advanced statistical methods. Additionally, autocorrelation functions can be examined to assess regularity of periodicity or seasonality. Several statistical tests have been introduced for studying the cyclical variation of time series data. For example, Edwards [1] developed a statistical test that locates weights corresponding to the number of observed cases for each month at 12 equally spaced points on a circle. The test is said to be significant if the calculated centre of the mass significantly deviates from the circle's centre. Jones et al [2] developed a test for determining whether incidence data for two or more groups have the same seasonal pattern. Further, Marrero [3] compared the performance of several tests for seasonality by simulation, which can be used as a guideline for selecting appropriate tests for a given data set based on the size of the data set and the shape of the sinusoidal curve. To apply any of these tests, however, observations must be aggregated into 12 monthly data points.

Several alternative tests, which do not require aggregated data, have also been developed. These include Fisher's Kappa (FK), which tests whether the largest periodogram is statistically different from the mean of periodograms; Bartlett's Kolmogorov-Smirnov (BKS) test, which statistically compares the normalized cumulative periodogram with the cumulative distribution function of a uniform zero and one random variable; and the X-11 procedure as used by the census bureaus in the United States and Canada [2, 4–11]. These tests utilize the frequency and time domain to detect seasonality. Each test provides an indication of the presence or absence of statistical significance of seasonality, however, they do not provide a sense of the magnitude of seasonality or how much variance is explained by seasonal occurrence in the data. This is particularly important in health care, as the presence of statistically significant seasonality may not translate into either etiologic or policy relevance.

In an effort to address the shortcomings in existing statistical methods, we propose the application of autoregressive regression models as a means for assessing the degree of accuracy to which a new observation can be predicted by stable (seasonal factors are constant over time) seasonal variation and use it for quantifying the strength of the seasonality within a set of serially correlated observations. In classical regression analysis the coefficient of determination, *R*
^{2}, is a standard statistical tool for estimating the proportion of total variation of the dependent variable, which can be explained by explanatory variables. A crucial point in standard regression is that observations are independent of one another. However, time series observations can be serially autocorrelated and this correlation must be taken into account.

*k*years monthly,

*n*= 12

*k*, trend removed and centred (mean deleted) observations. Let ,

*i*= 1,2,...,12 denote the monthly average. The monthly averages,

*s*, can be interpreted as crude estimates of stable seasonal factors, therefore, the range of parameter estimates is a good estimate of the magnitude of seasonal variation. For estimating one defines 11 dummy variables

*m*

_{ i }= 1 if month equals

*i*,

*m*

_{ i }= 0 otherwise and then regress

*m*

_{ i }s on

*y*

_{ t }

*s*. It is not difficult to show the ordinary least squared estimates of the parameters of the regression equation are and . In practice and parameters β

_{ i }

*s*are estimated simultaneously, therefore, the estimated parameters

*s*can be used as seasonal factor estimates.

The coefficient of determination, which lies between 0 and 1, can be used as a measure for the strength of the stable seasonality because it measures how well the next value can be predicted using month as the only predictor. When is zero, there is no seasonality. When is equal to 1, observations can be perfectly predicted for each month, which means that the variable month explains 100% of the variation in the data. In other words, there is a perfect seasonality. In practice we may characterize the strength of the seasonality based on different ranges of values for . Similar to other measures of correlation or goodness of fit, we can interpret as follows: values ranging from 0 to less than 0.4 may be characterized as non-existent to weak seasonality, 0.4 to less than 0.7 represent moderate to strong seasonality, and values ranging from 0.7 to 1 represent strong to perfect seasonality. The coefficient of determination, , does not quantify the magnitude of the seasonal effect (the difference between peak and trough, which can be estimated by the difference between maximum and minimum parameter estimates of the regression equation) but rather it quantifies its strength (i.e., how well new observations can be predicted when month is the only predictor).

The purpose of this paper is to evaluate the utility of R-squared autoregression in explaining variance in assessing stable seasonality. To this end, we have examined the performance of the R-squared autoregression through a simulation study and using two data sets known to demonstrate statistically significant weak and strong seasonality: monthly hospitalizations for atrial fibrillation and asthma.

## Methods

### Statistical methods

The autoregressive linear regression model for monthly observations is defined as:

*Y*
_{
t
}= *X*
_{
t
}β + ε_{
t
}

ε_{
t
}= -φ_{1}ε_{
t-1}- φ_{2}ε_{
t-2}- … - φ_{
p
}ε_{
t-p
}+ *e*
_{
t
}

where *Y*
_{
t
}is the observed time series, *X*
_{
t
}is the design matrix (a *k* × 12 matrix of 0 and 1), β = (μ, β_{1}, …, β_{11})' is the vector of parameters, ε_{
t
}is the error term that follows an autoregressive model of order *p*. Also, we assume that *e*
_{
t
}is normally and independently distributed with mean zero and variance
[12]. Standard statistical packages (e.g. SAS) can be used for estimating the parameters and the coefficient of determination,
, after adjusting for correlated error terms.

### Simulation

In order to assess the performance of the proposed for measuring the strength of the seasonality of a time series, a simulation study was conducted. Following this, the proposed technique was applied to two real data sets. The SAS software, version 8.2 (SAS Institute Inc. Cary, North Carolina) is used for simulating monthly observations and calculating .

We simulated 1000 replications of monthly observations over 10 years from the following model:

_{1}, φ

_{2}, and φ

_{12}the coefficients, this model generates observations with a cyclical trend component of period 12, observations from a seasonal ARMA model with a seasonal period of 12, and a combination of both. By changing the coefficients of the model we can generate pure white noise to highly correlated data with seasonal patterns. For example, a model with φ

_{1}= φ

_{2}= φ

_{12}= 0 generates a series of observations, which is a mixture of white noise plus a cyclical trend. The size of α controls the contribution of the seasonal trend in the generated observations. Parameters φ

_{1}and φ

_{2}control the correlation structure of the simulated data. When φ

_{1}= 0.9, for example, highly correlated observations are generated. When parameter φ

_{12}is nonzero, the model generates observations with a stochastic seasonal component. Similarly, nonzero φ

_{12}combined with nonzero φ

_{1}or φ

_{2}generates a series with a stochastic seasonal component which is correlated with the non-seasonal components. When all parameters are nonzero the generated observations have a complex structure which depends on a cyclical trend, a stochastic seasonal component, and its correlation structure. In our simulation, we set the order of the error terms in the autoregression model to 2 (for all simulated observations). The mean and standard deviation of 1000 calculated procedures are given in Table 1.

Simulation results

Parameters | ||||||
---|---|---|---|---|---|---|

φ | φ | φ | α = 0 | α = 1 | α = 2 | α = 4 |

0 | 0 | 0 | 0.099 (0.041) | 0.400 (0.070) | 0.703 (0.051) | 0.902 (0.020) |

0.5 | 0 | 0 | 0.098 (0.040) | 0.409 (0.070) | 0.712 (0.049) | 0.906 (0.019) |

-0.9 | 0 | 0 | 0.103 (0.042) | 0.396 (0.069) | 0.698 (0.050) | 0.899 (0.020) |

0.5 | -0.8 | 0 | 0.098 (0.039) | 0.393 (0.065) | 0.699 (0.044) | 0.899 (0.017) |

0 | 0 | 0.5 | 0.256 (0.088) | 0.703 (0.064) | 0.896 (0.026) | 0.971 (0.007) |

0 | 0 | 0.7 | 0.410 (0.114) | 0.847 (0.043) | 0.953 (0.014) | 0.988 (0.004) |

0 | 0 | 0.9 | 0.716 (0.101) | 0.974 (0.009) | 0.993 (0.002) | 0.998 (0.001) |

0.3 | -0.2 | 0.3 | 0.171 (0.066) | 0.602 (0.072) | 0.847 (0.034) | 0.957 (0.011) |

0.3 | -0.2 | 0.5 | 0.260 (0.095) | 0.764 (0.060) | 0.924 (0.021) | 0.980 (0.006) |

0.3 | -0.2 | 0.7 | 0.470 (0.149) | 0.934 (0.024) | 0.982 (0.006) | 0.995 (0.002) |

When α and φ_{12} are zero,
indicates no significant seasonality even for highly correlated data (e.g. φ_{1} = -0.9). The calculated
increases as α increases regardless of the sign and magnitude of the other parameters. When α is zero (there is no cyclical trend), the
increases in all cases as φ_{12} increases. This demonstrates the ability and usefulness of
in quantifying the strength of the seasonality in time series data.

To investigate the linkages between the
and *p*, the order of the autoregression model, we repeated the simulation experiments with *p* = 1,2,4,8, and 13. Also an additional simulation experiment using the stepwise autoregression method was conducted to select the order of the autoregressive error model whereby the maximum possible autoregressive order was set equal to 13. The stepwise autoregression method involves fitting a high order model and then sequentially removing parameters until all remaining autoregressive parameters remain statistically significant [13]. For the fixed order, *p*, simulation results (not shown here) showed that
is robust to over-fitting the data and *p* does not significantly affect the estimated
. Results from the stepwise experiments showed that relative to fixed orders, *p*, there were no significant differences in the
. Results were, however, slightly more conservative using the stepwise method.

In order to apply the autoregression procedure to actual data it is important to eliminate the nonstationarity in the mean and variance of the observations. For the data used in the study the Dickey-Fuller unit root test [14] is used to test the stationarity of the series and determine the order of differencing required for the nonstationary series. The SAS procedure, AUTOREG, was used for calculating
. In all cases *p*, the order of the autoregressive model for error terms was selected using the stepwise autoregression method. The maximum possible order was set equal to 13.

### Data Sources

The data were derived from two retrospective, population-based, cross-sectional time series studies assessing temporal patterns in all discharge separations for asthma (from April 1, 1988 to March 31, 2000) and atrial fibrillation (from April 1, 1988 to March 31, 2001) for the population of Ontario. Approximately 14 million residents of Ontario, Canada eligible for universal health care coverage during this time were included for analysis. The database used was the Canadian Institute for Health Information (CIHI) Discharge Abstract Database which records discharges from all Ontario acute care hospitals. All records with a most responsible discharge diagnosis of atrial fibrillation (ICD-9 code: 427.3) and asthma (ICD-9 code: 493) were selected. The numerator consisted of the total number of discharge separations for each month. Denominators were constructed from annual census data provided by Statistics Canada for each age group for residents of Ontario. Monthly population estimates were created through linear interpolation.

## Results

### Atrial fibrillation

### Asthma

## Discussion

The results of this study show that the
can be useful in distinguishing weak from strong stable seasonal effects in both simulation and in actual data sets. While the statistical tests such as BKS and Fisher's Kappa support statistical evidence of seasonality in the data, the
allows quantification of the strength of that stable seasonality, as demonstrated by the simulation results. Regardless of the values of the parameters φ_{1} and φ_{2}, when the parameters α and φ_{12} were zero, the
was small, and when one or both of those parameters increased,
increased proportionally. This is important because any proposed statistics for measuring the strength of the seasonality must be invariant of the correlation structure. The simulation results showed the power of the
in adequately quantifying the strength of the seasonality of the simulated observations for all models. The magnitude of
shows how well the next value can be predicted by using month as the only predictor. In other words it shows the contribution of seasonality in the total variation of the data.

When the technique was applied to the two data sets, it corroborated the visual evidence that asthma is more conspicuously seasonal than atrial fibrillation. The seasonality of asthma has been conclusively demonstrated in several studies and is likely a key to understanding the etiology of exacerbations of asthma [15–17]. The strength and consistency of the effect is likely of relevance to health policy and planning. The seasonality of atrial fibrillation has been reported, but outcomes were reported as relative risks [18]. The analysis provided here indicates that the seasonality of admissions for atrial fibrillation is not likely of policy or clinical significance as the magnitude is quite small.

One important question that remains to be answered is how the magnitude of seasonal factor changes over time affect the . For non-stable seasonal variation, a proper transformation such as log may be required to transfer a non-stable seasonal variation to a stable one. The value of may change if the sampling period changes (e.g. monthly data converted to weekly data). By including year and month as predictors we can adjust for moving seasonality, however, further research is required.

## Conclusions

The proposed autoregression method is a statistical technique well suited to the study of seasonality in health data. Although monthly data was used for this analysis, it can easily be applied to weekly or seasonal data. The approach allows researchers to quantify and compare the strength of the seasonality for different genders and age groups. The coefficient of determination is easy to calculate and interpret. And finally, it is well known to health care researchers and is frequently used as a measure for goodness of fit. For the purposes of health services research and population health measurement, evidence of the statistical presence of stable seasonality is insufficient to determine the etiologic, clinical and policy relevance of findings. Measurement of the strength of the seasonal effect is also required in order to provide a robust sense of seasonality. We believe that this autoregression technique, in concert with statistical testing, graphical representation and measures of the absolute magnitude of seasonal effect, is an important component to this robust approach.

## Authors Contributions

RM and RU initiated the idea for the study. MM and EC contributed critical intellectual input to the project. RM wrote the initial draft, and all authors contributed to the subsequent drafts. All have read and approve of the final draft.

## Declarations

### Acknowledgements

The authors would like to thank the editor and reviewers for their very helpful suggestions that significantly improved the paper. Also we would like to thank Shari Gruman for her expert assistance in formatting the manuscript. RU is supported by a New Investigator Award from the Canadian Institutes of Health Research and a Research Scholar Award from the Department of Family and Community Medicine, University of Toronto. This research was supported by an operating grant from the Canadian Institutes of Health Research.

## Authors’ Affiliations

## References

- Edwards J:
**The recognition and estimation of cyclic trends.***Ann Hum Genet*1961,**25:**83-86.View ArticlePubMedGoogle Scholar - Jones RH, Ford PM, Hamman RF:
**Seasonality comparisons among groups using incidence data.***Biometrics*1988,**44:**1131-1144.View ArticlePubMedGoogle Scholar - Marrero O:
**The performance of several statistical tests for seasonality in monthly data.***J Statist Comput Simulation*1983,**17:**275-296.View ArticleGoogle Scholar - Fuller W:
*Introduction to Statistical Time Series*New York: John Wiley 1976.Google Scholar - Franses P:
**Testing for seasonality.***Econ Lett*1992,**38:**259-262. 10.1016/0165-1765(92)90067-9View ArticleGoogle Scholar - Marrero O:
**L'analyse de la variation saisonniere quand l'amplitude et la taille sont faibles.***Canad J Statist*1999,**27:**857-882.View ArticleGoogle Scholar - Marrero O:
**Statistical testing for seasonality in data with multiple peaks and troughs.***Biom J*1984,**26:**591-608.View ArticleGoogle Scholar - Marrero O:
**The power of a nonparametric test for seasonality.***Biom J*1988,**30:**495-502.View ArticleGoogle Scholar - Normolle D, Brown M:
**Identification of aperiodic seasonality in non-Gaussian time series.***Biometrics*1994,**50:**798-812.View ArticlePubMedGoogle Scholar - Fellman J, Eriksson AW:
**Statistical analysis of the seasonal variation in demographic data.***Hum Biol*2000,**72:**851-876.PubMedGoogle Scholar - Harvey H:
**Testing in unobserved components models.***J Forecasting*2001,**20:**1-19. http://dx.doi.org/10.1002/1099-131X(200101)20:1%3C1::AID-FOR764%3E3.0.CO;2-3 View ArticleGoogle Scholar - Hamilton J:
*Time Series Analysis*Princeton, NJ: Princeton University Press 1994.,**Chapter 4:**Google Scholar - SAS/TES User's Guide Cary, NC: SAS Institute Inc. 1999.Google Scholar
- Dickey D, Hasza D, Fuller W:
**Testing for unit roots in seasonal time series.***J Am Stat Assoc*1984,**79:**355-367.View ArticleGoogle Scholar - Crighton EJ, Mamdani MM, Upshur RE:
**A population based time series analysis of asthma hospitalisations in Ontario, Canada: 1988 to 2000.***BMC Health Serv Res*2001,**1:**7. 10.1186/1472-6963-1-7View ArticlePubMedPubMed CentralGoogle Scholar - Kimbell-Dunn M, Pearce N, Beasley R:
**Seasonal variation in asthma hospitalizations and death rates in New Zealand.***Respirology*2000,**5:**241-246. 10.1046/j.1440-1843.2000.00255.xView ArticlePubMedGoogle Scholar - Harju T, Keistinen T, Tuuponen T, Kivela SL:
**Seasonal variation in childhood asthma hospitalisations in Finland, 1972–1992.***Eur J Pediatr*1997,**156:**436-439. 10.1007/s004310050632View ArticlePubMedGoogle Scholar - Frost L, Johnsen SP, Pedersen L, Husted S, Engholm G, Sorensen HT, Rothman KJ:
**Seasonal variation in hospital discharge diagnosis of atrial fibrillation: a population-based study.***Epidemiology*2002,**13:**211-215. 10.1097/00001648-200203000-00017View ArticlePubMedGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article: verbatim copying and redistribution of this article are permitted in all media for any purpose, provided this notice is preserved along with the article's original URL.