This paper develops a new test of true versus spurious long memory, based on log-periodogram estimation of the long memory parameter using skip-sampled data. A correction factor is derived to overcome the bias in this estimator due to aliasing. The procedure is designed to be used in the context of a conventional test of significance of the long memory parameter, and a composite test procedure is described that has the properties of known asymptotic size and consistency. The test is implemented using the bootstrap, with the distribution under the null hypothesis being approximated using a dependent-sample bootstrap technique to approximate short-run dependence following fractional differencing. The properties of the test are investigated in a set of Monte Carlo experiments. The procedure is illustrated by applications to exchange rate volatility and dividend growth series.
Estimation of the long memory parameter by the method due to Geweke and Porter-Hudak (1983), or one of its variants, is a popular methodology in time series analysis. This estimator (henceforth, GPH) exploits fact that the autocovariances of a long memory process are nonsummable, and the spectral density accordingly diverges at the origin at a particular rate, with as . GPH estimates by regressing the logarithms of the periodogram points in the neighbourhood of zero onto a suitable trend. However, except in very large samples this method has well-known limitations. As documented by Agiakloglou, Newbold, and Wohar (1993), the neglect of components of representing short-run autocorrelation imply omitted terms in the regression, resulting in potentially substantial bias. In particular, the method is problematic as a basis for testing the null hypothesis of short memory, the case , since the conventional Wald statistic can be severely over-sized.
A simple illustration of this difficulty is provided by the observational equivalence between the fractionally integrated process with and the autoregressive process with . The ARFI model
can exhibit a characteristically bimodal likelihood function when either of the parameters and is close to unity in the process generating the sample. For every finite sample size, there exists a close enough to unity to bias the GPH estimator of significantly, when its true value is zero. It is desirable to have a means of distinguishing the cases of true and spurious , and goodness-of-fit criteria are an unreliable guide.
The approach explored in this paper is to devise a test with null and alternative interchanged. Recent research has highlighted the well-known property of self-similarity of hyperbolic decay processes under transformations such as periodic aggregation and periodic sub-sampling, otherwise known as skip-sampling. Chambers (1998) was the first to point out that if a long memory process is recorded at different rates, the rate of decay of the autocovariances is invariant to the rate of observation.
There are two ways to conceive of lowering the observation rate. Temporal aggregation means taking the sums of successive observations to create the new sequence. This is the natural transformation for flow data, such that (for example) quarterly flows are each the sum of three successive monthly flows. Ohanissian, Russell, and Tsay (2008) implement a test of long memory based on comparing log-periodogram estimates under different rates of temporal aggregation.
Skip-sampling, by contrast, means taking every th observation and discarding the remainder. This is the natural way of lowering the observation rate for stock or price data, although for the present purpose the nature of the observations is irrelevant, since the required properties of the skip-sampled series hold in all cases. Consider this transformation in the context of hyperbolic memory decay. Let the parameter index the rate of decay such that the autocovariance sequence of a stationary process satisfies
for some . The hyperbolic memory class includes short memory processes having summable autocovariances, such that , and long memory processes where for , and hence . It is immediately evident that, for any fixed, finite ,
It follows that for the long memory class, the property of the spectral density near the origin should likewise be invariant to the sampling frequency.
This is in contrast to the case of exponential memory decay where for every finite , but there exists such that
In this case, note that
so that the memory decay parameter rises from to following skip-sampling. Since the estimator of (spurious) in the exponential decay case is inevitably sensitive to the value of , this suggests that comparing estimates under different rates of sampling might yield a useful test of the null hypothesis of long memory.
A range of nonlinear models, such as threshold autoregressive and Markov-switching processes, are often thought of as likely to be to be mistaken for long memory, since they can exhibit local patterns of apparent persistence, switches of local mean, for example or, in the case of ESTAR threshold models, unit root-like behaviour in the neighbourhood of the origin. As for the linear autoregressive model, the essential difference between these models and the long-memory case is that the serial dependence decays exponentially as the lag increases beyond a certain point, whereas long memory implies hyperbolic decay. Whether linear or nonlinear, stable difference equations of finite order necessarily exhibit exponential decay (see Gallant and White 1988; Davidson 1994), whereas unstable difference equations are nonstationary, featuring unit roots or explosive behaviour.
The class of cases of  with count as instances of the alternative hypothesis for present purposes, because the autocovariances are summable. Models with this characteristic have not been significantly exploited in econometrics to date except in two rather special contexts, over-differenced fractional models (where and there is the additional “anti-persistence” property of the autocovariances summing to zero) and stochastic volatility modelling. The FIGARCH (Baillie, Bollerslev and Mikkelsen 1996) and HYGARCH (Davidson 2004) models are cases of the ARCH() model where the lag weights in the conditional variance equation decline hyperbolically but are nonetheless summable. The co-moments of fourth order (when they exist) are likewise summable in these latter models. In the present case, by contrast, our null hypothesis is that of true long memory with .
This paper considers tests of the long memory hypothesis based on a comparison of the log-periodogram estimator of the parameter in skip-sampled data with that from the original data. The test statistic is asymptotically standard Gaussian under the null hypothesis, given the usual assumptions of this literature (notably Gaussianity of the observations, see Robinson 1995; Hurvich, Deo, and Brodsky 1998). Convergence to the limit may be slow, and the formulation adopted depends on ancillary assumptions. The test is therefore implemented, for evaluation purposes, both as an asymptotic test and as a bootstrap test. We further recognize that the test is not consistent, for the rejection probabilities must be ultimately decreasing in sample size under the alternative hypothesis. However, we propose that the procedure be utilized as a component of a composite test, in combination with the Wald significance test on the fractional integration parameter in which the roles of null and alternative are reversed. The composite test can be formalized by the construction of a pseudo--value, and we show that this defines a test of the null of long memory that is both consistent and asymptotically correctly sized.
The paper is organized as follows. Section 2 reviews the important issue of aliasing in skip-sampled data, and its consequences for the form of the periodogram. Section 3 derives a bias-corrected form of the GPH estimator appropriate to skip-samples. Next, Section 4 describes the test procedure and derives the null asymptotic distribution of the statistic. Section 5 describes the implementation of the bootstrap version of the test. Section 6 describes the composite test, and Section 7 comments on the nonstationary case of the null hypothesis. Monte Carlo findings reported in Section 9, Section 9 describes two contrasting applications, and Section 10 contains concluding comments. Some proofs are gathered in the appendix.
The distribution of the GPH estimator in skip-sampled data has been studied inter alia by Smith and Souza (2002, 2004) and Souza (2005). Skip-sampling induces a bias in the estimator due to the effect of aliasing on the form of the spectral density. For a comprehensive analysis of the aliasing phenomenon, see Hassler (2011). The essential result is that the spectral density of the skip-sampled data can be represented as an average of the spectral densities over the range of aliased frequencies.
Ifis a discrete stationary stochastic process with spectral densityandforand, the spectral density of the processis
The straightforward proof is given in the appendix. Note that cycles of frequency in the original data become cycles of frequency in the skip-sampled data, and frequencies above are no longer identifiable. Hence, these contributions to the variance of the series are effectively aggregated with the identifiable frequencies.
In the fractionally integrated case where
with , we find that cannot be directly log-linearized in the GPH manner. What can be done, following the suggestion of Smith and Souza (2002), is to write
There is, evidently, an omitted term in the log-periodogram regression in skip-sampled data, depending on as well as The omission of this term will be liable to produce a bias in the GPH regression, and its omission is not rendered negligible by taking frequencies close to the origin. Indeed, what is commonly observed is that estimates of obtained from skip-sampled data are substantially closer to zero than those from the original data.
Note the implication for the standard analysis of a model such as , which is revealed to be specifically linked to the frequency of observation. Without this assumption, there is no reason to suppose that the function does not also depend on , nor that it is constant near the origin. In this light, the standard long memory analysis appears a little more fragile than is commonly taken for granted. Nonetheless, in this paper we shall work with the standard assumptions for the purposes of developing a test.
3 The Bias-Corrected Estimator
The test we propose is based on the comparison of two narrow-band regression estimators of the memory parameter , one based on the full sample, the other based on skip-sampling of the test series. As before, let denote the periodicity of the skips. Skip-sampling is done by taking every th observation, so yielding a sample of size where denotes the largest integer below . This can be done times, by off-setting the initial observation, so that the skip-samples can be represented as , where, for ,
Each of these samples can be used to compute a modified log-periodogram estimator, which we denote for .
As shown in Section 2, the conventional GPH estimator applied to skip-sampled data is biased. To be more specific, it exhibits a bias different in character from the well-known case of data with short run dependence in the fractional differences, being present even if the original series is a pure fractional without short-run components. Moreover, the bias is not attenuated by choosing a narrow bandwidth. The logarithm of in  is a missing term in the log-periodogram regression, and bias correction involves finding a computable surrogate for this function.
Expression  as a function of depends in the first place on the unknown , and the natural approximation is to replace this with the asymptotically unbiased estimator . It also depends on the unknown spectral density component evaluated at different points, and except in the case of the pure fractional model, the term in  varies with in general over the whole of the interval , including points close to the origin. Approximating it by a constant, in the manner of dealing with in the narrow-band estimator, is therefore not an attractive option.
Possible methods for estimating this term include constructing a kernel estimator of from the spectrum of the fractional differences. However, in this implementation we have adopted a semiparametric approach. Let the null hypothesis specify that the random sequence has a representation of the ARFI form
where is an invertible lag polynomial, of possibly infinite order, and where “NI” denotes independent Gaussian. To approximate we use the Durbin–Levinson algorithm to fit an autoregression of order to the fractional differences , where is the estimator of based on the full sample. This yields an estimated lag polynomial , and we then approximate by
It suffices for our application that converges in probability to pointwise in a neighbourhood of zero, and since is a smooth differentiable function of the data this property should in fact hold in a wider class of processes than . Absolute summability of the autocovariances of the fractional differences should hold in processes for which log-periodogram regression has good properties. Since the estimators in question depend only on second moments, they will yield the same consistency properties if in  is merely white noise. With caveats concerning invertibility, the Wold theorem therefore extends validity to the general covariance stationary case. The issues arising here are carefully analysed, in the bootstrap context, by Kreiss, Paparoditis, and Politis (2011). They show that Gaussianity of the series is certainly sufficient and this is, in any case, an assumption adopted for our subsequent asymptotic analysis and imposed in our experiments. Of course, these considerations strictly relate to the case where is known, and the largest source of error in finite samples will be due to the replacement of by .
Letting as usual, the skip-sampled series consists of observations, and the frequencies at which the periodogram is evaluated are for where , for , represents the usual GPH bandwidth function of sample size. In practice should be chosen according to the established prescriptions of the literature and, following Hurvich, Deo, and Brodsky (1998) (henceforth HDB), setting ensures limiting Gaussianity of the estimator, with bias of Let denote the periodogram computed from the th skip-sampled data set with period , and let denote the formula in  approximated as described, using the estimated parameters and the representation of the short-run spectral density in . The th bias-corrected skip-sample estimator then takes the form
where . Provided is treated as fixed and not linked to sample size note that where , and this is the assumption we maintain henceforth.
While the formula in  employs an estimator of the function as a component of the aliasing correction, note that this estimator has not been included in the log-periodiogram regression itself, and this remains a narrow-band estimator. Be careful to note that depends on the narrow-band estimator of based on the full sample, which is used to fractionally difference the data, and hence it does not provide a direct route to a broad-band estimation procedure of the type proposed by Moulines and Soulier (1999), for example.
4 The Skip-Sampling Test
Letting the conventional GPH estimator based on the complete sample be denoted , the test statistic we consider is
where . We use the signed statistic and perform a one-tailed test, on the assumption that the leading cases of the alternative will give rise to a smaller value of in the skip-sampled data. Also note that, in view of the form of the estimator, using the average of the estimates from the skip samples is equivalent to adopting the average of the log-periodogram points across the offset samples as regressand. This scheme makes the most efficient use of the available data.
When the sample is large enough, both the conventional GPH estimator and the skip-sampled estimator defined in  can be analysed using the techniques developed in HDB. These authors obtain their results from the following assumptions, which here relate to our null hypothesis under test.
The processis stationary and Gaussian with the spectral density given in  with.
, and, are bounded for allin a neighbourhood of zero.
Letting there exists a function such that (analogous to the expression in HDB page 42)
where and . Under our assumptions, the first right-hand side term is . In the case , such that there is no skip sampling and , and . In the cases with , on the other hand,
Since is twice-differentiable with respect to and is -consistent under our assumptions, we can expand as
Then, using Lemma 1 of HDB, and letting
Note that the relevant properties of the random variables extend from the full-sample to the skip-sampled case, specifically, that their distribution has finite second moments that asymptotically do not depend on nuisance parameters – see Lemmas 2 and 6–8 of HDB. Since the regressors are the same for each , we further find
where and In the appendix, we show the following.
For fixed finite, andsuch thatis a weakly dependent process, converges in probability to a finite nonstochastic limit.
The next thing to note using further results from HDB is that, with ,
Also note that has the same limit in distribution for each , where , and so similarly,
It follows that under these conditions,
where , , and
To derive a formula for analytically would entail quite a challenging calculation, and we have circumvented the need for this by a numerical evaluation. Note that is the logarithm of the periodogram point of an independent Gaussian series (having and constant) whereas the are the log-periodograms of the corresponding skip-sampled series. Therefore, can be approximated as closely as desired, for given and , by a simulation based on a sufficiently large sample. The accuracy of the approximation can be monitored by computing the sample variances of the components at the same time, and checking how close these lie to their known asymptotic counterpart of . We have performed the simulation with 200,000 replications in a sample size of 2000, with the results shown in Table 1.
Denoting by the variance formula computed using these approximations, replacing by its full-sample GPH estimator, the test statistic is calculated as
This statistic is used as the basis for a one-tailed test with rejections in the upper tail.
5 The Bootstrap Test
A difficulty with the semiparametric approach to estimation is the slow convergence to the asymptote, at the rate rather than . The mean and variance approximations derived in the previous section are accordingly slow to improve, especially with the reduction in effective sample size following skip-sampling. This suggests that the bootstrap may have a useful role to play in implementing the test, while not overlooking that the parametric bootstrap is likewise dependent on slowly converging estimated parameters. Nonetheless, a comparison of the two procedures, asymptotic and bootstrap, may serve to triangulate the uncertainty.
The bootstrap distribution of the statistic has to be estimated by simulating the null hypothesis as a fractionally integrated process, while allowing for the possibility of short-run dependence of the fractional differences. Given an estimator of the fractional parameter and test statistic computed from the sample, the calculation is performed as follows.
Compute the fractional differences where and , for .
Repeat the following steps for
Draw a random sample from the distribution of using a method that preserves the dependence structure; see Remark 1 below.
Construct the sequence
where where , for , and is explained in Remark 2 below.
Compute the bootstrap test statistic as in  for the sample .
Compute the estimated -value for the test as if or else as
where is the th order statistic for the bootstrap statistics .
Methods for constructing the drawings include the stationary bootstrap of Politis and Romano (1994) and the sieve autoregression method of Bühlmann (1997). Note that the latter calculation is also used to obtain expression , and the same remarks apply regarding the validity of the sieve AR method in this context; see Kreiss, Paparoditis, and Politis (2011).
The correction terms are constructed using Gaussian drawings and weights computed from the estimated parameters to have a covariance structure matching the components omitted through truncating the innovation sequence at . These replace the sample initial condition which has been truncated in step 1. The resulting sequence is approximately stationary for . If the data are modelled in differences, replacing by , and the simulation is then integrated using the first observation for the initial condition. Nonstationary processes generated by this procedure converge after normalization to Type I Brownian motion. For details of the simulation procedure, see Davidson and Hashimzade (2009).
In practice, different estimators of , employing different bandwidths in particular, might be used to compute the statistic and to implement the bootstrap resampling as in Step 1. Using a wider bandwidth in the former case could increase power by emphasizing bias under the alternative, while a different balance between bias and variance might be advantageous in achieving the best bootstrap distribution. Such choices can be guided by simulation experiments.
6 The Composite Test
Testing the degree of persistence of a time series is a problem that has attracted a degree of controversy, as documented by one of the present authors (Davidson 2009). This is one of a class of problems have been characterized by Dufour (1997) as “ill-posed”, and has close links with the testing frameworks critically analysed by Pötscher (2002) and Faust (1996, 1999), inter alia. Tests of the null hypothesis that the series has summable autocovariances – the “I(0) hypothesis” – face a common difficulty for valid inference. This difficulty manifests itself in different ways in different contexts, but the essential common feature is that the null hypothesis constitutes an open set in the parameter space. It follows that test power cannot exceed test size, where the latter is defined as the supremum of the rejection probabilities over the null set of the model space. While this problem extends to more general parameterizations it is most transparent in the case where the “I(0)” property relates to the modulus of the maximal autoregressive root. The null hypothesis is represented by the interval with its closure containing the leading case of the alternative.
Although the null and alternative are interchanged, the present case is clearly similar. The null hypothesis relating to the value of is the open interval , with its closure containing the cases of the alternative with This is another situation where, under a literal interpretation, power cannot exceed size. The test is based on a comparison of two estimators of , where under the alternative, one (the full-sample estimator) is expected to exhibit more bias than the other (the skip-sampled estimator) as estimators of zero. Since the estimators being compared are both consistent, albeit biased in finite samples, the test is evidently inconsistent. The probability of exceeding the rejection criteria under the alternative evidently cannot be monotone nondecreasing in sample size.
While the test might therefore appear of doubtful value in applications, this conclusion overlooks the context in which such a test might be applied. The question actually being posed, in most cases, is whether a “significantly positive” estimator of should be treated as a biased estimator of zero. If the significance test does not result in rejection, then we might on these grounds decide to reject the null hypothesis of long memory and either forego the skip-sampling test or, at least, to overlook a non-rejection in the latter test.
To formalize this idea, consider a composite test in which the skip-sampling test is performed in partnership with a one-tailed Wald test of the hypothesis with alternative . With contamination by short-run positive autocorrelation, we anticipate possible over-rejection in this latter test. Non-rejection in the Wald test implies effective rejection of the null hypothesis of long memory, and there is, arguably, no need to proceed to the skip-sampling test. Here’s how we can compute a composite test which delivers a -value taking account of the outcome of the initial Wald test. Suppose that the Wald test delivers a -value in a sample of size , and the skip-sampling test a -value . Consider the pseudo- value calculated as
The test “rejectwhen” is consistent and asymptotically of size
To prove the proposition, first consider the behaviour of the statistic in the case (the null hypothesis). Let the Wald statistic be denoted . The corresponding -value is the area under the upper tail of the null distribution (standard normal) bounded by . Since the Wald test is consistent with , we have
Hence, consider , and note that for every ,
It follows that in probability. Since the null hypothesis is true, converges to in distribution according to . Therefore, is asymptotically uniformly distributed on the unit interval. By the indicated convergence in probability the composite test shares this property, and so rejects asymptotically with probability in an -level test.
Next, suppose that the null is false, with . Recalling that the Wald test is one-sided, is asymptotically uniformly distributed on the unit interval in this case and so, in particular, . When , it follows that for any ,
as , and the proposition is proved.
The convergence of to the uniform distribution under the null must be somewhat slower than that of , depending on the choice of . The smaller that is chosen, the nearer is to unity in any given sample size and the smaller is the size distortion ceteris paribus, while not overlooking the fact that the test based on may itself exhibit size distortion in one direction or another, so that the net distortion in a given sample size is unpredictable. On the other hand, the larger is chosen, the more rapidly approaches 0 under the alternative. Hence, the choice of represents a trade-off of power against size.
Take care to note that the consistency of the composite test holds whether or not under the alternative. To appreciate the contribution of the skip-sampling test, it may be helpful to envisage the “test” based on simply drawing a uniform random number from at the second stage, instead of computing the quantile of the skip-sampling statistic. Proposition 6.1 holds also for this test! What we have done here is to give an alternative way of formalizing the properties of the Wald test. The tendency of this test to over-reject the conventional null hypothesis , due to bias, is converted into a case of low finite-sample power to reject the hypothesis . However, the expectation is that the power of the composite test in finite samples is greater, to the extent that is distributed closer to zero than a uniform variate under the alternative. The simulation experiments reported in Section 8 show that such improvements, judged by the performance of the basic skip-sampling test, can be large.
We emphasize once again that the composite test does not need to be taken literally as an operational procedure. We can think of it as a formalization of the procedure of taking two test results into account in making a decision. If we cannot reject the hypothesis on the Wald test, we are unlikely to proceed to the second stage. If we do find “significantly positive” on conventional criteria, then we want to know how far this outcome might be attributable to bias, and the skip-sampling test can in this case provide countervailing evidence.
7 The Nonstationary Case
As the observational equivalence issue raised in the introduction would lead us to predict, autoregressively generated series with a root in the stable region but close to unity characteristically yield an estimated in the nonstationary range . However, it is known (from, e.g. Velasco 1999; Kim and Phillips 2006) that log-periodogram regression in this range is consistent, and also asymptotically normal, under regularity conditions, for . Our test should exhibit similar characteristics in stationary and nonstationary cases of the null hypothesis, and this conjecture is borne out by the simulation experiments reported in Section 8.
In a well-known paper, Diebold and Inoue (2001) point out that in certain models exhibiting structural change, in which the frequency of change has a particular relation with sample size, there is the appearance of hyperbolic memory decay. In some of their examples, the processes in question are “revealed” as really I(1) (stationary in differences) as is extended with fixed parameters. To understand how the skip-sample test might behave in these cases, we must not overlook the fact that a unit root process, like a serially independent process, is technically a case of the null hypothesis. Both cases exhibit the invariance of memory to skip-sampling characteristic of fractional integration. Thus, a skip-sampled unit root remains a unit root. For this reason we should not expect the present test to have greatest power against local-to unity autoregressive alternatives. The natural approach, faced with a time series that does not exhibit mean reversion, might be to test for hyperbolic memory in the differences. Diebold and Inoue also propose examples in which processes appearing to show hyperbolic decay in a given sample size are “revealed” as I(0) as increases, and here our test should perform better. In particular, they consider a simple independent process subject to Markov-switching, which is one of the cases to be studied in the next section.
8 Monte Carlo Experiments
We present some experiments using three sample sizes, , and , with 5,000 replications in each case. Following preliminary investigations a bandwidth for the GPH estimator of was chosen to compute the tests, with a skip-sampling period of . A relatively wide bandwidth, emphasizing bias, is intended to optimize the performance of the test under the alternative. The skip-sampling periods and have also been tried, although the properties of the test do not appear very sensitive to this setting. These settings emerged as the best compromise in performance in null and alternative cases. The experiments returned both asymptotic and bootstrap -values, using 300 bootstrap replications. The fractional differencing of the series prior to resampling has to be performed with an estimated , as described in Section 5, and for this purpose a narrower bandwidth was used, to attenuate the bias. The simple bootstrap with independent resampling was used in the simulations of the pure fractional null hypothesis.
Table 2 shows the results obtained in nominal 5% tests for three cases of the null hypothesis. Under , the data are generated as
where is an independent Gaussian term generated by the method of Davidson and Hashimzade (2009), such that the sequence is stationary (see also Remark 2 of Section 5). The chosen values of are shown in the column headings. The table entries show the proportion of replications in which the asymptotic and bootstrap -values, respectively, fell below 0.05. The rows of the table show the performance in the three sample sizes of the basic skip-sampling test, and three cases of the composite test based on , with the values of shown in the first column of the table. The bootstrap -values were computed using the algorithm from Section 5 using simple independent sampling to draw the fractional differences.
Under-sizing of the asymptotic test occurs in all sample sizes, and also in the bootstrap test in the larger samples, suggesting that the convergence is non-monotone. If it is found surprising that these errors in rejection probability do not diminish more quickly, it is as well to remember that the components of the statistic depend upon as few as periodogram points, a mere 34 even in the case of . The convergence to the asymptote is inevitably slow.
The issue of size distortion would clearly benefit from further study, and alternative estimators and bandwidths could certainly be considered. However, we note that under-rejection is a relatively benign problem provided the rejection rates under null and alternative differ sufficiently. Moreover, because the composite test yields a pseudo -value that is always smaller than the bootstrap -value, under-rejection is a desirable feature in the sense that the composite test is less prone to over-rejection under the null. Considering the alternative cases of in the composite test, the trade-off between power and size is evident here. The rate of over-rejection by the pseudo -value under the null can be unacceptably large, but even with this is a problem chiefly in small samples, or when is close to zero. The latter results are not surprising, because when is small, in  is on average closer to 1 then otherwise, and hence smaller and over-rejection more acute. It is interesting that this effect persists in the largest sample considered but, again, it is important to remember that the slow rate of convergence. It is, if anything, more surprising to see how well the test can perform in modest sample sizes.
Next, in Tables 3 and 4 some cases of the alternative are shown. The performance of the simple and composite tests with the data generated as first-order autoregressions is reported in Table 3. This is the exponential decay model
with , for three different values of . Note that these are rejection rates, not power estimates, since the test sizes are uncorrected. With our compound null hypothesis, no consistent scheme for correcting rejection rates can be defined. This table therefore needs to be read in conjunction with Table 2. Its interesting to note that the rejection rate increases as is increased. This reflects the fact that the test has most power when autocorrelation is substantial but not hyperbolic. When the amount of autocorrelation is small, it is correspondingly difficult for the test to discriminate between exponential and hyperbolic decay, and this fact is reflected in the lower rejection rates observed.
Table 4 shows rejection rates, in the basic test only, against a range of nonlinear processes. Here, we report the averages over replications of the log-periodogram estimates of , side by side with the rejection frequencies by the bootstrap test, where we know in each case that the true is zero. The models reported are as follows, where in each case
“Bilinear” is of the form
with and .
“ESTAR” is the exponential self-exciting threshold AR case,
where , , .
“Markov Mean” is a model with Markov–switching intercepts. This takes the form
where , and or 2 with .
“Markov AR” is an autoregressive Markov-switching model,
where , and and .
Our final set of experiments examines rejection rates (basic test only) under what we might call the “contaminated null hypothesis”, in other words, models in which the fractional differences of the process are autocorrelated. In such cases we have to consider resampling the differences using a bootstrap for dependent data. We consider two cases of the stationary bootstrap of Politis and Romano (1994) with exponential block length distributions with mean block lengths of 5 and 10 observations, and also the sieve-autoregressive method of Bühlmann (1997), where the lag length for the autoregression is chosen by the Akaike criterion up to a maximum of 10 lags. Including the asymptotic criterion, this makes for five test variants in all. Table 5 shows the results for models and sample sizes as in Table 2, but with an autoregressive component with . Thus, these are cases of the ARFIMA(1,,0) class
|Block Mean 5||Block Mean 10|
The asymptotic test acquits itself relatively well here, and the independent bootstrap fails seriously only in the smallest sample. Under-rejection in the larger samples is again a feature of the findings, with the poorest performance delivered by the stationary bootstrap with the block-length 10.
We report two applications of the skip-sampling test. The first case considers a volatility measure for daily exchange rates of the British pound sterling against six currencies, in excess of 9,500 observations covering the period January 1975 to October 2012, (source: Bank of England). The measure in question is the logarithm of the absolute value of daily appreciation (log-change) augmented by 0.005. Taking logarithms normalizes the distribution by alleviating asymmetry and excess kurtosis, while adding the small constant overcomes the problem of days when zero change was recorded, so that the volatility measure is undefined.
|Wald test||Skip-sample test||Bias test|
|New Zealand Dollar||0.523||0||0.977||0|
Table 6 reports in the first column the estimated from GPH estimation with a bandwidth of where is sample size. The succeeding columns show the bootstrap -values for three tests: the usual Wald test (-test) of significance of , the skip-sampling test (skip period 8, bandwidth of ), and lastly the bias test of Davidson and Sibbertsen (2009) with a bandwidth of 0.92. The latter test has the null hypothesis of a pure fractional process, and tests for the presence of short-run autocorrelation in the fractional differences. The stationary block-bootstrap with a mean block-length of 5 was implemented with 299 bootstrap replications. Note that in these cases, given the significance test outcomes, the composite test could not return a result different from the simple skip-sampling test.
As can be seen, none of these skip-sampling tests leads to a rejection, so that the skip-sampling test reinforces the evidence from the Wald test that these series are long memory. The skip-sampling -values tend to appear at the upper end of the unit interval, which is expected behaviour of the bootstrap in samples of this size, given the Monte Carlo findings. Even allowing for this distortion, however, the evidence in favour of the null hypothesis appears unequivocal.
Our second application is to the growth (log-change) in Robert Shiller’s S&P500 monthly real dividends series for January 1871–June 2012. First, consider the sub-period starting in January 1946 (798 observations) with the results shown in the first row of the Table 7. The estimation and test settings are the same here as for the previous example and, once again, note that the composite test cannot return different findings. This result suggests that the long memory indicated by the Wald test is spurious. However, we have a more direct check on this finding in the present case, by extending the sample. The result with the full set of 1,697 observations, starting in February 1871, is shown in the second row of Table 7, where the Wald test -value falls emphatically in the non-rejection region of the 1-tailed test. This result shows how the biases in log-periodogram regression can persist in large samples, but also how the skip-sampling test offers the possibility of providing counter-evidence to this spurious significance.
|Wald test||Skip-sample test||Bias test|
In this paper we have investigated the performance of a test for the null hypothesis of long memory, based on the self-similarity property of sequences with hyperbolic memory decay. The idea is to compare GPH log-periodogram estimators in original and skip-sampled versions of the data set. The aliasing phenomenon, which introduces an estimation bias in skip-samples, poses a problem for the implementation of this test, but a bias-corrected estimator permits the construction of an asymptotically pivotal statistic.
The use of a semiparametric method to construct the estimators and test statistic, with correspondingly slow convergence to the asymptote, inevitably poses a challenge for the implementation of the test, and the bootstrap variant of the test performs relatively well, in spite of being implemented using semi-parametric estimates of the null distribution. This combination of factors proves to pose a problem of under-rejection even in quite large samples, in the Monte Carlo evaluations. A bias reduction strategy such as the double bootstrap (Beran 1988) might alleviate this problem, at the cost of a large computational overhead, but in the settings where the test might be applied, under-rejection is a relatively benign problem, and the experiments indicate reasonable power properties. Alternative choices of test settings, such as GPH bandwidths and skip period, could prove helpful, although to research these must take us beyond the scope of this study. Notwithstanding these qualifications, the test may prove a useful addition to the arsenal of diagnostic procedures for long memory models, beside the bias test of Davidson and Sibbertsen (2009), which compares log-periodogram estimates with different bandwidths, and the aggregation test of Ohanissian, Russell, and Tsay (2008).
A.1 Proof of Proposition 2.1
Let denote the th autocovariance, defined by the identity
For the skip-sampled data with sampling period , the autocovariances are where
where the third equality makes use of the fact that , and the fourth one makes the change of variable and the substitution
A.2 Proof of Proposition 4.1
note first that
We obtain a formula for the derivative in the second term, and show that this is bounded in the limit. The terms of the form  depend on because the data used to construct the sieve autoregressive estimates are the fractional differences of the measured data. Assume that is fixed, and let and so let be the normalized data matrix whose columns are the vectors for . Also, let for denote the matrix equal to except that the th column has been replaced by . Then, note that the coefficients in the autoregression of order can be written using Cramer’s rule as
Let these elements define the -vector by also putting .
Now, let denote the real part of the Fourier matrix with elements for . Setting and , note that
where is the -vector having elements and for . In this notation we have
and it remains to evaluate the second right-hand side factor.
Start with the elements of the matrices. Considering row , let denote the generic lag associated with a column of . Using the argument from Tanaka (1999), Section 3.1, the derivatives with respect to can be written as
where the last equality defines . We have from Magnus and Neudecker (1988, 149), that for
(defining ) where the denote the matrices with elements , with the value of defined as appropriate, according to the construction of . Letting denote the vector with elements and , for , we now have the result
Since is a weakly dependent process by hypothesis, the process is covariance stationary. It follows directly that, for every finite , converges in probability to a nonstochastic limit, depending on the autocovariances of . From the fact that converges in the same manner, and the Slutsky theorem, the proposition follows under the conditions stated.
Two simplifying assumptions have been made to reach this conclusion. The first is that has been constructed as an infinite order moving average, whereas in practice the sums will be truncated, containing only the first terms. However, since the truncation affects at most a finite number of terms, this cannot change the value of the limit. Second, the lag length has been assumed fixed. However, since is a weakly dependent process by hypothesis, the autocovariances are summable and hence equal zero for lags exceeding some finite value. Letting tend to infinity with cannot change the distribution of beyond some point, since the additional elements of and have sums converging to zero as increases.
We thank David Peel for helpful discussions on this problem, and an anonymous referee for perceptive comments which have materially improved the paper.
Agiakloglou, C., P.Newbold, and M.Wohar. 1993. “Bias in an Estimator of the Fractional Difference Parameter.” Journal of Time Series Analysis14:235–46. Search in Google Scholar
Baillie, R. T., T.Bollerslev, and H. O.Mikkelsen. 1996. “Fractionally Integrated Generalized Autoregressive Conditional Heteroscedasticity.” Journal of Econometrics74:3–30. Search in Google Scholar
Beran, R.1988. “Prepivoting Test Statistics: A Bootstrap View of Asymptotic Refinements.” Journal of the American Statistical Association83:687–97. Search in Google Scholar
Bühlmann, P.1997. “Sieve Bootstrap for Time Series.” Bernoulli3:123–48. Search in Google Scholar
Chambers, M. J.1998. “Long Memory and Aggregation in Macroeconomic Time Series.” International Economic Review39 (4). Symposium on Forecasting and Empirical Methods in Macroeconomics and Finance (Nov., 1998):1053–72. Search in Google Scholar
Davidson, J.1994. Stochastic Limit Theory: An Introduction for Econometricians. Oxford: Oxford University Press. Search in Google Scholar
Davidson, J.2004. “Moment and Memory Properties of Linear Conditional Heteroscedasticity Models, and a New Model.” Journal of Business and Economics Statistics22 (1):16–29. Search in Google Scholar
Davidson, J.2009. “When Is a Time Series I(0)?” Chapter 13, pages 322–342, of The Methodology and Practice of Econometrics, a Festschrift for David F. Hendry, edited by J. Castle and N. Shepherd. Oxford: Oxford University Press. Search in Google Scholar
Davidson, J., and N.Hashimzade. 2009. “Type I and Type II Fractional Brownian Motions: A Reconsideration.” Computational Statistics and Data Analysis53 (6):2089–106. Search in Google Scholar
Davidson, J., and P.Sibbertsen. 2009. “Tests of Bias in Log-Periodogram Regression.” Economics Letters102:83–6. Search in Google Scholar
Diebold, F. X., and A.Inoue. 2001. “Long memory and regime switching.”Journal of Econometrics105:131–59. Search in Google Scholar
Dufour, J. -M. 1997. “Some Impossibility Theorems in Econometrics with Applications to Structural and Dynamic Models.” Econometrica65 (6):1365–87. Search in Google Scholar
Faust, J. 1996. “Near Observational Equivalence and Theoretical Size Problems with Unit Root Tests.” Econometric Theory12 (4):724–31. Search in Google Scholar
Faust, J. 1999. “Conventional Confidence Intervals for Points on Spectrum Have Confidence Level Zero.” Econometrica67 (3):629–37. Search in Google Scholar
Gallant, A. R., and H.White. 1988. A Unified Theory of Estimation and Inference for Nonlinear Dynamic Models. Oxford: Basil Blackwell. Search in Google Scholar
Geweke, J., and S.Porter-Hudak. 1983. “The Estimation and Application of Long-Memory Time Series Models.” Journal of Time Series Analysis4:221–37. Search in Google Scholar
Hassler, U. 2011. “Estimation of Fractional Integration under Temporal Aggregation.” Journal of Econometrics162:240–7. Search in Google Scholar
Hurvich, C. M., R.Deo, and J.Brodsky. 1998. “The Mean Squared Error of Geweke and Porter-Hudak’s Estimator of a Long Memory Time Series.” Journal of Time Series Analysis19:19–46. Search in Google Scholar
Kim, C. S., and P. C. B.Phillips. 2006 “Log-Periodogram Regression: The Nonstationary Case.” Cowles Foundation Discussion Paper 1587, Yale University. Search in Google Scholar
Kreiss, J. -P., E.Paparoditis, and D. N.Politis. 2011 “On the Range of Validity of the Autoregressive Sieve Bootstrap.” Annals of Statistics39:2103–30. Search in Google Scholar
Magnus, J. R., and H.Neudecker. 1988. Matrix Differential Calculus with Applications in Statistics and Econometrics. Chichester: John Wiley & Sons. Search in Google Scholar
Moulines, E., and P.Soulier. 1999. “Broad Band Log-Periodogram Estimation of Time Series with Long-Range Dependence.” Annals of Statistics27:1415–39. Search in Google Scholar
Ohanissian, A., J. R.Russell, R. S.Tsay, et al. 2008. “True or Spurious Long Memory? a New Test.” Journal of Business and Economic Statistics26 (2):161–75. Search in Google Scholar
Politis, D. N., and J. P.Romano. 1994. “The Stationary Bootstrap.”Journal of the American Statistical Association89:1303–13. Search in Google Scholar
Pötscher, B. M. 2002. “Lower Risk Bounds and Properties of Confidence Sets for Ill-Posed Estimation Problems with Applications to Spectral Density and Persistence Estimation, Unit Roots and Estimation of Long Memory Parameters.” Econometrica70 (3):1035–65. Search in Google Scholar
Robinson, P. 1995. “Log-Periodogram Regression of Time Series with Long-Range Dependence.” Annals of Statistics23:1048–72. Search in Google Scholar
Smith, J., and L. R.Souza. 2002. “Bias in the Memory Parameter for Different Sampling Rates.” International Journal of Forecasting18:299–313. Search in Google Scholar
Smith, J., and L. R.Souza. 2004. “Effects of Temporal Aggregation on Estimates and Forecasts of Fractionally Integrated Processes: A Monte-Carlo Study.” International Journal of Forecasting20:487–502. Search in Google Scholar
Souza, L. R. 2005. “A Note on Chambers’s Long Memory and Aggregation in Macroeconomic Time Series.” International Economic Review46 (3):1059–62. Search in Google Scholar
Tanaka, K. 1999. “The Nonstationary Fractional Unit Root.” Econometric Theory15:549–82. Search in Google Scholar
Velasco, C. 1999. “Non-Stationary Log-Periodogram Regression.” Journal of Econometrics91:325–71. Search in Google Scholar
©2015 by De Gruyter