Distribution of the Least Squares Estimator in a First-Order Autoregressive Model





By

Mukhtar M. Ali

Department of Economics

University of Kentucky

Lexington, KY 40506

E-mail: MMALI1@POP.UKY.EDU




















Preliminary. Not to be quoted without February, 1996

permission from the author. Revised Oct., 1996

Distribution of the Least Squares Estimator in a First-Order Autoregressive Model




SUMMARY

This paper investigates the finite sample distribution of the least squares estimator of the autoregressive parameter in a first-order autoregressive model. Uniform asymptotic expansion for the distribution applicable to both stationary and nonstationary cases is obtained. Accuracy of the approximation to the distribution by a first few terms of this expansion is then investigated. It is found that the leading term of this expansion approximates well the distribution. The approximation is, in almost all cases, accurate to the second decimal place throughout the distribution. In the literature, there exists a number of approximations to this distribution which are specifically designed to apply in some special cases of this model. The present approximation compares favorably with those approximations and in fact, its accuracy is, with almost no exception, as good as or better than these other approximations. Convenience of numerical computations seems also to favor the present approximations over the others. An application of the finding is illustrated with examples.

JEL Classification: C13, C22

Key Words: Unit Root; Saddlepoint Approximation; Asymptotic expansion.

Distribution of the Least Squares Estimator in a First-Order Autoregressive Model

1. INTRODUCTION

Consider the first-order autoregressive model

(1.1) yt = ßyt-1 + et, t = 1, 2, ..., n

where ß is a real constant and the errors et's are identically independently distributed normal variables each with mean = 0 and variance = se2. The initial observation, y0 can be treated as fixed or stochastic. When y0 is stochastic, it is assumed to have a normal distribution with mean 0 and variance se2/(1 - ß2), ½ß½ < 1, and is independent of et's. For ease of exposition, these cases are identified with three models, each corresponding to a different nature of y0. When y0 is fixed at 0, it is named Model A; and Model B when y0 is fixed at a non-zero real constant; and Model C when y0 is N(0, se2/(1 - ß2)) and ½ß½ < 1. When y0 is stochastic, the model is stationary. When y0 is fixed, if ½ß½ < 1, the model is stationary asymptotically. If ½ß½ = 1, this is a well-known random walk model and if ½ß½ > 1, the model is explosive. The random walk model implies unit root hypothesis. Recently, there has been an enormous interest in testing for unit root. They include, among others, Dickey (1976), Dickey and Fuller (1979, 1981), Evans and Savin (1981, 1984), Fuller (1976), Hasza and Fuller (1979), Perron and Phillips (1987), Phillips and Perron (1988), Schwert (1987) and Stock and Watson (1989). Diebold and Nerlove (1990) have given an excellent survey of works in this area.

The unknown parameter ß is customarily estimated by its least squares estimator

(1.2) .

Under the assumption that et's are normally distributed, this is also the maximum likelihood estimate of ß. The distribution of has been studied extensively. Unfortunately, the exact distribution of , in closed form is unknown. Asymptotically, it has a normal distribution (Mann and Wald, 1943) if ½ß½ < 1, a Cauchy distribution (White, 1958) if ½ß½ > 1 and a non-standard distribution (White, 1958, Rao, 1978) if ½ß½ = 1. These asymptotic distributions can be used to approximate the finite sample distribution of . Such an approximation results in a nonsmooth transformation from a normal distribution to a non-standard distribution to a Cauchy distribution. Because the exact distribution of is continuous for all values of , this suggests that the asymptotic distribution would not approximate adequately the finite sample distribution especially near the discontinuity point of = 1. In fact it has been found that, unless ß is close to zero, these asymptotic distributions do not approximate well the distribution in finite samples (Evans and Savin, 1981; Tsui, 1989). The non-standard limiting distribution when ½ß½ = 1 (Rao, 1978) seems to approximate well the finite sample distribution when ½ß½ is close to 1, but it is too complicated for practical use. An accurate approximation to this limiting distribution can, however be obtained from the asymptotic expansion given by Abadir (1993) .

The distribution of is not known in closed form but it can be obtained numerically. The exact distribution has been numerically computed by Phillips (1977, 1978), Evans and Savin (1981) for selected values of sample size n and the autoregressive coefficient ß, and comprehensively for a wide range of (n, ß) by Tsui (1989) and Tsui and Ali (1994). However, these numerical approaches are often computationally demanding and expensive even with modern high speed computers. Besides the numerical methods, several authors (Dickey, 1976; Fuller, 1976) have performed Monte Carlo experiments to tabulate the distribution in the case of ß = 1 and y0 = 0 (special case of Model A). These distributions can be used to test for unit root hypothesis but are not much of use for further inference about ß.

As an alternative to numerically computing (or simulating) the exact distribution, several authors have attempted to obtain convenient approximations to the distribution of in finite samples. Phillips (1977, 1978), Satchell (1984), and Tsui and Ali (1992) have examined approximations by Edgeworth expansion and found them, except for ß close to 0 and only at the center of the distribution, to be unsatisfactory. Tsui and Ali (1992) examined also approximations by Cornish-Fisher-type (Cornish and Fisher, 1937; Fisher and Cornish, 1960; Hill and Davis, 1968) expansions and the four-parameter Pearson distributions. Accuracy of these approximations was found to depend substantially on sample size and the values of the autoregressive coefficient. None of these approximations was found to be reliable when the autoregressive coefficient is moderately large and the sample size is small. In the case of Model C, Phillips (1978) derived a saddlepoint approximation to the probability density function. An approximation to the distribution can be obtained by numerically integrating this approximate density function. He found this approximation to be exceptionally accurate, certainly for sample sizes as large as 30. Unfortunately, the approximation was not defined over a sizable region of the tail for values of the autoregressive parameter greater than 0.4. For the same case of Model C, Wang (1992) derived a saddlepoint approximation to the distribution function which is available over the entire range of . Wang's approximation was found to be exceptionally accurate. Lieberman (1994) derived an alternative saddlepoint approximation to the probability density of which is also available over the entire range of . From an illustrative check for the accuracy of the distribution function derived from this approximate density, the approximation was found, for the Model C, to be excellent for both n = 10 and n = 30 and for all four ß values of .2, .4, .6 and .8 that were examined. The approximation seems promising but its accuracy has been tested for only a few values of n and of ß and for only the model C. Furthermore, this approximation can be computationally demanding (even to the point of being impractical), especially for large sample sizes. This is because a crucial step in the implementation of the approximation is to obtain a solution of a highly non-linear equation which requires either a computation of eigenvalues of an nxn matrix or repeated inversion of nxn matrices. Moreover, expensive numerical integration may be needed to obtain the distribution function from the approximate density function.

Phillips (1988) considered a near-integrated random process where the autoregressive parameter is defined by

(1.3) = exp(c/n).

c is a real constant measuring the deviation from the unit root case. Cavanagh (1986), Nabeya and Tanaka (1990) and Perron (1989) examined , in the cases of Model A and Model B, the limiting distribution of under (1.3) as an approximation to the finite sample distribution of . They claim the approximation to be quite good in the case of Model A. It is computationally demanding but this limiting distribution can be computed numerically and it has been tabulated by these authors. Larsson (1995) has, however provided a convenient approximation to this limiting distribution in the case of Model A. This approximation may serve as an approximation to the finite sample distribution of . In search of further approximation to the distribution of , Perron (1991a, b) considered the continuous time Ornstein-Uhlenbeck process dyt = ytdt + dwt, where wt is a Wiener process. Let be the continuous time maximum likelihood (conditional upon y0) estimator of based upon a single path of data of length T. He then advanced the exact distribution of as an approximation to the finite sample distribution of . Again, it is computationally demanding but the distribution of can be computed numerically. Perron (1991a, b) have provided some selected percentage points of this distribution.

In summary, it seems, with all the shortcomings, only viable approximations available in the literature are those by Phillips (1978), Perron (1991a), Wang (1992) and Lieberman (1994) in the case of Model C, those by Perron (1991b), Lieberman (1994) and Larsson (1995) in the case of Model A and that by Perron (1991b) in the case of Model B.

In this paper, an uniform asymptotic expansion for the distribution (not the density function) of the least squares estimator of ß is obtained. This expansion is applicable for all the three models, Model A, Model B and Model C. An alternative expression of the joint characteristic function of the numerator and denominator of as a product in trigonometric functions avoids the costly computation of eigenvalues or repeated inversion of high order matrices that requires in implementing Lieberman (1994) approximation. Accuracy of approximation to the distribution by a first few terms of this expansion is investigated for a wide range of values of n and of ß and for all the three models. It is found that the very first term of this expansion approximates well the distribution, especially at the extreme tails. The approximation is, in almost all cases, accurate to the second decimal place throughout the distribution. The accuracy improves by including further term beyond the first term of this expansion in the approximation. But occasionally the accuracy of such an approximation with additional term(s) deteriorates and sometimes it deteriorates to the point of being useless giving probability value which is outside the range of (0, 1). The first-term approximation compares favorably with the approximations given by Phillips (1978), Wang (1992) and Lieberman (1994) in the case of Model C, is as good as and in general, better than the approximations given by Larsson (1995) in the case of Model A , Perron (1991b) in the case of Model A and Model B and Perron(1991a) in the case of Model C.

The plan of this study is as follows. The uniform asymptotic expansion for the distribution of is obtained in section 2. The accuracy of the expansion to the distribution is examined in section 3. Accuracy of the first-term approximation is then compared with those given by Phillips (1978), Perron (1991a), Wang(1992) and Lieberman (1994) in the case of Model C; Perron (1991b), Lieberman (1994) and Larsson (1995) in the case of Model A; and Perron(1991b) in the case of Model B. These comparisons are also reported in section 3. Examples illustrating the use of the results of this paper are reported in section 4. Some concluding remarks are given in section 5.

2. ASYMPTOTIC EXPANSION

There seems to have been some confusion in the literature in regards to the sample size n. Following Hurwicz (1950), the sample size is taken as the number of stochastic yt in (1.1). Thus, the sample size is n in Models A and B where y0 is fixed and it is (n+1) in Model C where y0 is stochastic. Defining

(2.1) P = and Q = ,

the least squares estimator for all the three models (A, B and C) is = P/Q.

Let y(iu, iv) = E(eiuP+ivQ) be the joint characteristic function of P and Q. Then, by Theorem 1 of Gurland (1948, p. 229), the cumulative distribution function of is given as

(2.2) G(w) = Pr(< w) = Pr(P/Q < w)

=

A change of variable (new u = old ui) in (2.2) leads to

(2.3) G(w) =

where L is a path of integration made up of two segments: from -¥i to -di, and from di to ¥i. It can be shown that the integrand has a simple pole at u = 0 and y(0, 0) = 1, so that, by Cauchy's integral formula,

(2.4)

where C is any closed curve encircling no singularity other than u = 0, in the positive (counterclockwise) direction. Suppose that the curve C is a circle of radius d, then we can rewrite (2.3) as

(2.5) G(w) = ,

where the new path of integration R is obtained by adding that half of the circle C for which Re(u) > 0 to the original path L. It can be shown that the integrand is analytic for Re(u) > 0. Thus, by the well known theorem of Cauchy, the path of integration R can be modified to obtain

(2.6) G(w) = , c > 0.

It can be seen that the integrand in (2.6) has a simple pole at u = 0 and following Lieberman (1994), it can be shown that the integrand has only one saddlepoint. Defining, h(u) = -(1/n)lny(u, -uw), the integral in (2.6) is, then exactly in the form of that in equation (65) in Rice (1968). Hence, from equation (68) of Rice (1968), the uniform asymptotic expansion for G(w) is given by

(2.7) G(w) = ,

where F(.) and f(.) are, respectively the distribution and probability density function of a standard normal variable,

pj = , j ³ 0,

u1 is the saddlepoint, h(0) = 0, h(r)(u) = , r ³ 0, h(0) = h, h1(r) = h(r)(u1), h1 = h(u1), h(1)(u1) = 0,

x = (-h1)(1/2)u1/½u1½, z = u1(h1(2)/2)(1/2), (q)0 = 1, (q)j = q(q + 1) ... (q + j - 1), b00 =1, b0i = 0, for i ³ 1, and

bk+1,i+1 = ,

ds = -2h1(s+2)/[(s+2)!h1(2)], s ³ 1.

In deriving the expansion in (2.7), it is assumed that the origin, at which there is a simple pole for the integrand in (2.6), does not coincide with the saddlepoint. If the saddlepoint is at the origin, a classical saddlepoint analysis (Lugannani and Rice, 1980, p. 479) can be applied to obtain an asymptotic expansion for G(w) as

(2.8)

where qi = h(i)(0)/(i![h(2)(0)](i/2)), i = 3, 4, ....

3. ACCURACY OF THE EXPANSION

The distribution can be computed numerically from (2.7) (or (2.8)). A major problem is the evaluation of y(u, -uw) and hence of the function h(u) that appears in the integrand in (2.6). Lieberman (1994) obtained y(u, -uw) as a determinant of an nxn matrix depending on w. He also expressed it as an elementary function of eigenvalues of the same nxn matrix. For either case, as it requires determinant or eigenvalues of an nxn matrix for each w, the computation becomes prohibitively time consuming as the sample size n becomes large, say larger than 200. From White (1958), an expression for y(u, -uw) can be derived in closed form. It, however involves in raising some (real) expressions to the power of sample size n. Thus, for large sample sizes, it becomes problematic to maintain reasonable numerical accuracy. Alternatively, following White (1961), y(u, -uw) can be expressed as a polynomial of degree n, the sample size. Unfortunately, this creates computational problems in large sample sizes because this polynomial of high degree often contains a large number of terms each of which is negligible individually but significant collectively. The expression for y(u, -uw) that is found to be most convenient is the one obtained by Tsui (1989) and Tsui and Ali (1994). This expression is in trigonometric functions and is given as

(3.1) y(u, -uw) = Dn(u, -uw)-1/2, for Model A

= Dn(u, -uw)-1/2, for Model B

= Ö(1 - ß2)[Dn+1(u, -uw) - ß2Dn(u, -uw)]-1/2, for Model C

where a = y0/se,

(3.2) Dn(u, -uw) = Sn-1(u, -uw) - (ß + u)2Sn-2(u, -uw), and

(3.3) Sn(u, -uw) = .

Using expression (3.1), it is a routine matter to compute h(u) and its various derivatives that are required to obtain the distribution function from (2.7) (or, (2.8)). In this computations, the basic function to evaluate is Sn(u, -uw) and its derivatives. As Sn(u, -uw) = Sgn[Sn(u, -uw)]½Sn(u, -uw)½, where Sgn(x) = 1, if x ³ 0 and = -1, if x < 0, the function Sn(u, -uw) and its derivatives can be obtained from those of ½Sn(u, -uw)½ and its derivatives. In turn, the function ½Sn(u, -uw)½ and its derivatives may be evaluated utilizing the relation

(3.3) ln½Sn(u, v)½ = .

It may be noted that the saddlepoint u1 is a solution to the equation

(3.4) h(1)(u1) = 0.

The equation (3.4) can be solved by the iterative Newton-Raphson method starting at u1 = 0 and modifying, if necessary the step size at each iteration so that y(u, -uw) is positive.

The expansion (2.7) (or, (2.8)) can be used to compute the distribution function, G(w). Unfortunately, most asymptotic expansions are nonconvergent, with the magnitude of successive terms tracking a J curve of initial decline followed by a steep rise. Fortunately, often a few of the beginning terms provide adequate approximation. To check for the accuracy of the approximation to the distribution of by a few terms in the expansion (2.7) (or, (2.8)), we consider three approximations, LEAD, LEAD2 and LEAD3 which are obtained by truncating the expansion (2.7) (or (2.8)) at the leading term, the second term and the third term, respectively. More specifically,

(3.5) LEAD: expansion (2.7) truncated at j = 0 (or, expansion (2.8) truncated to the terms at most of order O(n-1/2))

(3.6) LEAD2: expansion (2.7) truncated at j = 1 (or, expansion (2.8) truncated to the terms at most of order O(n-3/2))

(3.7) LEAD3: expansion (2.7) truncated at j = 2 (or, expansion (2.8) truncated to the terms at most of order O(n-5/2))

To check for the accuracy of these approximations, the exact distributions obtained by evaluating numerically the integral in (2.2) (see Tsui (1989) and Tsui and Ali (1994)) are compared with these approximations. A fortran program is written to implement these approximations and can be obtained on request. For a thorough investigation, the distributions and the approximations were computed comprehensively for Models A, B and C. In all cases, we take various sample size n = 10, 20, 30, 40, 50, 75, 100, 150, 200, 250, 300 and 500. For Models A and B, choices of autoregressive coefficient, ß include 0.4, 0.6, 0.8, 0.9, 0.95, 0.99, 1.0, 1.01. For Model B, the values of the parameter a are: 1 and 4. As for Model C, choices of autoregressive coefficient, ß are: 0.4, 0.6, 0.8, 0.9, 0.95, 0.975, 0.99. In each case, the distribution G(w) is computed at 33 percentile points, w = x/g(n) + ß (x = -16.0, -12.0, -8.0, -6.0, -4.0, -3.5, -3.0, -2.8, -2.6, -2.4, -2.2, -2.0, -1.8, -1.6, -1.4, -1.2, -1.0, -0.8, -0.6, -0.4, -0.2, 0.0, 0.2, 0.4, 0.6, 0.8, 1.0, 1.2, 1.4, 1.6, 1.8, 2.0 and 4.0) where g(n) = , if ½ß½ < 1, = , if ½ß½ = 1.0 and = , if ½ß½ > 1. It may be worth observing that in this experimentation, we have taken only positive values of ß. This is because the distribution of for a given ß is the mirror-image of that for - ß (see Cryer, Nankervis and Savin, 1989). To save space, Tables 1, 2 and 3 reproduce some selected but representative results for Models A, B and C, respectively. All the exact and approximate distributions that are computed can be obtained upon request.

Comparing the approximations with the exact, some definite conclusions can be drawn. In particular, it is found that, for all cases considered, LEAD approximates exceptionally well, especially at the tails of the distributions where most of the inferential interests lie. For example, for sample size as small as 10, the LEAD approximations of 0.0223, 0.0177 and 0.0357 to Pr(g(n)( - ) < x) for x = -6.0 and = 1.0 in Model A, for x = -6.0, = 1.0 and = 1 in Model B and x = -6.0, = 0.95 in Model C , respectively match quite closely to the corresponding exact values of 0.0208, 0.0165 and 0.0331 (see Tables 1, 2 and 3). As can be seen from the entries in Tables 1, 2, and 3, LEAD approximation matches, in almost all the cases to the second decimal place and quite often to the third decimal place to the corresponding exact value. Almost without exception, the LEAD approximates exceptionally well at the tails of the distribution (where most of the inferential interests lie) matching the exact value at the third decimal place. Quite often LEAD2 or LEAD3 provides more accurate approximation than that given by LEAD. Thus, for example, while the exact value of Pr(g(n)( - ) < x) for n = 10, x = -6.0, = 1.0 in Model A is 0.0208, the approximate values from LEAD, LEAD2 and LEAD3 are, respectively 0.0223, 0.0203 and 0.0211 (see Table 1). Unfortunately, in a number of cases the accuracy of LEAD2 and LEAD3 deteriorates to the point of being useless giving probability outside the range of (0, 1). For example, LEAD approximation of 0.6428 to Pr(g(n)( - ) < x) for x = -0.4, = 1.01, n = 10 in Model A (see Table 1) matches well the exact value of 0.6396 but both approximations LEAD2 and LEAD3 provide values which are outside the range of (0, 1).

Preceding analysis suggests that, for all practical purposes, LEAD provides excellent approximation to Pr(g(n)( - ) < x) for all three Models A, B and C. To shed further light on its accuracy, in what follows we compare its accuracy with several other approximations that are available in the literature, namely the approximations by Perron (1991b), Lieberman (1994) and Larsson (1995) in the case of Model A; Perron (1991b) in the case of Model B; and Phillips (1978), Perron (1991a), Wang (1992) and Lieberman (1994) in the case of Model C.

In the case of Model A, Perron (1991b) provides an approximation to the distribution of the least squares estimator and tabulates a selected set of percentage points as approximations to the percentage points of the distribution of . These percentage points were tabulated for a number of parameter values of c where = exp(c/n). At these percentage points, we evaluate the Lieberman (1994), Larsson (1995) and our LEAD approximations to the the distribution function for . We also obtain "exact" evaluation of the distribution function at these percentage points by simulation with one-half of one million replications. A selected but representative set of these evaluations of the distribution function is reported in Table 4. Lieberman (1994) approximation matches the LEAD to the fourth decimal place throughout our comparison and hence it is not shown in this table. It can be seen from the entries in this table that Larsson (1995) approximation is not available for a part of the region of the distribution function but whenever it is available it provides an accurate approximation to the Perron (1991b) approximation; and that LEAD approximates well and matches the distribution function, almost always to the second decimal place. Accuracy of Perron (1991b) or Larsson (1995) approximation is rather disappointing. For example, for the case when c = -5.0 (see Panel A in the table), at the percentage point of -10.4633, while the exact evaluation of the distribution function, at n = 10, = 0.6065 is 0.0039, Perron and Larsson approximations are , respectively 0.050 and 0.0580. The accuracy of both Perron and Larsson approximations improves as the sample size and/or is increased. Thus, in the above example, at n = 50, = 0.9048, the exact evaluation of the distribution function is 0.0358 which is closer to the Perron and Larsson approximations of 0.050 and 0.0580 than it was when n = 10, = 0.6065. Even then, for a sample of size as large as 50, the accuracy of both Perron and Larsson approximations is questionable. However, it may be of some comfort to know that both Perron and Larsson provide more accurate approximation to the distribution at the right tail than at the left tail and both approximations are reasonably accurate when and/or n are sufficiently large. Thus, in the case of c = 2.0, n = 50 and = 1.041, at the percentage point of 0.7585, the exact evaluation of the distribution function of 0.9425 is reasonably close to the Perron approximation of 0.95 and also to the Larsson approximation of 0.9332.

In the case of Model B, Perron (1991b) provides an approximation to the distribution of the least squares estimator and tabulates a selected set of percentage points as approximations to the percentage points of the distribution of . These percentage points were tabulated for a number of parameter values of c and where = exp(c/n) and = n1/2. At these percentage points, we evaluate our LEAD approximation to the distribution function for . We also obtain "exact" evaluation of the distribution function at these percentage points by simulation with one-half of one million replications. A selected but representative set of these evaluations of the distribution function is reported in Table 5. Examining the entries, it is clear that, at every percentage point, the LEAD approximation outperforms the Perron approximation. While the LEAD approximation matches the exact evaluation, almost always to the third decimal place and often to the fourth decimal place, the Perron approximation fails often to match even to the second decimal place. Accuracy of the Perron approximation improves as the sample size is increased but it has a questionable accuracy even for a sample of size as large as 50. For example, for c = -5.0, = 0.5, n = 50, = 0.9048, = 3.54, at the percentage point of -8.7550, while (see Table 5, Panel A) the exact evaluation of the distribution is 0.0369, the Perron approximation is 0.050.

In the case of Model C, Perron (1991a) provides an approximation to the distribution of the least squares estimator and tabulates a selected set of percentage points as approximations to the percentage points of the distribution of . These percentage points were tabulated for a number of parameter values of c where = exp(c/n). At these percentage points, we evaluate our LEAD approximation to the distribution function for . We also obtain "exact" evaluation of the distribution function at these percentage points by simulation with one-half of one million replications. A selected but representative set of these evaluations of the distribution function is reported in Table 6. Examining the entries, it is clear that, at every percentage point, the LEAD approximation outperforms the Perron approximation. As in the case of Model B, while the LEAD approximation matches the exact evaluation, almost always to the third decimal place and often to the fourth decimal place, the Perron approximation fails often to match even to the second decimal place except at the right tail of the distribution. Accuracy of the Perron approximation improves as the sample size is increased but it has a questionable accuracy even for a sample of size as large as 50. For example, for c = -10.0, n = 50, = 0.8187, at the percentage point of -11.834, while (see Table 6, Panel A) the exact evaluation of the distribution is 0.0278, the Perron approximation is 0.050.

There are several other approximations in the case of Model C. Phillips (1978), Wang (1992) and Lieberman (1994) have provided approximations to the Pr. For a number of sample sizes of n , a variety of values of , and at a number of percentage points, x, Phillips (1978) tabulates his approximation (wherever it is available) to this probability along with its exact value. We evaluate our LEAD approximation for each of these combinations of n, and x. A selected but representative set of these evaluations is reported in Table 7. Lieberman (1994) approximation matches the LEAD to the fourth decimal place throughout our comparison and hence it is not shown in the table. Note, however that the approximations reported by Lieberman (1994, Table 1) do not match the corresponding LEAD approximations but they match the LEAD approximations to the left-tailed probability, namely the Pr. It is clear from the entries in Table 7 that all four approximations, Phillips (1978, whenever it is available), Wang (1992), Lieberman (1994) and the LEAD are highly accurate and they are all of comparable accuaracy.

4. ILLUSTRATIVE EXAMPLES

Preceding analysis suggests that, for all practical purposes, the distribution of the least squares estimator , G(w) = Pr( < w) can be well approximated by the first term in the expansion (2.7) (or, (2.8)), specifically by

(4.1) AG(w) = F() - f()(), if the saddlepoint u1 ¹ 0

= 0.5 + (2pn)q3, if the saddlepoint u1 = 0

where u1 is the saddlepoint satisfying h(1)(u) = 0, h(u) is as defined preceding the equation (2.7), h1 = h(u1), x = (-h1)(1/2)u1/½u1½, z = u1(h1(2)/2)(1/2), and q3 = h(3)(0)/(6[h(2)(0)](3/2)). The approximation is applicable for all the Models A, B and C and for all possible values of the parameter ß (½ß½ < 1, ½ß½ = 1, or ½ß½ > 1).

The approximation AG(w) can then be used to make inference about the parameter ß. In most applications, it is of interest to know whether ½ß½ < 1, or, ½ß½ = 1, or ½ß½ > 1. Such inference can be made from an appropriate confidence interval estimate for ß. These confidence interval estimates can be obtained using the approximation AG(w). It can be shown that the lower and upper limit of the central 100(1 - g)% confidence interval are, respectively the solutions of

(4.2) AG() = 1 - g/2, and

(4.3) AG() = g/2

where is the least squares estimate of ß. These equations can be solved by any of a variety of successive approximation methods (see Abramowitz and Stegun, 1970, p. 18). A fortran program which can be obtained on request is written to solve these equations using the Newton's Rule. To start the approximation method, both the lower and upper limit can be taken to be the estimate .

Confidence interval is a convenient tool to make inference. However, if there is a specific inference needs to be made, it may be more convenient to make such inference by testing specific hypothesis. Thus, if one is interested to know whether there is a root = 1 in the autoregressive polynomial or not, one may test the hypothesis H0: ß = 1, against the alternatives Ha: ß < 1. We would reject H0 if the p-value given by Pr( < sample ½ ß = 1) is smaller than the level of significance. Using our approximation AG(w) to the distribution of , this p-value can be computed.

To illustrate the use of these results, we have analyzed three sets of data generated from the Model A. In each case, we have chosen n = 25 and se = 1. The first data set is generated from the model with ß = .95 (asymptotically stationary model), the second with ß = 1 (random-walk model) and the third with ß = 1.05 (explosive model). The three data sets, estimate based on these data and the central 95% confidence interval estimates are found as follows.

Data Set 1: 0.86, 1.26, 2.39, 2.60, 2.81, 4.15, 3.36, 1.25, 1.17, 0.16, -0.09, 0.54, -0.57

-2.62, -3.10, -1.30, 0.19, 1.56, 1.60, 1.49, 3.62, 3.96, 3.03, 2.49, 3.64

= 0.930; 95% confidence interval: [0.772, 1.186]

The p-value, in testing H0: ß = 1 against Ha: ß < 1, is 0.371

Data Set 2: 0.86, 1.31, 2.50, 2.82, 3.16, 4.64, 4.06, 2.12, 2.11, 1.15, 0.91, 1.54, 0.45,

-1.62, -2.23, -0.59, 0.83, 2.21, 2.33, 2.30, 4.51, 5.03, 4.29, 3.91, 5.18

= 0.991; 95% confidence interval: [0.874, 1.204]

The p-value, in testing H0: ß = 1 against Ha: ß < 1, is 0.626

Data Set 3: 0.86, 1.35, 2.61, 3.06, 3.56, 5.22, 4.89, 3.20, 3.34, 2.56, 2.44, 3.19, 2.27

0.31, -0.29, 1.33, 2.83, 4.35, 4.68, 4.89, 7.34, 8.23, 7.91, 7.91, 9.58

= 1.066; 95% confidence interval: [0.989, 1.222]

The p-value, in testing H0: ß = 1 against Ha: ß < 1, is 0.965

Based on these confidence intervals and p-values, one would not reject the hypothesis of unit root in all these three cases. However, for data set 2, unit root is located near to the center of this interval; for data set 1, it is close to the upper limit of the interval and for data set 3, it is close to the lower limit of the interval. This evidence may be interpreted to mean that the model generating the data set 1 is likely to have root (the parameter ß) less than 1, that generating the data set 2 is likely to have unit root and that generating the data set 3 to have root larger than 1.

5. CONCLUDING REMARKS

Autoregressive models have been found to be most prominent among models in describing time movement of a time series variable. Inference on the parameters of these models are then of utmost interest to the practitioners. Often, such inference has been based on the least squares estimators. As the exact distributions of these least squares estimators are rarely known, the practitioners have been forced to rely on the known asymptotic distributions for inferential purposes. There are at least two disadvantages to such procedures. First, the asymptotic distributions are either too complicated for practical use or provide impractical poor approximations to exact distributions, especially for sample sizes that are usually available in practice. Second, the form of the asymptotic distributions depend upon the parameter values. More specifically, asymptotic distribution that is appropriate when all the roots of the autoregressive polynomial are less than 1 in magnitude is different from the one that is appropriate if at least one of the roots is equal to 1, and these asymptotic distributions are, in turn different from the one appropriate when some of the roots are greater than 1 in magnitudes. This nonsmooth transformation of the asymptotic distribution from the case when the roots are less than 1 to the case when some roots are equal to 1 to the case when some roots are greater than 1 is, to say the least counterintuitive and it makes it essential prerequisite to have the knowledge of any root of the autoregressive polynomial to be equal to 1 or greater than 1 before one can use the appropriate asymptotic distribution for inference.

This paper considers the autoregressive model of order one. An uniform asymptotic expansion for the distribution of the least squares estimator for the autoregressive coefficient is derived. This is a valid expansion irrespective of the size of the root of the autoregressive polynomial. Unfortunately, this asymptotic expansion is, like most asymptotic expansions nonconvergent. Fortunately, however, it is found, after a detailed investigation that the leading term of the expansion provides excellent approximation to the exact distribution. Thus, this approximation can be used, for all practical purposes to make inference. To implement this approximation, one is required to evaluate a certain characteristic function and its first two derivatives. A convenient closed-form expression for the characteristic function is given which should facilitate such computations.

There are a number of alternative approximations to the distribution of the least squares estimator. Of these approximations, the most promising ones seem to be those given by Perron (1991b), Lieberman (1994) and Larsson (1995) in the case of Model A; Perron (1991b) in the case of Model B; and Phillips (1978), Perron (1991a), Wang (1992) and Lieberman (1994) in the case of Model C. A thorough comparison of accuracy of these approximations with that of our approximation (LEAD), we find that the LEAD is exceptionally accurate, and its accuracy is far superior to those given by Perron (1991b) and Larsson (1995) in the case of Model A, Perron (1991b) in the case of Model B, and Perron (1991a) in the case of Model C; and is as accurate or better than those given by Lieberman (1994) in the case of Model A and Phillips (1978), Wang (1992) and Lieberman (1994) in the case of Model C. In short, in the case of Model B, approximation from Perron (1991b) has questionable accuracy and there is no rival to the LEAD approximation. In the case of Model A, approximations from Perron (1991b) and Larsson (1995) have questionable accuracy, and both the LEAD and the approximation from Lieberman (1994) provide excellent accuracy. Thus, in this case of Model A, there are two excellent approximations. However, on the ground of computational convenience, the LEAD is preferable to that given by Lieberman (1994), especially if the sample size is large. Finally, in the case of Model C, approximation from Perron (1991a) provide questionable accuracy but the LEAD and the remaining three approximations from Phillips (1978), Wang (1992) and Lieberman (1994) provide excellent accuracy. Of the four viable alternative approximations, the approximation from Phillips (1978) is not available for a sizable region in the tail of the distribution where most of the inferential interest lie; the computational burden to implement the approximation from Lieberman (1994) can be prohibitive, especially if the sample size is large; and computationally, the approximation from Wang (1992) is as convenient as the LEAD. Thus, in the case of Model C, the LEAD and the approximation from Wang (1992) are preferable to all the others and there is not much reason to prefer Wang (1992) to the LEAD . In summary, the LEAD provides the most desirable approximation in all the cases of Model A, Model B and Model C and only in the case of Model C, the approximation from Wang (1992) can be as desirable as the LEAD.

Table 1: Exact and Approximations (LEAD, LEAD2, LEAD3) to Pr(g(n)( - ) < x)

for Model A, n = 10

x ß = 1.0 ß = 1.01

EXACT LEAD LEAD2 LEAD3 EXACT LEAD LEAD2 LEAD3

-16.0 0.0000 0.0000 0.0000 0.0000 0.2160 0.2333 0.2169 0.2034

-12.0 0.0001 0.0001 0.0001 0.0001 0.2861 0.3045 0.2930 0.2755

-8.0 0.0047 0.0050 0.0046 0.0047 0.3800 0.3943 0.3906 0.3754

-6.0 0.0208 0.0223 0.0203 0.0211 0.4382 0.4482 0.4480 0.4344

-4.0 0.0730 0.0791 0.0705 0.0739 0.5032 0.5096 0.5121 0.5012

-3.5 0.0975 0.1059 0.0942 0.0975 0.5205 0.5263 0.5293 0.5187

-3.0 0.1293 0.1405 0.1257 0.1265 0.5382 0.5435 0.5471 0.5381

-2.8 0.1445 0.1570 0.1410 0.1399 0.5455 0.5506 0.5543 0.5469

-2.6 0.1614 0.1753 0.1584 0.1545 0.5528 0.5577 0.5616 0.5569

-2.4 0.1802 0.1954 0.1780 0.1710 0.5602 0.5649 0.5689 0.5689

-2.2 0.2011 0.2175 0.2002 0.1897 0.5678 0.5723 0.5763 0.5849

-2.0 0.2243 0.2417 0.2252 0.2115 0.5754 0.5797 0.5838 0.6092

-1.8 0.2501 0.2683 0.2533 0.2370 0.5831 0.5873 0.5870 *

-1.6 0.2789 0.2974 0.2845 0.2669 0.5909 0.5948 0.5956 *

-1.4 0.3111 0.3292 0.3191 0.3013 0.5988 0.6025 0.6035 *

-1.2 0.3471 0.3640 0.3571 0.3401 0.6068 0.6103 0.6101 *

-1.0 0.3876 0.4022 0.3984 0.3828 0.6149 0.6182 0.6131 *

-0.8 0.4328 0.4441 0.4433 0.4292 0.6230 0.6261 0.6031 *

-0.6 0.4826 0.4903 0.4920 0.4812 0.6313 0.6342 0.5265 *

-0.4 0.5361 0.5414 0.5450 0.5357 0.6396 0.6428 * *

-0.2 0.5939 0.5976 0.5989 * 0.6480 0.6558 * *

0.0 0.6566 0.6586 0.6646 0.6550 0.6565 0.6585 0.6645 0.6550

0.2 0.7227 0.7226 0.7304 0.6487 0.6650 0.6624 * *

0.4 0.7883 0.7859 0.7946 0.7874 0.6736 0.6745 * *

0.6 0.8475 0.8433 0.8531 0.8468 0.6823 0.6833 0.7737 *

0.8 0.8952 0.8904 0.8999 0.8944 0.6910 0.6917 0.7169 *

1.0 0.9300 0.9255 0.9336 0.9292 0.6997 0.7002 0.7116 *

1.2 0.9536 0.9500 0.9562 0.9530 0.7085 0.7087 0.7148 0.9275

1.4 0.9691 0.9664 0.9709 0.9687 0.7173 0.7172 0.7191 *

1.6 0.9792 0.9772 0.9804 0.9790 0.7260 0.7256 0.7338 0.6746

1.8 0.9858 0.9843 0.9866 0.9857 0.7348 0.7340 0.7422 0.7135

2.0 0.9902 0.9891 0.9907 0.9901 0.7435 0.7424 0.7508 0.7339

4.0 0.9995 0.9995 0.9995 0.9995 0.8265 0.8224 0.8327 0.8260

g(n) = , if ½ß½ < 1, = , if ½ß½ = 1 and = , if ½ß½ =1; EXACT is the exact probability computed by numerical integration (Tsui, 1989); LEAD, LEAD2 and LEAD3 are, respectively the approximations from the leading one, two and three terms in the expansion (2.7) or (2.8).

* the number is outside the range of (0, 1).

Table 2: Exact and Approximations (LEAD, LEAD2, LEAD3) to Pr(g(n)( - ) < x)

for Model B, n = 10, = 1

x ß = 1.0 ß = 1.01

EXACT LEAD LEAD2 LEAD3 EXACT LEAD LEAD2 LEAD3

-16.0 0.0000 0.0000 0.0000 0.0000 0.1967 0.2130 0.1969 0.1839

-12.0 0.0001 0.0001 0.0001 0.0001 0.2659 0.2839 0.2727 0.2533

-8.0 0.0035 0.0037 0.0034 0.0035 0.3612 0.3754 0.3730 0.3548

-6.0 0.0165 0.0177 0.0162 0.0167 0.4216 0.4311 0.4331 0.4163

-4.0 0.0622 0.0672 0.0601 0.0631 0.4901 0.4956 0.5008 0.4872

-3.5 0.0845 0.0916 0.0816 0.0850 0.5085 0.5133 0.5192 0.5055

-3.0 0.1140 0.1239 0.1106 0.1124 0.5274 0.5317 0.5382 0.5254

-2.8 0.1284 0.1395 0.1249 0.1251 0.5352 0.5392 0.5459 0.5341

-2.6 0.1445 0.1570 0.1413 0.1390 0.5431 0.5469 0.5538 0.5437

-2.4 0.1625 0.1763 0.1599 0.1544 0.5511 0.5546 0.5617 0.5545

-2.2 0.1827 0.1979 0.1812 0.1719 0.5592 0.5625 0.5697 0.5678

-2.0 0.2054 0.2218 0.2056 0.1923 0.5674 0.5705 0.5778 0.5862

-1.8 0.2308 0.2482 0.2333 0.2166 0.5758 0.5787 0.5859 0.6157

-1.6 0.2594 0.2774 0.2646 0.2455 0.5842 0.5869 0.5903 *

-1.4 0.2916 0.3095 0.2998 0.2796 0.5928 0.5953 0.5994 *

-1.2 0.3281 0.3450 0.3387 0.3190 0.6015 0.6037 0.6075 *

-1.0 0.3696 0.3842 0.3817 0.3630 0.6103 0.6123 0.6129 *

-0.8 0.4166 0.4276 0.4287 0.4114 0.6192 0.6210 0.6092 *

-0.6 0.4689 0.4760 0.4801 0.4642 0.6282 0.6298 0.5588 *

-0.4 0.5257 0.5301 0.5364 0.5235 0.6373 0.6390 0.0750 *

-0.2 0.5879 0.5905 0.5943 * 0.6465 0.6522 * *

0.0 0.6561 0.6568 0.6656 0.6542 0.6558 0.6564 0.6655 0.6541

0.2 0.7284 0.7270 0.7371 0.6866 0.6652 0.6620 * *

0.4 0.7995 0.7961 0.8065 0.7990 0.6747 0.6742 * *

0.6 0.8618 0.8570 0.8674 0.8611 0.6842 0.6838 0.7505 *

0.8 0.9094 0.9045 0.9138 0.9087 0.6937 0.6931 0.7149 *

1.0 0.9420 0.9379 0.9451 0.9414 0.7033 0.7024 0.7142 0.2491

1.2 0.9629 0.9597 0.9651 0.9625 0.7130 0.7118 0.7187 *

1.4 0.9760 0.9737 0.9774 0.9757 0.7226 0.7210 0.7318 0.6456

1.6 0.9842 0.9826 0.9852 0.9841 0.7322 0.7304 0.7410 0.7039

1.8 0.9894 0.9883 0.9900 0.9893 0.7418 0.7396 0.7503 0.7301

2.0 0.9928 0.9920 0.9932 0.9927 0.7514 0.7489 0.7596 0.7464

4.0 0.9997 0.9996 0.9997 0.9997 0.8405 0.8355 0.8470 0.8397

g(n) = , if ½ß½ < 1, = , if ½ß½ = 1 and = , if ½ß½ =1; EXACT is the exact probability computed by numerical integration (Tsui, 1989); LEAD, LEAD2 and LEAD3 are, respectively the approximations from the leading one, two and three terms in the expansion (2.7) or (2.8).

* the number is outside the range of (0, 1).

Table 3: Exact and Approximations (LEAD, LEAD2, LEAD3) to Pr( )

for Model C, n = 10

x ß = 0.95 ß = 0.99

EXACT LEAD LEAD2 LEAD3 EXACT LEAD LEAD2 LEAD3

-16.0 0.0001 0.0001 0.0001 0.0001 0.0086 0.0094 0.0083 0.0088

-12.0 0.0012 0.0013 0.0012 0.0012 0.0209 0.0232 0.0199 0.0215

-8.0 0.0123 0.0131 0.0120 0.0124 0.0505 0.0566 0.0481 0.0495

-6.0 0.0331 0.0357 0.0322 0.0336 0.0803 0.0899 0.0790 0.0744

-4.0 0.0852 0.0923 0.0824 0.0854 0.1332 0.1450 0.1376 0.1261

-3.5 0.1077 0.1168 0.1046 0.1062 0.1530 0.1642 0.1592 0.1469

-3.0 0.1365 0.1479 0.1336 0.1313 0.1771 0.1868 0.1848 0.1720

-2.8 0.1502 0.1625 0.1477 0.1430 0.1882 0.1971 0.1965 0.1835

-2.6 0.1654 0.1787 0.1637 0.1560 0.2003 0.2082 0.2092 0.1961

-2.4 0.1822 0.1964 0.1817 0.1708 0.2136 0.2203 0.2231 0.2099

-2.2 0.2009 0.2159 0.2019 0.1877 0.2283 0.2335 0.2382 0.2252

-2.0 0.2218 0.2373 0.2248 0.2075 0.2446 0.2481 0.2550 0.2423

-1.8 0.2451 0.2609 0.2505 0.2306 0.2627 0.2643 0.2736 0.2616

-1.6 0.2713 0.2868 0.2793 0.2575 0.2832 0.2825 0.2945 0.2835

-1.4 0.3008 0.3154 0.3114 0.2884 0.3063 0.3031 0.3181 0.3086

-1.2 0.3341 0.3470 0.3470 0.3235 0.3328 0.3266 0.3449 0.3376

-1.0 0.3719 0.3820 0.3866 0.3628 0.3634 0.3539 0.3757 0.3712

-0.8 0.4146 0.4209 0.4304 0.4064 0.3992 0.3855 0.4113 0.4104

-0.6 0.4627 0.4644 0.4790 0.4554 0.4408 0.4226 0.4523 0.4564

-0.4 0.5165 0.5131 0.5328 0.5110 0.4891 0.4659 0.4994 0.5177

-0.2 0.5753 0.5673 0.5892 0.9894 0.5439 0.5157 0.5467 *

0.0 0.6385 0.6267 0.6553 0.6353 0.6037 0.5709 0.6117 0.6289

0.2 0.7041 0.6896 0.7207 0.6758 0.6654 0.6294 0.6773 0.0854

0.4 0.7683 0.7526 0.7836 0.7581 0.7246 0.6874 0.7321 0.7474

0.6 0.8261 0.8109 0.8395 0.8169 0.7776 0.7411 0.7858 0.8027

0.8 0.8738 0.8605 0.8846 0.8658 0.8222 0.7881 0.8316 0.8443

1.0 0.9101 0.8995 0.9185 0.9039 0.8583 0.8277 0.8687 0.8760

1.2 0.9365 0.9283 0.9426 0.9322 0.8868 0.8601 0.8977 0.8997

1.4 0.9551 0.9490 0.9594 0.9523 0.9092 0.8864 0.9199 0.9174

1.6 0.9680 0.9635 0.9710 0.9663 0.9266 0.9075 0.9367 0.9311

1.8 0.9770 0.9737 0.9790 0.9759 0.9404 0.9244 0.9495 0.9420

2.0 0.9832 0.9809 0.9847 0.9826 0.9512 0.9380 0.9592 0.9510

4.0 0.9988 0.9987 0.9989 0.9988 0.9912 0.9891 0.9924 0.9905

EXACT is exact the probability computed by numerical integration (Tsui, 1989); LEAD, LEAD2 and LEAD3 are, respectively the approximations from the leading one, two and three terms in the expansion (2.7) or (2.8).

* the number is outside the range of (0, 1).

Table 4: Exact and Approximations (PERRON, LARSSON, LEAD) to

for Model A

Panel A (c = -5.0)

n = 10, = .6065 n = 25, = .8187 n = 50, = .9048

x PERRON LARSSON EXACT LEAD EXACT LEAD EXACT LEAD

-16.7378 0.010 0.0110 0.0000 0.0000 0.0020 0.0020 0.0050 0.0051

-13.1743 0.025 0.0282 0.0004 0.0004 0.0083 0.0088 0.0156 0.0160

-10.4633 0.050 0.0580 0.0039 0.0038 0.0238 0.0250 0.0358 0.0369

-7.7242 0.100 0.1216 0.0242 0.0246 0.0627 0.0655 0.0808 0.0831

2.4561 0.900 NA 0.9127 0.9137 0.9068 0.9095 0.9038 0.9068

3.0637 0.950 NA 0.9573 0.9561 0.9551 0.9556 0.9532 0.9541

3.5239 0.975 NA 0.9761 0.9998 0.9773 0.9768 0.9766 0.9765

4.0342 0.990 NA 0.9869 0.9862 0.9899 0.9893 0.9902 0.9898

Panel B (c = 0.0)

n = 10, = 1.0 n = 25, = 1.0 n = 50, = 1.0

x PERRON LARSSON EXACT LEAD EXACT LEAD EXACT LEAD

-13.6919 0.010 0.0115 0.0011 0.0011 0.0049 0.0054 0.0076 0.0078

-10.4399 0.025 0.0297 0.0079 0.0082 0.0169 0.0181 0.0210 0.0221

-8.0383 0.050 0.0616 0.0262 0.0276 0.0392 0.0424 0.0445 0.0478

-5.7133 0.100 0.1307 0.0724 0.0773 0.0871 0.0952 0.0941 0.1014

0.9280 0.900 0.8505 0.8618 0.8578 0.8850 0.8825 0.8931 0.8903

1.2854 0.950 0.9379 0.9151 0.9110 0.9376 0.9340 0.9442 0.9409

1.6122 0.975 0.9711 0.9473 0.9436 0.9656 0.9628 0.9705 0.9682

2.0325 0.990 0.9889 0.9710 0.9687 0.9844 0.9825 0.9875 0.9860

Panel C (c = 2.0)

n = 10, = 1.221 n = 25, = 1.083 n = 50, = 1.041

x PERRON LARSSON EXACT LEAD EXACT LEAD EXACT LEAD

-12.0557 0.010 0.0125 0.0040 0.0043 0.0071 0.0079 0.0085 0.0093

-8.8762 0.025 0.0351 0.0174 0.0193 0.0214 0.0242 0.0234 0.0260

-6.5587 0.050 0.0890 0.0438 0.0501 0.0470 0.0537 0.0489 0.0551

-4.3431 0.100 0.9664 0.1002 0.1117 0.0988 0.1116 0.1003 0.1119

0.5191 0.900 0.8376 0.8445 0.8219 0.8788 0.8576 0.8901 0.8692

0.7585 0.950 0.9332 0.9034 0.8860 0.9334 0.9192 0.9425 0.9292

1.0067 0.975 0.9694 0.9398 0.9282 0.9631 0.9550 0.9699 0.9624

1.3550 0.990 0.9884 0.9675 0.9611 0.9831 0.9793 0.9870 0.9839

c is a reparameterization of (see Perron, 1991b) with = exp(-c/n); EXACT is the exact probability computed by simulation with one-half of one million replications; PERRON, LARSSON and LEAD are, respectively the approximations from Perron (1991b), Larsson (1995) and the leading term in the expansion (2.7) or (2.8).

NA: not available

Table 5: Exact and Approximations (PERRON, LEAD) to

for Model B

Panel A (c = -5.0, = 0.5)

n = 10, = .6065, n = 25, = .8187, n = 50, = .9048,

= 1.58 = 2.50 = 3.54

x PERRON EXACT LEAD EXACT LEAD EXACT LEAD

-13.8955 0.010 0.0000 0.0000 0.0021 0.0021 0.0050 0.0052

-10.9832 0.025 0.0003 0.0003 0.0090 0.0091 0.0160 0.0161

-8.7550 0.050 0.0039 0.0039 0.0253 0.0253 0.0369 0.0371

-6.4899 0.100 0.0250 0.0251 0.0656 0.0660 0.0825 0.0832

2.2695 0.900 0.9244 0.9249 0.9106 0.9129 0.9061 0.9082

2.8257 0.950 0.9684 0.9677 0.9591 0.9598 0.9547 0.9561

3.2413 0.975 0.9855 0.9846 0.9808 0.9807 0.9779 0.9786

3.6788 0.990 0.9937 0.9933 0.9924 0.9921 0.9913 0.9913

Panel B (c = 0.0, = 0.5)

n = 10, = 1.00, n = 25, = 1.00, n = 50, = 1.00,

= 1.58 = 2.50 = 3.54

x PERRON EXACT LEAD EXACT LEAD EXACT LEAD

-10.9570 0.010 0.0011 0.0011 0.0056 0.0055 0.0076 0.0078

-8.3535 0.025 0.0081 0.0083 0.0177 0.0183 0.0213 0.0222

-6.4315 0.050 0.0264 0.0278 0.0405 0.0426 0.0456 0.0479

-4.5709 0.100 0.0731 0.0755 0.0898 0.0954 0.0951 0.1015

0.7643 0.900 0.8774 0.8717 0.8904 0.8870 0.8951 0.8920

1.0481 0.950 0.9303 0.9261 0.9423 0.9392 0.9460 0.9433

1.3042 0.975 0.9607 0.9577 0.9698 0.9673 0.9722 0.9703

1.6338 0.990 0.9814 0.9800 0.9872 0.9857 0.9884 0.9875

Panel C (c = 2.0, = 0.5)

n = 10, = 1.2214, n = 25, = 1.0833, n = 50, = 1.0408,

= 1.58 = 2.50 = 3.54

x PERRON EXACT LEAD EXACT LEAD EXACT LEAD

-9.2554 0.010 0.0040 0.0043 0.0075 0.0080 0.0088 0.0095

-6.7080 0.025 0.0174 0.0194 0.0220 0.0246 0.0236 0.0265

-4.8444 0.050 0.0438 0.0496 0.0477 0.0540 0.0488 0.0556

-3.0206 0.100 0.0986 0.1038 0.1001 0.1068 0.0999 0.1078

0.3756 0.900 0.8646 0.8438 0.8859 0.8663 0.8929 0.8737

0.5401 0.950 0.9232 0.9091 0.9402 0.9271 0.9449 0.9327

0.7143 0.975 0.9579 0.9490 0.9689 0.9612 0.9717 0.9649

0.9680 0.990 0.9811 0.9771 0.9869 0.9835 0.9881 0.9855

c and are reparameterizations of and = y0/ (see Perron, 1991b) with = exp(-c/n) and = n1/2; EXACT is the exact probability computed by simulation with one-half of one million replications; PERRON and LEAD are, respectively the approximations from Perron (1991b), and the leading term in the expansion (2.7) or (2.8).

Table 6: Exact and Approximations (PERRON, LEAD) to

for Model C

Panel A (c = -10.0)

n = 10, = 0.3679 n = 25, = 0.6703 n = 50, = 0.8187

x PERRON EXACT LEAD EXACT LEAD EXACT LEAD

-18.478 0.010 0.0000 0.0000 0.0006 0.0005 0.0032 0.0032

-14.728 0.025 0.0000 0.0000 0.0037 0.0037 0.0111 0.0114

-11.834 0.050 0.0002 0.0002 0.0129 0.0136 0.0278 0.0286

-8.846 0.100 0.0045 0.0044 0.0426 0.0441 0.0688 0.0698

3.793 0.900 0.9553 0.9563 0.9267 0.9278 0.9143 0.9154

4.729 0.950 0.9861 0.9861 0.9704 0.9709 0.9613 0.9621

5.436 0.975 0.9947 0.9945 0.9879 0.9882 0.9828 0.9896

6.154 0.990 0.9979 0.9977 0.9961 0.9961 0.9939 0.9965

Panel B (c = -1.0)

n = 10, = 0.9048 n = 25, = 0.9608 n = 50, = 0.9802

x PERRON EXACT LEAD EXACT LEAD EXACT LEAD

-11.848 0.010 0.0009 0.0009 0.0049 0.0051 0.0073 0.0075

-8.942 0.025 0.0066 0.0068 0.0161 0.0169 0.0203 0.0213

-6.831 0.050 0.0222 0.0229 0.0375 0.0395 0.0434 0.0459

-4.818 0.100 0.0634 0.0657 0.0842 0.0889 0.0922 0.0970

0.971 0.900 0.8864 0.8806 0.8940 0.8892 0.8976 0.8921

1.322 0.950 0.9357 0.9327 0.9447 0.9411 0.9475 0.9439

1.639 0.975 0.9628 0.9609 0.9709 0.9683 0.9730 0.9707

2.042 0.990 0.9814 0.9803 0.9873 0.9858 0.9888 0.9875

Panel C (c = -0.1)

n = 10, = 0.9900 n = 25, = 0.9960 n = 50, = 0.9980

x PERRON EXACT LEAD EXACT LEAD EXACT LEAD

-8.168 0.010 0.0037 0.0041 0.0074 0.0077 0.0086 0.0092

-5.634 0.025 0.0154 0.0165 0.0214 0.0227 0.0234 0.0249

-3.906 0.050 0.0380 0.0416 0.0459 0.0496 0.0481 0.0523

-2.405 0.100 0.0880 0.0957 0.0959 0.1042 0.0978 0.1071

0.495 0.900 0.8864 0.8616 0.8942 0.8685 0.8981 0.8708

0.727 0.950 0.9377 0.9233 0.9446 0.9302 0.9474 0.9325

0.971 0.975 0.9654 0.9576 0.9713 0.9636 0.9730 0.9655

1.312 0.990 0.9837 0.9802 0.9877 0.9846 0.9886 0.9859

c is a reparameterization of (see Perron, 1991a) with = exp(-c/n); EXACT is the exact probability computed by simulation with one-half of one million replications; PERRON and LEAD are, respectively the approximations from Perron (1991a), and the leading term in the expansion (2.7) or (2.8).

Table 7: Exact and Approximations (PHILLIPS, WANG, LEAD) to

for Model C, n = 10

= 0.2 = 0.8

x EXACT PHILLIPS WANG LEAD EXACT PHILLIPS WANG LEAD

1.0 0.2987 0.2926 0.2987 0.2988 0.3444 NA 0.3582 0.3572

1.4 0.1400 0.1283 0.1400 0.1400 0.2280 NA 0.2371 0.2372

1.8 0.0551 0.0274 0.0551 0.0550 0.1565 NA 0.1628 0.1629

2.2 0.0191 NA 0.0189 0.0190 0.1083 NA 0.1126 0.1126

2.6 0.0060 NA 0.0057 0.0056 0.0745 NA 0.0775 0.0774

3.0 0.0018 NA 0.0015 0.0015 0.0507 NA 0.0522 0.0525

EXACT is the exact probability as reported in Phillips (1978); PHILLIPS is the saddlepoint approximation given in Phillips (1978); WANG is the approximation given in Wang (1992); LEAD is the approximation from the leading term of the expansion (2.7) or (2.8).

NA: not available; the approximate density is not defined.

REFERENCES

Abadir, K. M. (1993). "The Limiting Distribution of the Autocorrelation Coefficient Under a Unit Root", The Annals of Statistics, 21, no. 2, 1058-70.

Abramowitz, M. and I. A. Stegun (1970). Handbook of Mathematical Functions, Dover edition, ninth printing, New York: Dover Publications, Inc.

Cavanagh, C. (1986). "Roots Local to Unity", mimeo, Harvard University.

Cornish, E. A. and Fisher, R. A. (1937). "Moments and Cumulants in the Specification of Distribution', Revue de l'institut Internat. de Statist., 4, 307-20.

Cryer, J. D., J. C. Nankervis and N. E. Savin (1989). "Mirror-Image Distributions in AR(1) Models", Econometric Theory, 5, 36-52.

Dickey, D. A. (1976). Estimation and Hypothesis Testing for Nonstationary Time Series, Unpublished Ph. D. Dissertation, Iowa State University, Ames, IW.

Dickey, D. A. and W. A. Fuller (1979). "Distribution of the Estimators for Autoregressive Time Series with a Unit Root", J. Amer. Statist. Assn., 74, 427-31.

Dickey, D. A. and W. A. Fuller (1981). "Likelihood Ratio Test for Autoregressive Time Series with a Unit Root", Econometrica, 49, 1057-1072.

Diebold, F. X. and M. Nerlove (1990). "Unit Roots in Economic Time Series: A Selective Survey", Advances in Econometrics, 8, 3-69.

Evans, G. B. and N. E. Savin (1981). "Testing for Unit Roots: I", Econometrica, 49, 753-79.

Evans, G. B. and N. E. Savin (1984). "Testing for Unit Roots: II", Econometrica, 52, 1241-69.

Fisher, R. A. and E. A. Cornish (1960). "The Percentile Points of Distributions Having Unknown Cumulants", Technometrics, 2, 209-26.

Fuller, W. A. (1976). Introduction to Statistical Time Series, New York: Wiley

Gurland, J. (1948). "Inversion Formula for the Distribution of Ratios", Annals of Mathematical Statistics, 19, 228-37.

Hasza, D. P. and W. A. Fuller (1979). "Estimation for Autoregressive Processes with Unit Roots", Annals of Statistics, 7, 1106-20.

Hill, G. W. and A. W. Davis (1968). "Generalized Asymptotic Expansions of Cornish-Fisher Type", Annals of Mathematical Statistics, 39, 1264-73.

Hurwicz, L. (1950). "Least Squares Bias in Time Series", in Statistical Inference in Dynamic Models, ed. by T. C. Koopmans, New York: Wiley, 365-83.

Larsson, R. (1995). "The Asymptotic Distributions of Some Test Statistics in Near-Integrated AR Processes", Econometric Theory, 11, 306-30.

Lieberman, O. (1994). "Saddlepoint Approximation for the Least Squares Estimator in First-Order Autoregression", Biometrika, 81, no. 4, 807-11.

Lugannani, R. and S. O. Rice (1980). "Saddle Point Approximation for the Distribution of the Sum of Independent Random Variables", Adv. Appl. Prob., 12, 475-90.

Mann, H. B. and A. Wald (1943). "On the Statistical Treatment of Linear Stochastic Difference Equations", Econometrica, 11, 173-220.

Nabeya, S. and K. Tanaka (1990). "A General Approach to the Limiting Distribution for Estimators in Time Series Regression with Nonstable Autoregressive Errors", Econometrica, 58, 145-63.

Perron, P. (1989). "The Calculation of the Limiting Distribution of the Least Squares Estimator in a Near-Integrated Model", Econometric Theory, 5, 241-55.

Perron, P. (1991a). "A Continuous Time Approximation to the Stationary First-Order Autoregressive Model", Econometric Theory, 5, 241-55.

Perron, P. (1991b). "A Continuous Time Approximation to the Unstable First-Order Autoregressive Process: The Case Without an Intercept", Econometrica, 59, no. 1, 211-36.

Perron, P. and P. C. B. Phillips (1987). "Does GNP have a Unit Root? A Reevaluation", Economic Letters, 23, 139-45.

Phillips, P. C. B. (1977). "Approximations to Some Finite Sample Distributions Associated with the First Order Stochastic Difference Equation", Econometrica, 45, 463-85.

Phillips, P. C. B. (1978). "Edgeworth and Saddlepoint Approxiamtions in a First Order Noncircular Autoregression", Biometrika, 59, 79-84.

Phillips, P. C. B. (1988). "Regression Theory for Near-Integrated Time Series", Econometrica, 56, 1021-1044.

Phillips, P. C. B. and P. Perron (1988). "Testing for a Unit Root in Time Series Regression",Biometrika, 75, 335-46.

Rao, M. M. (1978). "Asymptotic Distribution of an Estimator of the Boundary Parameter of an Unstable Process", Annals of Statistics, 16, 185-90.

Rice, S. O. (1968). "Uniform Asymptotic Expansions for Saddle Point Integrals - Applications to a Probability Distribution Occurring in Noise Theory", Bell Syst. Tech. J., 47, 1971-2013.

Satchell, S. E. (1984). "Approximations to the Finite Sample Distributions for Non-Stable First Order Difference Equations", Econometrica, 52, 1271-88.

Schwert, G. W. (1987). "Effects of Model Specification on Tests of Unit Roots in Macroeconomic Data", J. Monetary Econ., 20, 73-103.

Stock, J. H. and W. W. Watson (1989). "Interpreting the Evidence on Money-Income Causality", J. of Econometrics, 40, 161-81.

Tsui, A. K. (1989). On the Finite Sample Distribution of a Least Squares Estimator in a First Order Autoregressive Model, Unpublished Ph. D. Dissertation, University of Kentucky, Lexington, Kentucky.

Tsui, A. K. and M. M. Ali (1992). "Approximations to the Distribution of the Least SquaresEstimatorin a First Order Stationary Autoregressive Model", Communications in Statistics - Simulation, 21, no. 2, 463-84.

Tsui, A. K. and M. M. Ali (1994). "Exact Distributions, Density Functions and Moments of the Least Squares Estimator in a First-Order Autoregressive Model", Computational Statistics & Data Analysis, 17, 433-54.

White, J. S. (1958). "The Limiting Distribution of the Serial Correlation Coefficient in the Explosive Case", Annals of Mathematical Statistics, 29, 1188-97.

White, J. S. (1961). "Asymptotic Expansions for the Mean and Variance of the Serial Correlation Coefficient", Biometrika, 48, 85-95.