dp00 15

Download as pdf or txt
Download as pdf or txt
You are on page 1of 20

A robust data-driven version of the Berlin Method

Yuanhua Feng and Siegfried Heiler

University of Konstanz
Abstract
In this paper a robust data-driven procedure for decomposing seasonal time series based on a generalized Berlin Method (BV, Berliner Verfahren) as proposed by Heiler and Michels (1994) is discussed. The basic robust algorithm used here is an adaptation of the LOWESS (LOcally Weighted Scatterplot Smoothing) procedure (Cleveland, 1979). For selecting the optimal bandwidth the simple double smoothing rule (Heiler and Feng, 1999) is used. The optimal order of the local polynomial is selected with a BIC criterion. The proposed procedure is applied to the macroeconomic time series used in the recent empirical studies carried out by the German Federal Statistical O ce (Speth, 1994 and Hopfner, 1998).

1 Introduction
Decomposing seasonal time series into unobservable trend-cyclical and seasonal components has a long tradition. It is a very important issue of econometrics and provides adjusted data for a prospective analysis (e.g. for a current business cycle analysis). There exists a large number of di erent methodical approaches and also ready-made software systems to perform this. See Heiler (1995) for a survey of methods in this eld from the beginning of the 1960s. See also Eurostat (1998). Heiler (1966, 1970) developed a decomposition procedure based on local regression with polynomials and trigonometric
This paper summarizes a part of the dissertation of Dr. Y. Feng: \Kernel- and Locally Weighted Regression { with applications to time series decomposition", Berlin, 1999, which awarded the Forderpreis of the Gerhard-Furst-Preis of the German Federal Statistical O ce, 1999. The original work was supported in part by a Doctoral-Fellowship of the State Baden-Wuttemberg, Germany and by the SFB 178, University of Konstanz, Germany.

functions as local regressors. This idea became the basis of the so-called Berlin Method (BV, Berliner Verfahren), which in its fourth version (BV 4) is being applied by the German Federal Statistical O ce since 1983. Comparisons among di erent approaches for time series decomposition including the BV 4 with empirical macroeconomic data were carried out by the German Federal Statistical O ce (Speth, 1994 and Hopfner, 1998). The traditional idea of local regression was generalized by Stone (1977) and Cleveland (1979) to a so-called locally weighted regression (LWR), which became the most attractive nonparametric approach in recent years (see the monograph of Fan and Gijbels, 1996). This approach together with other new developments in the area of nonparametric statistics and in the area of computer science since the eighties allows us to propose an improved version of the BV. Heiler and Michels (1994) generalized the BV based on LWR by introducing a kernel as weight function. Like with other nonparametric approaches, e ective use of the generalized BV requires the choice of some smoothing parameters, such as the order of the polynomial and the bandwidth. Hence, the development of some data-driven procedures for carrying out the generalized BV automatically is a crucial theoretical and practical problem. The rst data-driven version of the BV was developed by Heiler and Feng (1996, 1999). In this procedure techniques to deal with outliers in the data are not considered yet. However, the locally weighted regression estimator is susceptible to outliers among the data due to the fact that at a point t only a part of the observations is used in the local smoothing. The proposal of Cleveland (1979) is a robust approach, which works well for common nonparametric regression. In this paper we adapt at rst the idea of Cleveland (1979) to the decomposition of seasonal time series with given smoothing parameters. Then we develop a robust data-driven procedure for time series decomposition using the simple double smoothing (DS) rule for the bandwidth selection (Feng and Heiler, 1999) and the BIC (Bayesian information criterion) for the choice of the polynomial degree (Feng, 1999). The proposed robust data-driven procedure is applied to the macroeconomic time series used by the recent empirical studies carried out by the German Federal Statistical O ce to show its usefulness in practice.

2 The estimators
2.1 Generalized Berlin Method
The BV is based on local least squares. In BV 4 locally weighted least squares are introduced with a xed weighting function (Nourney, 1983). This approach is generalized by Heiler and Michels (1994) based on LWR. The basic idea is as follows. Let Yt, t = 1 ::: n, be an equidistant time series. Assume that (possibly after some transformation of the original data) Yt follows an additive components model

Yt = G(t) + S (t) +

t = 1 2 ::: n

(1)

where the t are assumed to be i.i.d. random variables with E ( t ) = 0 and var( t ) = 2 . G is the trend-cyclical component, S is the seasonal component with seasonal period s and m := G + S is the mean function. The trend-cyclical component is assumed to have some smoothness properties, precisely, to be at least (p+1) times di erentiable, so that it can be expanded in a Taylor series around a point t0, yielding a local polynomial representation of order p p : X j G(t) = 1j (t0 )(t ; t0 ) :
j =0

In a similar way for the seasonal component it is assumed that it can be locally modeled by a Fourier series q X : S (t) = 2j (t0 ) cos j (t ; t0 ) + 3j (t0 ) sin j (t ; t0 )]
j =1

where q = s=2] with x] = largest integer x, j = j 1 , for j = 2 ::: q . Put


1 (t0 )

= 2 =s is the seasonal frequency and

= ( 10(t0 ) :::

1p (t0 ))

2 (t0 )

= ( 21(t0 ) (t0) = ( 1(t0 )T

31 (t0 ) ::: 2q (t0 ) T T 2 (t0 ) )

3q (t0 )])

and X = (X1...X2) with the rows x(t)T, where X is the n (p + s)-regressor matrix. The last terms of 2 and x2 in ], respectively, are only necessary for odd s, for even s they have to be omitted. Let K (u), the weighting function of LWR, be a second order kernel with compact support -1, 1], h 2 be a bandwidth of the observation time, y = (Y1 ::: Yn)T denote the data vector and K = diag(kt) be a weight matrix with 8 < K ( t;t0 ) t 2 t0 ; h t0 + h] kt = : h+0:5 0 otherwise:
I N

= (1 (t ; t0) ::: (t ; t0 )p)T x2 (t) = (cos 1 (t ; t0 ) sin 1 (t ; t0 ) ::: cos q (t ; t0) sin q (t ; t0 )])T x(t) = (x1 (t)T x2 (t)T )T
x1 (t)

Let ej denote the j th (p + 1) 1 unit vector, its odd entries and 0 elsewhere.

denote an (s ; 1) 1 vector having 1 in

The locally weighted least squares criterion n X Yt ; x1 (t)T 1(t0 ) ; x2(t)T 2 (t0)]2 kt ) min leads to the solutions and
t=1

^(t0) = (X(t0 )TK(t0)X(t0));1X(t0 )TK(t0)y

(2) (3) (4) (5)

m ^ (t0) = (eT 1 ^ (t0) = (eT G 1


^(t0) = (0T S

T )(XT KX);1 XT Ky =: wT y s Ty 0T )(XTKX);1 XT Ky =: w1 T )(XT KX);1 XT Ky =: wTy s 2

and

where 0 denotes a vector of zeros of appropriate dimension. ^ and S ^ are all linear smoothers and only observations with t 2 t0 ; h t0 + h] m ^, G obtain non-zero weights. The non-zero parts of w, w1 and w2 will be called the weighting systems. It can be shown that w in (3) satis es: 1: wi(t) = 0 8 <1 n P 2: wi(t)((i ; t)=n)j = : i=1 0 8 P n > < =1 wi (t) cos( j (i ; t)) = 1 0 2 : > iP n : wi (t) sin( j (i ; t)) = 0 i=1 0 ~ ;1 w ~K ~ = min! 3: w if ji ; tj > b j=0 1 j<k

j = 1 :::
~ with respect to w

(6)

~ is the non-zero part of K and w ~ is the non-zero part of w. On the other hand, where K weighting system satis es (6) is the solution of (3). Properties 2 and 2' in (6) ensure that the proposed estimators are unbiased for a polynomial trend-cyclical component and for an exactly periodic seasonal component. For estimations at another instant in the central part of the time series (t0 2 h + 1 n ; h]) we obtain the same weighting systems (i.e. the procedure works like a moving average in the central part of the time series). As in local polynomial tting the th derivative of the trend-cyclical component, 0 < p, can be estimated by ^ ( ) (t0 ) = !(eT+1 0T)(XTKX);1XTKy G =: (w )Ty: (7)

Finite sample and asymptotic properties of these estimators may be found in Feng (1999).

2.2 Approach at the boundary


A point t 2 1 h] n;h+1 n] is called a boundary point. At such a point the observations introduced in the estimate are not symmetric around t. The quality of the estimation at a boundary point is hence usually worse than that at a point in the central part of the 5

time series. This is the so-called boundary problem. In order to cope with this problem we distinguish a left bandwidth hl and a right bandwidth hr and put h = max(hl hr). It is assumed that hl < t, hr n ; t and hT > p + s, where hT = hl + hr + 1 is called the total bandwidth. If hl and hr are xed and t 2 hl + 1 n ; hr], then the estimators m ^, ^, G ^ ( ) and S ^ work as moving averages with symmetric weighting systems for hl = hr G or asymmetric weighting systems for hl 6= hr. In the proposal of Heiler and Feng (1996, 1999) both, hT and p are allowed to change from point to point at the boundary in order to obtain optimal decomposition results. However, the estimations obtained by this procedure are sometimes not stable at the boundary. This problem is solved here by putting hT and p xed, i.e. just one optimal total bandwidth ^ hT together with one optimal order of polynomial p ^ will be selected for the whole time series. In this case the proposed estimators are like k-NN estimators. The use of k-NN estimators to solve the boundary problem was proposed by Gasser et al. (1985). A k-NN estimator is an estimator with a xed total bandwidth hT = hl + hr + 1, which is kept constant at the boundary as well as in the central part. The left bandwidth hl and the right bandwidth hr depend on t. Here only odd integers hT 2 p + s + 1 n] will be considered as possible total bandwidth such that the weighting system in the central part is symmetric. Let hm = (hT ; 1)=2. Then the left and right bandwidths at each point t are determined by

hl = t ; 1 hr = hT ; t hl = hm hr = hm hl = hT ; (n ; t) ; 1 hr = n ; t

if t hm if hm < t n ; hm otherwise:

(8)

A nonrobust data-driven procedure for choosing hm and p for the k-NN estimator will be introduced in the next section.

3 Nonrobust data-driven procedure


From here on only estimation of m, G and S will be considered. The optimal choice of the parameters hT and p is the one, which minimizes a given error criterion as a distance 6

measure between the estimate and the (unknown) underlying function. In this paper m is considered as the target function, i.e. m ^ will be optimized with respect to m. As error criterion for selecting the bandwidth we use the mean averaged squared error (MASE) n X M = n;1 E m ^ t ; mt ]2 : (9)
t=1

It is well known that the MASE splits up into a variance part V and a bias part B , i.e. M = V + B . Under the assumption that t are iid random variables the variance part of a k-NN estimator is n n ^ = ^ 2n;1 X X(wi(t))2 V
2 where ^ 2 is an estimator of 2 . In this paper the bootstrap variance estimator ^ 2 := ^B 2 is de ned by as proposed by Heiler and Feng (1999) will be used. ^B n 2 = n;1 X r 2 ^B i
i=1 t=1 i=1

where the ri's are the residuals of a data-driven pilot smoothing. This estimator is just the averaged residual sum of squares of a pilot estimate. A simple estimate of M was proposed by Rice (1984): n n X ~ = n;1 X(m R ^ t ; yt)2 + (n;1 2wt(t) ; 1)^ 2 : (10)
t=1 t=1

This idea will be used as a pilot method for selecting the parameters. However, the pilot ^ := max(R ~V ^ ), called the R-statistic, due to estimate of M in our program is actually R ^ depends on the couple fhT pg. The optimal bandwidth selected the fact that M V . R ^ T p, is the one among all possible hT's, which by the R-statistic for a given p, denoted by h ^ (p). minimizes R In Heiler and Feng (1996) p is also selected with the error criterion itself. In this paper we propose selecting p by means of the information criteria BIC (Schwarz, 1978 and Akaike, 1979). However, the optimal polynomial order will only be chosen from p = 0 1 ::: 4, because it should not be too high. Theoretically, an odd p is more preferable (see Fan and Gijbels, 1995, 1996). But m ^ with p even performs sometimes the best for nite sample. The original BIC was proposed for model selection in a parametric case, where BIC is 7

a consistent criterion for model choice. In the current case we will use the following de nition of BIC: ^(p)) + ln(n)(p + 1)=n: BIC(p) = ln(R (11) The term ln(n)(p + 1)=n is a penalty for increasing the polynomial order. The BIC ^T R p depends on the couple fhT pg, too. The optimal choice is the couple fh ^g, which minimizes the BIC. For a given p the optimal bandwidth selected by BIC is the same as ^T R p ^(p) itself. But the nal selected optimal couple fh that selected by R ^g by BIC may be di erent from that directly selected by the R-statistic. ^T R p The procedure to search fh ^g for a k-NN estimator with xed p is much simpler than that proposed in Heiler and Feng (1996). Let hmax = n, if n is odd, or hmax = n ; 1, if n is even. For p = 0 1 ::: 4, let hmin = p + s + 2, if p + s is odd, or hmin = p + s + 3, ^ T p which minimizes BIC(p). Select the couple fh ^T R p if p + s is even. Search h ^g with ^ T p pg. fh ^T R p the smallest value of BIC from all fh ^g are then the optimal parameters selection by BIC. The polynomial order p will only be selected once by BIC means the R-criterion. In the following we will discuss how to select a more e ective bandwidth by the double smoothing rule. From now on it is assumed that we have selected a p ^. For the DS rule one also needs a pilot estimate. As shown in Heiler and Feng (1996), the polynomial order in the pilot smoothing should be higher than that in the main smoothing. Hence, p ^p = p ^ + 2 will be used in the pilot smoothing. The DS estimate of MASE is then de ned by ^D = V ^ +B ^D M (12) where ^D = n;1 XfX wi(t)m B ^pi ;m ^ p tg2
n n t=1 i=1

^T g and where m ^ p is the pilot estimate of m obtained with pp = p + 2 and bandwidth h ^ T D is the one among all possible selected by the R-statistic. The optimal bandwidth h ^ D . A simple nonrobust data-driven procedure is as total bandwidths, which minimizes M follows:
2 following the proposal in Heiler and Feng (1999). 1. Estimate the variance with ^B

2. 2. Select an optimal order of polynomial p ^ following the BIC with ^B 2 and p = p ^ p following the R-statistic with ^B 3. Select an optimal total bandwidth h ^+2. p Calculate the pilot estimate m ^ p. 2, p ^ T following the DS criterion with ^B 4. Select an optimal total bandwidth h ^ and m ^ p. ^ T. Calculate all estimators one needs with h

This is a simpli ed procedure of the proposal of Heiler and Feng (1996, 1999) due to the use of the k-NN approach and a xed order of polynomial for all observation points of the time series.

4 Robust procedure with given parameters


It is shown that LOWESS works well in common nonparametric regression (see the examples given in Cleveland, 1979 and in Fan and Gijbels, 1996). Cleveland et al. (1990) adapted the LOWESS to time series decomposition, where seasonal uctuations are treated in a di erent way. In this section we will propose a robust procedure for time series decomposition based on another adaptation of the LOWESS. Our proposal di ers from the original LOWESS in two points: 1. The so-called season-dependent medians are introduced, so that the LOWESS may be easily adapted to our model. 2. A stability criterion is used, which allows the number of the robust iterations (NRI) to be decided by the data. In this section the parameters hT and p are assumed to be chosen by hand. The robust data-driven version of the BV will be introduced in the next section. Detailed discussion on the properties of such a robust procedure is omitted. The basic idea of the LOWESS is as follows. Fit m ^ 0 in the 0-th (nonrobust) iteration with weights k0 i(t) = ki(t) as de ned in section 2.1. Calculate the residuals rj t = yt ; m ^ j;1(t) in the j -th (1 j J ) iteration from the estimate obtained in the j ; 1-th iteration, where J is the number of robust iterations (NRI) given beforehand. Assign the robustness weight to each observation as j t = B (rj t=(6 j )), t = 1 2 ::: n, where j denotes the 9

median of jrj tj and B (u) = (1 ; u2)21I ;1 1] is the bisquare weighting function. Fit m ^j with the weights k0 i(t) being replaced by kj i(t) = j ik0 i(t). However, in the case of time series decomposition the residuals often depend on the seasonal component. Hence, a uniform median may not be suitable in this context. Therefore it will be natural to treat the residuals within each seasonal period di erently. For this purpose we will use the season-dependent medians, j t, rather than a uniform median j , where j t is the median of jrj ij for all i such that (i ; t)=s is an integer. This means that the residuals are at rst divided into s groups. The robustness weights are then calculated in each group following the idea of Cleveland (1979). We will see that the use of season-dependent medians works well for decomposing seasonal time series. As in Cleveland (1979) the function B (u) = (1 ; u2)21I ;1 1] will be used to calculate the robustness weights, which can, of cause, be replaced by another symmetric nonnegative kernel function. Assume that xed p and hT are used in all iterations. Then the adaptation of the LOWESS into the time series decomposition context works as follows: 1. In the 0-th iteration of this procedure decompose the time series by locally weighted regression using the weights k0 i(t) = ki(t) and obtain m ^ 0. 2. In the j th iteration, let rj t = yt ; m ^ j;1(t), t = 1 2 ::: n, denote the residuals obtained from the (j -1)th iteration. De ne the robustness weights by ! r jt t = 1 2 ::: n jt = B 6 jt where j t is the season-dependent median as described above. 3. Decompose the time series with ki(t) replaced by the modi ed weights kj i(t) = j i k0 i (t). 4. Repeatedly carry out steps 2 and 3 until a stability criterion is ful lled or up to a given number of iterations. An important question is how many robust iterations should be carried out in oder to obtain satisfactory results. It is clear that at least two robust iterations are needed, since 10

in the rst step the robustness weights are calculated from of residuals obtained by a nonrobust procedure. In the following we will propose a stability criterion such that the number NRI can be determined by the data. To simplify the description we put 0 i 1 for all i in the 0-th iteration. For given p and bandwidth hT the weights kj i(t) in the j th iteration depend only on j i. Hence the estimates in the j th iteration will be close to those in the (j -1)th iteration, if 1 for all j and i, we can use the averaged j i ' (j ;1) i for all i. Observing that 0 ji absolute di erence (AAD) between j i and (j;1) i: n X AADj = n;1 j j i ; (j;1) ij j = 1 2 ::: as a measure of stability of the robust estimates in the j th iteration. For given p and hT AADj will converge to zero, as j ! 1, if the robust procedure is stable. Hence we can choose a small positive constant c0 and stop the iterative procedure when AADj < c0 . Such an estimator will be called a stable robust estimate, or simply a robust estimate, if no confusion is to be expected. The robust estimate depends strongly on c0. If c0 is too large, the results will not be satisfactory. If c0 is too small, the computer time will be unnecessarily large. In this paper c0 = 0:0125 will be used.
i=1

5 Data-driven robust procedure


The error criteria given in section 3 are proposed for a linear smoother under the assumption that the t are iid. It is clear that the robust locally weighted regression estimators are nonlinear, since the robustness weights j t for j > 0 depend on the data. In this case the weighting system w also depends on the data. From the robust procedure described in the last section we can see that the dependence of the weighting system on the data is very complex. In this paper, however, we will simply use the R-statistic and DS criterion as approximate methods to select the order of polynomial and the bandwidth in a robust iteration. This is just an attempt to develop a robust data-driven time series decomposition procedure. There are still many open questions in this area (see the nal remarks). 11

The extension of the bandwidth selection procedure to a robust iteration, when the nonlinearity is ignored, is straightforward. Both, the pilot bandwidth and the main bandwidth have to be reselected in each iterative step, since the optimal bandwidth in the next robust iteration may be di erent from that in the last one. Hence the data-driven robust procedure needs large computing time. The local regression estimates depend strongly on the bandwidth hT. Besides the stability condition AADj < c0 an additional stability condition on the bandwidth will also be used. Denote the bandwidth for the main smoothing ^ T j. Assume that j 2. Then the data-driven procedure will be in the j th iteration as h ^T j = h ^ T (j;1). When the procedure is not stable, it will stopped when AADj < c0 and h be stopped after J iterations. In this case the best NRI has to be chosen subjectively by analyzing the detailed results at each iteration. The proposed data-driven robust time series decomposition procedure reads as ^p 0, h ^ T 0 and compute m 1. For j = 0 obtain p ^, the pilot bandwidth h ^ 0 following the nonrobust data-driven procedure as given in section 3. 2. In the j -th, j > 0, iteration compute the robustness weights from m ^ (j;1) obtained in the (j -1)-th iteration. ^ p j following the R-statistic with the robust 3. Select the optimal pilot bandwidth h procedure. Calculate the robust pilot estimate m ^ pj . ^ T j following the DS criterion with the robust proce4. Select the optimal bandwidth h dure. Calculate the robust estimate m ^ j. 5. For j 2 put j0 := j and go to step 6, if the stability conditions are satis ed or j0 = J . Otherwise, put j := j + 1 and go back to step 2. 6. Calculate other estimates with the selected parameters and the robust procedure. The selected parameters of such a procedure is the polynomial order p ^ and a sequence of ^T 0 ^ bandwidths fh hT 1 ::: ^ hT j0 g. The estimates at the end of this procedure depend not ^ T j0 but on the whole sequence fh ^T 0 h ^ T 1 ::: h ^ T j0 g. If we let the robust procedure only on h 12

^ T j, proposed in the last section, run j0 + 1 times with p ^ and corresponding bandwidth, h in each iteration, we will obtained the same estimates.

6 Applications
In this section the proposed procedure will be applied to the seven examples used by the recent empirical studies carried out by the German Federal Statistical O ce (Speth, 1994 and Hopfner, 1998). These are observations in the old States in Germany from January 1976 to June 1987 for the following time series: 1. Unemployment (UNEMP), 2. Production index of production industries (PIPIN), 3. Production index of automobile industry (PIAUT), 4. Production index of manufacture of tobacco (PITOB), 5. Production index of chemical industry (PICHE), 6. Index of orders received in engineering industry (OIENG), 7. Index of orders received in building construction (OIBUI). Here c0 = 0:0125 and J = 20 are used for decomposing these time series. The estimated 2, p ^T = h ^ T j0 are listed in Table 1. The bandwidths selected in parameters ^B ^, j0 and h other iterations are omitted. From Table 1 we see that the selected parameters for these time series are quite di erent. This shows that a model with given parameters can not be suitable for all data sets in practice. Hence a data-driven procedure is required. The ^ T depend on the variance 2 of a time series and also on the structure of the selected p ^ and h ^ T. On the other hand, trend-cyclical component. The larger 2 (relatively), the larger h ^ T. Furthermore, p ^ T depend also mutually on the more complex G is, the smaller is h ^ and h 13

^ T, and vice versa. For all of the examples we have each other. The higher p ^, the larger h obtained stables robust decomposition results, except for the time series OIBIU, for which ^ T = 49 we have j0 = J . For this time series the optimal bandwidth switches between h ^ T = 37. If J = 19 or J = 21 were used, we would have obtained h ^ T as the nal and h result.
2, p ^ T selected for all examples Table 1: ^B ^, j0 and h

Time Series UNEMP PIPIN PIAUT PITOB PICHE OIENG OIBUI

2 ^B 470.79 3.767 57.20 15.55 2.734 28.01 5.709

p ^ 4 2 0 0 2 1 2

j0 10 7 4 8 7 4 20

^T h 55 55 63 43 37 41 49

Decomposition results for the three time series PIAUT, PITOB and OIENG, which seem to be a ected by outliers at some points, are shown in gures 1 through 3. Each of the gure gives the original data together with the nonrobust as well as the stable robust estimates of G (upper), the nonrobust estimate of S (middle) as well the stable robust estimate of S (lower), respectively. We see, in all of these examples both, the estimates of G and S are improved by the robust procedure at instants, where there seem to be some outliers. Figure 2 shows that (see the observations around t = 80) the e ect of outliers, which are di cult to make out at the rst glance, can be corrected by means of the season-dependent medians. And the corrected seasonal component looks much better than the nonrobust one. For more examples see Feng (1999).

14

7 Final remarks
In this paper a robust data-driven version of the BV is proposed. Examples from the practice show that the proposed procedure works well. However, there are some open questions to be resolved in future. For instance, some error criteria, which are more suitable for selecting optimal parameters by the robust time series decomposition procedure 2 used in as those proposed in section 4, should be developed. The variance estimator ^B this paper is nonrobust. A robust variance estimator for seasonal time series should be developed. If the variance of t depends strongly on the seasonal component, then both, the error criterion and the variance estimator should be adapted to this fact. The nite sample properties and asymptotic properties of the parameters selected by the robust data-driven procedure should be investigated. Moreover, the error process variables t are generally not independent. In this case it is more di cult to solve the above mentioned questions. Finally, note that the selected parameters are just optimal for the estimate of ^ nor for S ^. The development of procedures which yield m. They are neither optimal for G ^ or S ^ alone, are also very important. optimal parameters for G

8 Acknowledgements
This paper was supported in part by the Center of Finance and Econometrics at the University of Konstanz, Germany. We would like to thank the colleagues in the Department IIA, German Federal Statistical O ce for providing us with the data.

15

References
1] Akaike, H. (1979). A bayesian extension of the minimum AIC procedure of autoregressive model tting. Biometrika, 66, 1979, 237{242. 2] Cleveland, R.B., Cleveland, W.S., McRae, I.E. and Terpenning, I. (1990). STL: A seasonal-trend decomposition procedure based on LOWESS (with discussion). J. Ofcial Statistics, 6, 3{73. 3] Cleveland, W.S. (1979). Robust locally weighted regression and smoothing scatterplots. J. Amer. Statist. Assoc., 74, 829{836. 4] Eurostat (1998). Seasonal Adjustment Methods { A Comparison. Luxembourg 5] Fan, J. and Gijbels, I. (1995). Data-driven bandwidth selection in local polynomial tting: Variable bandwidth and spatial adaptation. J. Roy. Statist. Soc. Ser. B, 57 371{394. 6] Fan, J. and Gijbels, I. (1996). Local Polynomial Modeling and its Applications. Chipman and Hall, London. 7] Feng, Y. (1999). Kernel- and Locally Weighted Regression { with applications to time series decomposition. Verlag fur Wissenschaften und Forschung, Berlin. 8] Feng, Y. and Heiler, S. (1999). Selecting bandwidth for nonparametric regression based on double smoothing rule. Preprint, University of Konstanz. 9] Gasser, Th., Muller, H.G. and Mammitzsch, V. (1985). Kernels for nonparametric curve estimation. J. Roy. Statist. Soc. Ser. B, 47, 238{252. 10] Heiler, S. (1960). Analyse der Struktur Wirtschaftlicher Prozesse durch Zerlegung von Zeitreihen. Dissertation, University of Tubingen. 11] Heiler, S. (1970). Theoretische Grundlagen des `Berliner Verfahrens'. In Wetzel, W. (ed.): Neuere Entwicklungen auf dem Gebiet der Zeitreihenanalyse, Sonderheft 1 zum Allg. Statistischen Archiv, 67{93. 16

12] Heiler, S. (1995). Zur Glattung saisonaler Zeitreihen. In Rinne, H., Ruger, B. and Strecker, H. (eds.): Grundlagen der Statistik und ihre Anwendungen, Festschrift fur Kurt Weichselberger, Physika-Verlag, Heidelberg, 128{148. 13] Heiler, S. and Feng, Y. (1996). Datengesteuerte Zerlegung saisonaler Zeitreihen. ifo Studien, 3/1996, 41{73. 14] Heiler, S. and Feng, Y. (1999). Data-driven decomposition of seasonal time series. To appear in J. of Statistical Planning and Inference. 15] Heiler, S. and Michels, P. (1994). Deskriptive und Explorative Datenanalyse. Oldenbourg, Munchen. 16] Hopfner, B. (1998). Ein empirischer Vergleich neuerer Verfahren zur Saisonbereinigung und Komponentenzerlegung. Wirtschaft und Statistik, 12/1998, 949{959. 17] Nourney, M. (1983). Umstellung der Zeitreihenanalyse. Wirtschaft und Statistik, 11/1983, 841{852. 18] Rice, J. (1984). Bandwidth choice for nonparametric regression. Ann. Statist., 12, 1215{1230. 19] Speth, H.-Th. (1994). Vergleich von Verfahren zur Komponentenzerlegung von Zeitreihen. Wirtschaft und Statistik, 2/1994, 98{108. 20] Stone, C.J. (1977). Consistent nonparametric regression (with discussion). Ann. Statist., 5, 595{620. 21] Schwarz, G. (1978). Estimating the dimension of a model. Ann. Statist., 6, 461{464.

17

Figure 1: Nonrobust and stable robust decomposition results for the time series PIAUT. The gures show the original data together with the nonrobust (solid line) as well as the stable robust (dashes) estimates of G (upper), the nonrobust estimate of S (middle) as well the stable robust estimate of S (lower), respectively. 18

Figure 2: The same results as given in gure 1 but for the time series PITOB.

19

Figure 3: The same results as given in gure 1 but for the time series OIENG.

20

You might also like