1. Trang chủ
  2. » Tài Chính - Ngân Hàng

SAS/ETS 9.22 User''''s Guide 33 pptx

10 394 0

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 10
Dung lượng 250,77 KB

Nội dung

312 ✦ Chapter 7: The ARIMA Procedure Output 7.7.2 Airline Model with Outliers SERIES A: Chemical Process Concentration Readings The ARIMA Procedure Outlier Detection Summary Maximum number searched 3 Number found 3 Significance used 0.01 Outlier Details Approx Chi- Prob> Obs Type Estimate Square ChiSq 135 Additive -0.10310 12.63 0.0004 62 Additive -0.08872 12.33 0.0004 29 Additive 0.08686 11.66 0.0006 The output shows that a few outliers still remain to be accounted for and that the model could be refined further. References ✦ 313 References Akaike, H. (1974), “A New Look at the Statistical Model Identification,” IEEE Transaction on Automatic Control, AC–19, 716–723. Anderson, T. W. (1971), The Statistical Analysis of Time Series, New York: John Wiley & Sons. Andrews and Herzberg (1985), A Collection of Problems from Many Fields for the Student and Research Worker, New York: Springer–Verlag. Ansley, C. (1979), “An Algorithm for the Exact Likelihood of a Mixed Autoregressive Moving- Average Process,” Biometrika, 66, 59. Ansley, C. and Newbold, P. (1980), “Finite Sample Properties of Estimators for Autoregressive Moving-Average Models,” Journal of Econometrics, 13, 159. Bhansali, R. J. (1980), “Autoregressive and Window Estimates of the Inverse Correlation Function,” Biometrika, 67, 551–566. Box, G. E. P. and Jenkins, G. M. (1976), Time Series Analysis: Forecasting and Control, San Francisco: Holden-Day. Box, G. E. P., Jenkins, G. M., and Reinsel, G. C. (1994), Time Series Analysis: Forecasting and Control, Third Edition, Englewood Cliffs, NJ: Prentice Hall, 197–199. Box, G. E. P. and Tiao, G. C. (1975), “Intervention Analysis with Applications to Economic and Environmental Problems,” JASA, 70, 70–79. Brocklebank, J. C. and Dickey, D. A. (2003), SAS System for Forecasting Time Series, Second Edition, Cary, North Carolina: SAS Institute Inc. Brockwell, P. J. and Davis, R. A. (1991), Time Series: Theory and Methods, Second Edition, New York: Springer-Verlag. Chatfield, C. (1980), “Inverse Autocorrelations,” Journal of the Royal Statistical Society, A142, 363–377. Choi, ByoungSeon (1992), ARMA Model Identification, New York: Springer-Verlag, 129–132. Cleveland, W. S. (1972), “The Inverse Autocorrelations of a Time Series and Their Applications,” Technometrics, 14, 277. Cobb, G. W. (1978), “The Problem of the Nile: Conditional Solution to a Change Point Problem,” Biometrika, 65, 243–251. Davidson, J. (1981), “Problems with the Estimation of Moving-Average Models,” Journal of Econo- metrics, 16, 295. Davies, N., Triggs, C. M., and Newbold, P. (1977), “Significance Levels of the Box-Pierce Portman- teau Statistic in Finite Samples,” Biometrika, 64, 517–522. 314 ✦ Chapter 7: The ARIMA Procedure de Jong, P. and Penzer, J. (1998), “Diagnosing Shocks in Time Series,” Journal of the American Statistical Association, Vol. 93, No. 442. Dickey, D. A. (1976), “Estimation and Testing of Nonstationary Time Series,” unpublished Ph.D. thesis, Iowa State University, Ames. Dickey, D. A., and Fuller, W. A. (1979), “Distribution of the Estimators for Autoregressive Time Series with a Unit Root,” Journal of the American Statistical Association, 74 (366), 427–431. Dickey, D. A., Hasza, D. P., and Fuller, W. A. (1984), “Testing for Unit Roots in Seasonal Time Series,” Journal of the American Statistical Association, 79 (386), 355–367. Dunsmuir, William (1984), “Large Sample Properties of Estimation in Time Series Observed at Unequally Spaced Times,” in Time Series Analysis of Irregularly Observed Data, Emanuel Parzen, ed., New York: Springer-Verlag. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M. C., and Chen, B. C. (1998), “New Capabilities and Methods of the X-12-ARIMA Seasonal Adjustment Program,” Journal of Business and Economic Statistics, 16, 127–177. Fuller, W. A. (1976), Introduction to Statistical Time Series, New York: John Wiley & Sons. Hamilton, J. D. (1994), Time Series Analysis, Princeton: Princeton University Press. Hannan, E. J. and Rissanen, J. (1982), “Recursive Estimation of Mixed Autoregressive Moving- Average Order,” Biometrika, 69 (1), 81–94. Harvey, A. C. (1981), Time Series Models, New York: John Wiley & Sons. Jones, Richard H. (1980), “Maximum Likelihood Fitting of ARMA Models to Time Series with Missing Observations,” Technometrics, 22, 389–396. Kohn, R. and Ansley, C. (1985), “Efficient Estimation and Prediction in Time Series Regression Models,” Biometrika, 72, 3, 694–697. Ljung, G. M. and Box, G. E. P. (1978), “On a Measure of Lack of Fit in Time Series Models,” Biometrika, 65, 297–303. Montgomery, D. C. and Johnson, L. A. (1976), Forecasting and Time Series Analysis, New York: McGraw-Hill. Morf, M., Sidhu, G. S., and Kailath, T. (1974), “Some New Algorithms for Recursive Estimation on Constant Linear Discrete Time Systems,” IEEE Transactions on Automatic Control, AC–19, 315–323. Nelson, C. R. (1973), Applied Time Series for Managerial Forecasting, San Francisco: Holden-Day. Newbold, P. (1981), “Some Recent Developments in Time Series Analysis,” International Statistical Review, 49, 53–66. Newton, H. Joseph and Pagano, Marcello (1983), “The Finite Memory Prediction of Covariance Stationary Time Series,” SIAM Journal of Scientific and Statistical Computing, 4, 330–339. References ✦ 315 Pankratz, Alan (1983), Forecasting with Univariate Box-Jenkins Models: Concepts and Cases, New York: John Wiley & Sons. Pankratz, Alan (1991), Forecasting with Dynamic Regression Models, New York: John Wiley & Sons. Pearlman, J. G. (1980), “An Algorithm for the Exact Likelihood of a High-Order Autoregressive Moving-Average Process,” Biometrika, 67, 232–233. Priestly, M. B. (1981), Spectra Analysis and Time Series, Volume 1: Univariate Series, New York: Academic Press Schwarz, G. (1978), “Estimating the Dimension of a Model,” Annals of Statistics, 6, 461–464. Stoffer, D. and Toloi, C. (1992), “A Note on the Ljung-Box-Pierce Portmanteau Statistic with Missing Data,” Statistics & Probability Letters 13, 391–396. Tsay, R. S. and Tiao, G. C. (1984), “Consistent Estimates of Autoregressive Parameters and Extended Sample Autocorrelation Function for Stationary and Nonstationary ARMA Models,” JASA, 79 (385), 84–96. Tsay, R. S. and Tiao, G. C. (1985), “Use of Canonical Analysis in Time Series Model Identification,” Biometrika, 72 (2), 299–315. Woodfield, T. J. (1987), “Time Series Intervention Analysis Using SAS Software,” Proceedings of the Twelfth Annual SAS Users Group International Conference, 331–339. Cary, NC: SAS Institute Inc. 316 Chapter 8 The AUTOREG Procedure Contents Overview: AUTOREG Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . 318 Getting Started: AUTOREG Procedure . . . . . . . . . . . . . . . . . . . . . . . 320 Regression with Autocorrelated Errors . . . . . . . . . . . . . . . . . . . . 320 Forecasting Autoregressive Error Models . . . . . . . . . . . . . . . . . . . . 327 Testing for Autocorrelation . . . . . . . . . . . . . . . . . . . . . . . . . . 329 Stepwise Autoregression . . . . . . . . . . . . . . . . . . . . . . . . . . . . 332 Testing for Heteroscedasticity . . . . . . . . . . . . . . . . . . . . . . . . . 334 Heteroscedasticity and GARCH Models . . . . . . . . . . . . . . . . . . . . 338 Syntax: AUTOREG Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342 Functional Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342 PROC AUTOREG Statement . . . . . . . . . . . . . . . . . . . . . . . . . 346 BY Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347 CLASS Statement (Experimental) . . . . . . . . . . . . . . . . . . . . . . . . 347 MODEL Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 348 HETERO Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 362 NLOPTIONS Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364 RESTRICT Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364 TEST Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365 OUTPUT Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 367 Details: AUTOREG Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . 370 Missing Values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 370 Autoregressive Error Model . . . . . . . . . . . . . . . . . . . . . . . . . . 370 Alternative Autocorrelation Correction Methods . . . . . . . . . . . . . . . 374 GARCH Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375 Goodness-of-fit Measures and Information Criteria . . . . . . . . . . . . . . . 381 Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 384 Predicted Values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 405 OUT= Data Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 410 OUTEST= Data Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 410 Printed Output . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 412 ODS Table Names . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 413 ODS Graphics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 415 Examples: AUTOREG Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . 416 Example 8.1: Analysis of Real Output Series . . . . . . . . . . . . . . . . . 416 318 ✦ Chapter 8: The AUTOREG Procedure Example 8.2: Comparing Estimates and Models . . . . . . . . . . . . . . . 420 Example 8.3: Lack-of-Fit Study . . . . . . . . . . . . . . . . . . . . . . . . 425 Example 8.4: Missing Values . . . . . . . . . . . . . . . . . . . . . . . . . 429 Example 8.5: Money Demand Model . . . . . . . . . . . . . . . . . . . . . 434 Example 8.6: Estimation of ARCH(2) Process . . . . . . . . . . . . . . . . 439 Example 8.7: Estimation of GARCH-Type Models . . . . . . . . . . . . . . 442 Example 8.8: Illustration of ODS Graphics . . . . . . . . . . . . . . . . . . . 447 References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 457 Overview: AUTOREG Procedure The AUTOREG procedure estimates and forecasts linear regression models for time series data when the errors are autocorrelated or heteroscedastic. The autoregressive error model is used to correct for autocorrelation, and the generalized autoregressive conditional heteroscedasticity (GARCH) model and its variants are used to model and correct for heteroscedasticity. When time series data are used in regression analysis, often the error term is not independent through time. Instead, the errors are serially correlated (autocorrelated). If the error term is autocorrelated, the efficiency of ordinary least squares (OLS) parameter estimates is adversely affected and standard error estimates are biased. The autoregressive error model corrects for serial correlation. The AUTOREG procedure can fit autoregressive error models of any order and can fit subset autoregressive models. You can also specify stepwise autoregression to select the autoregressive error model automatically. To diagnose autocorrelation, the AUTOREG procedure produces generalized Durbin-Watson (DW) statistics and their marginal probabilities. Exact p-values are reported for generalized DW tests to any specified order. For models with lagged dependent regressors, PROC AUTOREG performs the Durbin t test and the Durbin h test for first-order autocorrelation and reports their marginal significance levels. Ordinary regression analysis assumes that the error variance is the same for all observations. When the error variance is not constant, the data are said to be heteroscedastic, and ordinary least squares estimates are inefficient. Heteroscedasticity also affects the accuracy of forecast confidence limits. More efficient use of the data and more accurate prediction error estimates can be made by models that take the heteroscedasticity into account. To test for heteroscedasticity, the AUTOREG procedure uses the portmanteau Q test statistics (McLeod and Li 1983), Engle’s Lagrange multiplier tests (Engle 1982), tests from Lee and King (1993), and tests from Wong and Li (1995). Test statistics and significance p-values are reported for conditional heteroscedasticity at lags 1 through 12. The Bera-Jarque normality test statistic and its significance level are also reported to test for conditional nonnormality of residuals. The following tests for independence are also supported by the AUTOREG procedure for residual analysis and diagnostic checking: Brock-Dechert-Scheinkman (BDS) test, runs test, turning point test, and the rank version of the von Neumann ratio test. Overview: AUTOREG Procedure ✦ 319 The family of GARCH models provides a means of estimating and correcting for the changing variability of the data. The GARCH process assumes that the errors, although uncorrelated, are not independent, and it models the conditional error variance as a function of the past realizations of the series. The AUTOREG procedure supports the following variations of the GARCH models:  generalized ARCH (GARCH)  integrated GARCH (IGARCH)  exponential GARCH (EGARCH)  quadratic GARCH (QGARCH)  threshold GARCH (TGARCH)  power GARCH (PGARCH)  GARCH-in-mean (GARCH-M) For GARCH-type models, the AUTOREG procedure produces the conditional prediction error variances in addition to parameter and covariance estimates. The AUTOREG procedure can also analyze models that combine autoregressive errors and GARCH- type heteroscedasticity. PROC AUTOREG can output predictions of the conditional mean and variance for models with autocorrelated disturbances and changing conditional error variances over time. Four estimation methods are supported for the autoregressive error model:  Yule-Walker  iterated Yule-Walker  unconditional least squares  exact maximum likelihood The maximum likelihood method is used for GARCH models and for mixed AR-GARCH models. The AUTOREG procedure produces forecasts and forecast confidence limits when future values of the independent variables are included in the input data set. PROC AUTOREG is a useful tool for forecasting because it uses the time series part of the model in addition to the systematic part in generating predicted values. The autoregressive error model takes into account recent departures from the trend in producing forecasts. The AUTOREG procedure permits embedded missing values for the independent or dependent variables. The procedure should be used only for ordered and equally spaced time series data. 320 ✦ Chapter 8: The AUTOREG Procedure Getting Started: AUTOREG Procedure Regression with Autocorrelated Errors Ordinary regression analysis is based on several statistical assumptions. One key assumption is that the errors are independent of each other. However, with time series data, the ordinary regression residuals usually are correlated over time. It is not desirable to use ordinary regression analysis for time series data since the assumptions on which the classical linear regression model is based will usually be violated. Violation of the independent errors assumption has three important consequences for ordinary regression. First, statistical tests of the significance of the parameters and the confidence limits for the predicted values are not correct. Second, the estimates of the regression coefficients are not as efficient as they would be if the autocorrelation were taken into account. Third, since the ordinary regression residuals are not independent, they contain information that can be used to improve the prediction of future values. The AUTOREG procedure solves this problem by augmenting the regression model with an autore- gressive model for the random error, thereby accounting for the autocorrelation of the errors. Instead of the usual regression model, the following autoregressive error model is used: y t D x 0 t ˇ C t  t D ' 1  t1  ' 2  t2  : : :  ' m  tm C  t  t  IN.0;  2 / The notation  t  IN.0;  2 / indicates that each  t is normally and independently distributed with mean 0 and variance  2 . By simultaneously estimating the regression coefficients ˇ and the autoregressive error model parameters ' i , the AUTOREG procedure corrects the regression estimates for autocorrelation. Thus, this kind of regression analysis is often called autoregressive error correction or serial correlation correction. Example of Autocorrelated Data A simulated time series is used to introduce the AUTOREG procedure. The following statements generate a simulated time series Y with second-order autocorrelation: Regression with Autocorrelated Errors ✦ 321 / * Regression with Autocorrelated Errors * / data a; ul = 0; ull = 0; do time = -10 to 36; u = + 1.3 * ul - .5 * ull + 2 * rannor(12346); y = 10 + .5 * time + u; if time > 0 then output; ull = ul; ul = u; end; run; The series Y is a time trend plus a second-order autoregressive error. The model simulated is y t D 10 C 0:5t C  t  t D 1:3 t1  0:5 t2 C  t  t  IN.0; 4/ The following statements plot the simulated time series Y. A linear regression trend line is shown for reference. title 'Autocorrelated Time Series'; proc sgplot data=a noautolegend; series x=time y=y / markers; reg x=time y=y/ lineattrs=(color=black); run; The plot of series Y and the regression line are shown in Figure 8.1. . Technometrics, 22, 3 89 396 . Kohn, R. and Ansley, C. ( 198 5), “Efficient Estimation and Prediction in Time Series Regression Models,” Biometrika, 72, 3, 694 – 697 . Ljung, G. M. and Box, G. E. P. ( 197 8), “On. Models,” JASA, 79 (385), 84 96 . Tsay, R. S. and Tiao, G. C. ( 198 5), “Use of Canonical Analysis in Time Series Model Identification,” Biometrika, 72 (2), 299 –315. Woodfield, T. J. ( 198 7), “Time Series. portmanteau Q test statistics (McLeod and Li 198 3), Engle’s Lagrange multiplier tests (Engle 198 2), tests from Lee and King ( 199 3), and tests from Wong and Li ( 199 5). Test statistics and significance p-values

Ngày đăng: 02/07/2014, 14:21

TỪ KHÓA LIÊN QUAN