Using econometrics a practical guide, 4th edition

617 97 0
Using econometrics a practical guide, 4th edition

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

Thông tin tài liệu

PART THE BASIC REGRESSION MODEL CHAPTER An Overview of Regression Analysis 1.1 What Is Econometrics? 1.2 What Is Regression Analysis? 1.3 The Estimated Regression Equation 1.4 A Simple Example of Regression Analysis 1.5 Using Regression to Explain Housing Prices 1.6 Summary and Exercises 1.1 What Is Econometrics? "Econometrics is too mathematical; it's the reason my best friend isn't majoring in economics." "There are two things you don't want to see in the making—sausage and econometric research "1 "Econometrics may be defined as the quantitative analysis of actual economic phenomena " "It's my experience that 'economy-tricks' is usually nothing more than a justification of what the author believed before the research was begun." Obviously, econometrics means different things to different people To beginning students, it may seem as if econometrics is an overly complex obstacle to an otherwise useful education To skeptical observers, econometric results should be trusted only when the steps that produced those results are completely known To professionals in the field, econometrics is a fascinating set Attributed to Edward E Learner 2, Paul A Samuelson, T C Koopmans, and J R Stone, "Repo rt of the Evaluative Committee for Econometrica," Econometrica, 1954, p 141 PART I • THE BASIC REGRESSION MODEL of techniques that allows the measurement and analysis of economic phenomena and the prediction of future economic trends You're probably thinking that such diverse points of view sound like the statements of blind people trying to describe an elephant based on what they happen to be touching, and you're partially right Econometrics has both a formal definition and a larger context Although you can easily memorize the formal definition, you'll get the complete picture only by understanding the many uses of and alternative approaches to econometrics That said, we need a formal definition Econometrics, literally "economic measurement," is the quantitative measurement and analysis of actual economic and business phenomena It attempts to quantify economic reality and bridge the gap between the abstract world of economic theory and the real world of human activity To many students, these worlds may seem far apart On the one hand, economists theorize equilibrium prices based on carefully conceived marginal costs and marginal revenues; on the other, many firms seem to operate as though they have never heard of such concepts Econometrics allows us to examine data and to quantify the actions of firms, consumers, and governments Such measurements have a number of different uses, and an examination of these uses is the first step to understanding econometrics 1.1.1 Uses of Econometrics Econometrics has three major uses: describing economic reality testing hypotheses about economic theory forecasting future economic activity The simplest use of econometrics is description We can use econometrics to quantify economic activity because econometrics allows us to put numbers in equations that previously contained only abstract symbols For example, consumer demand for a particular commodity often can be thought of as a relationship between the quantity demanded (Q) and the commodity's price (P), the price of a substitute good (P S), and disposable income (Yd) For most goods, the relationship between consumption and disposable income is expected to be positive, because an increase in disposable income will be associated with an increase in the consumption of the good Econometrics actually allows us to estimate that relationship based upon past consumption, income, and prices In other words, a general and purely theoretical functional relationship like: CHAPTER • AN OVERVIEW OF REGRESSION ANALYSIS Q = f(P, Ps , Yd) (1.1) Q=31.50 -0.73P+O.11Ps +0.23Yd (1.2) can become explicit: This technique gives a much more specific and descriptive picture of the function Let's compare Equations 1.1 and 1.2 Instead of expecting consumption merely to "increase" if there is an increase in disposable income, Equation 1.2 allows us to expect an increase of a specific amount (0.23 units for each unit of increased disposable income) The number 0.23 is called an estimated regression coefficient, and it is the ability to estimate these coefficients that makes econometrics valuable The second and perhaps the most common use of econometrics is hypothesis testing, the evaluation of alternative theories with quantitative evidence Much of economics involves building theoretical models and testing them against evidence, and hypothesis testing is vital to that scientific approach For example, you could test the hypothesis that the product in Equation 1.1 is what economists call a normal good (one for which the quantity demanded increases when disposable income increases) You could this by applying various statistical tests to the estimated coefficient (0.23) of disposable income (Yd) in Equation 1.2 At first glance, the evidence would seem to suppo rt this hypothesis because the coefficient's sign is positive, but the "statistical significance" of that estimate would have to be investigated before such a conclusion could be justified Even though the estimated coefficient is positive, as expected, it may not be sufficiently different from zero to imply that the true coefficient is indeed positive instead of zero Unfortunately, statistical tests of such hypotheses are not always easy, and there are times when two researchers can look at the same set of data and come to slightly different conclusions Even given this possibility, the use of econometrics in testing hypotheses is probably its most impo rt ant function The third and most difficult use of econometrics is to forecast or predict what is likely to happen next qua rt er, next year, or further into the future, based on what has happened in the past For example, economists use econometric models to make forecasts of variables like sales, profits, Gross Domestic Product (GDP), and the inflation rate The accuracy of such forecasts depends in large measure on the degree to which the past is a good guide to the future Business leaders and politicians tend to be especially in3 The results in Equation 1.2 are from a model of the demand for chicken that we will examine in more detail in Section 6.1 PART I • THE BASIC REGRESSION MODEL terested in this use of econometrics because they need to make decisions about the future, and the penalty for being wrong (bankruptcy for the entrepreneur and political defeat for the candidate) is high To the extent that econometrics can shed light on the impact of their policies, business and government leaders will be better equipped to make decisions For example, if the president of a company that sold the product modeled in Equation 1.1 wanted to decide whether to increase prices, forecasts of sales with and without the price increase could be calculated and compared to help make such a decision In this way, econometrics can be used not only for forecasting but also for policy analysis 1.1.2 Alternative Econometric Approaches There are many different approaches to quantitative work For example, the fields of biology, psychology, and physics all face quantitative questions similar to those faced in economics and business However, these fields tend to use somewhat different techniques for analysis because the problems they face aren't the same "We need a special field called econometrics, and textbooks about it, because it is generally accepted that economic data possess ce rt ain properties that are not considered in standard statistics texts or are not sufficiently emphasized there for use by economists " Different approaches also make sense within the field of economics The kind of econometric tools used to quantify a particular function depends in pa rt on the uses to which that equation will be put A model built solely for descriptive purposes might be different from a forecasting model, for example To get a better picture of these approaches, let's look at the steps necessary for any kind of quantitative research: specifying the models or relationships to be studied collecting the data needed to quantify the models quantifying the models with the data Steps and are similar in all quantitative work, but the techniques used in step 3, quantifying the models, differ widely between and within disciplines Choosing the best technique for a given model is a theory-based skill that is often referred to as the "a rt " of econometrics There are many alternative approaches to quantifying the same equation, and each approach may Clive Granger, "A Review of Some Recent Textbooks of Econometrics," Journal of Economic Literature, March 1994, p 117 CHAII'ER • AN OVERVIEW OF REGRESSION ANALYSIS give somewhat different results The choice of approach is left to the individual econometrician (the researcher using econometrics), but each researcher should be able to justify that choice This book will focus primarily on one particular econometric approach: single-equation linear regression analysis The majority of this book will thus concentrate on regression analysis, but it is impo rt ant for every econometrician to remember that regression is only one of many approaches to econometric quantification The impo rt ance of critical evaluation cannot be stressed enough; a good econometrician can diagnose faults in a particular approach and figure out how to repair them The limitations of the regression analysis approach must be fully perceived and appreciated by anyone attempting to use regression analysis or its findings The possibility of missing or inaccurate data, incorrectly formulated relationships, poorly chosen estimating techniques, or improper statistical testing procedures implies that the results from regression analyses should always be viewed with some caution 1.2 What Is Regression Analysis? Econometricians use regression analysis to make quantitative estimates of economic relationships that previously have been completely theoretical in nature After all, anybody can claim that the quantity of compact discs demanded will increase if the price of those discs decreases (holding everything else constant), but not many people can put specific numbers into an equation and estimate by how many compact discs the quantity demanded will increase for each dollar that price decreases To predict the direction of the change, you need a knowledge of economic theory and the general characteristics of the product in question To predict the amount of the change, though, you need a sample of data, and you need a way to estimate the relationship The most frequently used method to estimate such a relationship in econometrics is regression analysis 1.2.1 Dependent Variables, Independent Variables, and Causality Regression analysis is a statistical technique that attempts to "explain" movements in one variable, the dependent variable, as a function of movements in a set of other variables, called the independent (or explanatory) vari ables, through the quantification of a single equation For example in Equation 1.1: Q = f(P, Ps , Yd) (1.1) PART I ■ THE BASIC REGRESSION MODEL • Q is the dependent va ri able and P, P S, and Yd are the independent va ri ables Regression analysis is a natural tool for economists because most (though not all) economic propositions can be stated in such single-equation functional forms For example, the quantity demanded (dependent va ri able) is a function of price, the prices of substitutes, and income (independent variables) Much of economics and business is concerned with cause-and-effect propositions If the price of a good increases by one unit, then the quantity demanded decreases on average by a ce rt ain amount, depending on the price elasticity of demand (defined as the percentage change in the quantity demanded that is caused by a one percent change in price) Similarly, if the quantity of capital employed increases by one unit, then output increases by a ce rt ain amount, called the marginal productivity of capital Propositions such as these pose an if-then, or causal, relationship that logically postulates that a dependent variable's movements are causally determined by movements in a number of specific independent variables Don't be deceived by the words dependent and independent, however Although many economic relationships are causal by their very nature, a regression result, no matter how statistically significant, cannot prove causality All regression analysis can is test whether a significant quantitative relationship exists Judgments as to causality must also include a healthy dose of economic theory and common sense For example, the fact that the bell on the door of a flower shop ri ngs just before a customer enters and purchases some flowers by no means implies that the bell causes purchases! If events A and B are related statistically, it may be that A causes B, that B causes A, that some omitted factor causes both, or that a chance correlation exists between the two The cause-and-effect relationship is often so subtle that it fools even the most prominent economists For example, in the late nineteenth century, English economist Stanley Jevons hypothesized that sunspots caused an increase in economic activity To test this theory, he collected data on national output (the dependent variable) and sunspot activity (the independent variable) and showed that a significant positive relationship existed This result led him, and some others, to jump to the conclusion that sunspots did indeed cause output to rise Such a conclusion was unjustified because regression analysis cannot confirm causality; it can only test the strength and direction of the quantitative relationships involved 1.2.2 Single Equation Linear Models - The simplest single-equation linear regression model is: Y = 130 + 131X (1.3) CHAPTER • AN OVERVIEW OF REGRESSION ANALYSIS Equation 1.3 states that Y, the dependent va ri able, is a single-equation linear function of X, the independent variable The model is a single-equation model because no equation for X as a function of Y (or any other variable) has been specified The model is linear because if you were to plot Equation 1.3 on graph paper, it would be a straight line rather than a curve The [3s are the coefficients that determine the coordinates of the straight line at any point Ro is the constant or intercept term; it indicates the value of Y when X equals zero 13 is the slope coefficient, and it indicates the amount that Y will change when X increases by one unit The solid line in Figure 1.1 illustrates the relationship between the coefficients and the graphical meaning of the regression equation As can be seen from the diagram, Equation 1.3 is indeed linear The slope coefficient, [3 , shows the response of Y to a change in X Since being able to explain and predict changes in the dependent variable is the essential reason for quantifying behavioral relationships, much of the emphasis in regression analysis is on slope coefficients such as (3 In Figure 1.1 for example, if X were to increase from X to X2 (AX), the value of Y in Equation 1.3 would increase from Y to Y2 (AY) For linear (i.e., straight-line) regression models, the response in the predicted value of Y due to a change in X is constant and equal to the slope coefficient 13 : (Y2 Y1) AY (X2 _ X1 ) = AX — 131 where is used to denote a change in the variables Some readers may recognize this as the "rise" (AY) divided by the "mn" (AX) For a linear model, the slope is constant over the entire function We must distinguish between an equation that is linear in the variables and one that is linear in the coefficients This distinction is impo rt ant because if linear regression techniques are going to be applied to an equation, that equation must be linear in the coefficients An equation is linear in the variables if plotting the function in terms of X and Y generates a straight line For example, Equation 1.3: Y = R o + 13 1X (1.3) is linear in the variables, but Equation 1.4: Y = Ro + R1X2 (1.4) is not linear in the variables because if you were to plot Equation 1.4 it 602 APPENDIX A $4,000,000 — $400,000 + $4,000,000 = $7,600,000 and average wealth rises to $7,600,000/100 = $76,000 Such an error has no effect on the median, because it simply causes one high value to be replaced by another This exercise illustrates the general principle that the median is less sensitive than the mean to measurement errors 16-4 Because the numbers on each side are equally likely, we can reason directly that a six-sided die has an expected value of 3.5 and a foursided die has an expected value of 2.5 Because the possibilities are more spread out on the six-sided die (1 through versus through 4), we know that the six-sided die has the larger standard deviation 16-6 The z values and normal probabilities are: P[x > 270] = P P[x > 310] = P x µ > 270 - 266 16 x µ> 310 - 266 16 = P[z > 0.25] = 0.4013 = P[z > 2.75] = 0.003 16-8 The high-school seniors who take the SAT are not a random sample because this test is taken by students who intend to go to college; these are generally students with above-average scholastic aptitude The relationship between the fraction of a state's seniors that takes the SAT and the state's average SAT score is negative If a small fraction of the state's seniors takes the SAT, it will mostly consist of the state's best students As the fraction of a state's students taking the SAT increases, the group of students that takes the SAT is increasingly composed of weaker students, who bring down the state's average SAT 16 10 - The mean is 299,756.2174 and the standard deviation is 107.1146 Table B-4 in the appendix shows that with 23 — = 22 degrees of freedom, the appropriate t-value for a 99 percent confidence interval is 2.819 A 99 percent confidence interval does include the value 299,710.5 that is now accepted as the speed of light: ANSWERS TO EVEN-NUMBERED EXERCISES 603 x ± t *( ' s Vn 299,756.2174 ± 2.8191 \ = 299,756.2 ± 63.0 10 1146 " ^ V23 V 16-12 If x is N[215, 10] then for a random sample of size n = 20 P[x?257]=P 257-215 µ cr / ^ =P[z?18.8]~0 10/V20 _ X Dr Frank's patients may choose to be medical patients because they have heart problems Any trait they happen to share will then seemingly explain the heart disease; however, the standard statistical tests are not valid if these are not a random sample from the population of all people with earlobe creases 16-14 The null hypothesis is that the population mean is 33.4 percent The sample mean is 18.300, the standard deviation is 8.636, and the t value is —8.742: X t= µ "^ s/Vn _ 18.300 '— 33.4 / 8.636/ V 25 = 8.742 APPENDIX Statistical Tables The following tables present the critical values of various statistics used primarily for hypothesis testing The primary applications of each statistic are explained and illustrated The tables are: B-1 Critical Values of the t-Distribution B-2 Critical Values of the F-Statistic: Percent Level of Significance B-3 Critical Values of the F-Statistic: Percent Level of Significance B-4 Critical Values of the Durbin-Watson Test Statistics d L and du : Percent Level of Significance B-5 Critical Values of the Durbin-Watson Test Statistics d L and d u : 2.5 Percent Level of Significance B-6 Critical Values of the Durbin-Watson Test Statistics d L and d u : Percent Level of Significance B-7 The Normal Distribution B-8 The Chi-Square Distribution 605 B 606 APPENDIX B Table B-1: The t-Distribution The t-distribution is used in regression analysis to test whether an estimated slope coefficient (say Rk) is significantly different from a hypothesized value (such as 13 H0) The t-statistic is computed as ' — 13H0)/SE(K) tk — Ow Rk where (3 k is the estimated slope coefficient and SE(R k) is the estimated standard error of j3 k To test the one-sided hypothesis: H0: Pk 13H Pk > RH°HA: the computed t-value is compared with a critical t-value t c, found in the t-table on the opposite page in the column with the desired level of significance for a one-sided test (usually or 10 percent) and the row with n — K — degrees of freedom, where n is the number of observations and K is the number of explanatory variables If I tk > t c and if tk has the sign implied by the alte rn ative hypothesis, then reject H 0; otherwise, not reject H0 In most econometric applications, 13H0 is zero and most computer regression programs will calculate t k for 13 H0 = For example, for a percent onesided test with 15 degrees of freedom, t c = 1.753, so any positive tk larger than 1.753 would lead us to reject H and declare that Pk is statistically significant in the hypothesized direction at the 95 percent level of confidence For a two-sided test, H : 13 k = RHo the HA: Rk # RHo, the procedure is identical except that the column corresponding to the two-sided level of significance is used For example, for a percent two-sided test with 15 degrees of freedom, t, = 2.131, so any tk larger in absolute value than 2.131 would lead us to reject H and declare that (3 k is significantly different from 13 H0 at the 95 percent level of confidence Another use of the t-test is to determine whether a simple correlation coefficient (r) between two variables is statistically significant That is, the null hypothesis of no correlation between two variables can be tested with: tr = r\/(n — 2)/V(1 — r 2) where n is the number of observations This tr is then compared with the appropriate tc (n degrees of freedom) using the methods outlined above For more on the t-test, see Chapter STATISTICAL TABLES TABLE B-1 CRITICAL VALUES OF THE t-DISTRIBUTION Level of Significance 10% 20% 5% 10% 2.5% 5% 1% 2% 0.5% °/U 3.078 1.886 1.638 1.533 1.476 6.314 2.920 2.353 2.132 2.015 12.706 4.303 3.182 2.776 2.571 31.821 6.965 4.541 3.747 3.365 63.657 9.925 5.841 4.604 4.032 10 1.440 1.415 1.397 1.383 1.372 943 1.895 1.860 1.833 1.812 2.447 2.365 2.306 2.262 2.228 3.143 2.998 2.896 2.821 2.764 3.707 3.499 3.355 3.250 3.169 11 12 13 14 15 1.363 1.356 1.350 1.345 1.341 1.796 1.782 1.771 1.761 1.753 2.201 2.179 2.160 2.145 2.131 2.718 2.681 2.650 2.624 2.602 3.106 3.055 3.012 2.977 2.947 16 17 18 19 20 1.337 1.333 330 1.328 1.325 1.746 1.740 734 1.729 1.725 2.120 2.110 2.101 2.093 2.086 2.583 2.567 2.552 2.539 2.528 2.921 2.898 2.878 2.861 2.845 21 22 23 24 25 1.323 1.321 1.319 1.318 1.316 1.721 717 1.714 1.711 1.708 2.080 2.074 2.069 2.064 2.060 2.518 2.508 2.500 2.492 2.485 2.831 2.819 2.807 2.797 2.787 26 27 28 29 30 1.315 1.314 1.313 1.311 1.310 1.706 1.703 1.701 1.699 1.697 2.056 2.052 2.048 2.045 2.042 2.479 2.473 2.467 2.462 2.457 2.779 40 60 120 1.303 1.296 1.289 1.684 671 1.658 2.021 2.000 1.980 2.423 2.390 2.358 2.704 2.660 2.617 1.282 1.645 1.960 2.326 2.576 Degrees of Freedom One Sided: Two Sided: 2.771 2.763 2.756 2.750 (Normal) 00 Source: Reprinted from Table IV in Sir Ronald A Fisher, Statistical Methods for Research Workers, 14th ed (copyright © 1970, University of Adelaide) with permission of Hafner, a Division of the Macmillan Publishing Company, Inc 607 608 APPENDIX B Table B-2: The F-Distribution The F-distribution is used in regression analysis to test two-sided hypotheses about more than one regression coefficient at a time To test the most typical joint hypothesis (a test of the overall significance of the regression): Ho : R1=R2= HA: H o is not true = PK = O the computed F-value is compared with a critical F-value, found in one of the two tables that follow The F-statistic has two types of degrees of freedom, one for the numerator (columns) and one for the denominator (rows) For the null and alternative hypotheses above, there are K numerator (the number of restrictions implied by the null hypothesis) and n — K denominator degrees of freedom, where n is the number of observations and K is the number of explanatory variables in the equation This particular F-statistic is printed out by most computer regression programs For example, if K = and n = 30, there are numerator and 24 denominator degrees of freedom, and the critical F-value for a percent level of significance (Table B-2) is 2.62 A computed F-value greater than 2.62 would lead us to reject the null hypothesis and declare that the equation is statistically significant at the 95 percent level of confidence For more on the F-test, see Sections 5.5 and 7.7 STATISTICAL TABLES TABLE B-2 CRITICAL VALUES OF THE F-STATISTIC: PERCENT LEVEL OF SIGNIFICANCE v2 = Degrees of Free dom for Denom inator y1 = Degrees of Freedom for Numerator 2 161 18.5 10.1 7.71 6.61 200 19.0 9.55 6.94 5.79 10 5.99 5.59 5.32 5.12 4.96 5.14 4.76 4.53 4.74 4.35 4.12 4.46 4.07 3.84 4.26 3.86 3.63 4.10 3.71 3.48 11 12 13 14 15 4.84 3.98 4.75 3.89 4.67 3.81 4.60 3.74 4.54 3.68 16 17 18 19 20 4.49 4.45 4.41 4.38 4.35 21 22 23 24 25 4.32 3.47 3.07 2.84 2.68 4.30 3.44 3.05 2.82 2.66 4.28 3.42 3.03 2.80 2.64 4.26 3.40 3.01 2.78 2.62 4.24 3.39 2.99 2.76 2.60 30 40 60 120 x 4.17 4.08 4.00 3.92 3.84 3.63 3.59 3.55 3.52 3.49 216 225 19.2 19.2 9.28 9.12 6.59 6.39 5.41 5.19 230 19.3 9.01 6.26 5.05 234 19.3 8.94 6.16 4.95 10 12 20 00 237 239 242 19.4 19.4 19.4 8.89 8.85 8.79 6.09 6.04 5.96 4.88 4.82 4.74 244 248 254 19.4 19.4 19.5 8.74 8.66 8.53 5.91 5.80 5.63 4.68 4.56 4.36 4.39 4.28 3.97 3.87 3.69 3.58 3.48 3.37 3.33 3.22 4.21 3.79 3.50 3.29 3.14 4.15 4.06 3.73 3.64 3.44 3.35 3.23 3.14 3.07 2.98 4.00 3.87 3.67 3.57 3.44 3.23 3.28 3.15 2.93 3.07 2.94 2.71 2.91 2.77 2.54 3.59 3.49 3.41 3.34 3.29 3.36 3.20 3.09 3.26 3.11 3.00 3.18 3.03 2.92 3.11 2.96 2.85 3.06 2.90 2.79 3.01 2.91 2.83 2.76 2.71 2.95 2.85 2.77 2.70 2.64 2.79 2.65 2.40 2.69 2.54 2.30 2.60 2.46 2.21 2.53 2.39 2.13 2.48 2.33 2.07 3.24 3.20 3.16 3.13 3.10 3.01 2.96 2.93 2.90 2.87 2.85 2.75 2.67 2.60 2.54 2.85 2.74 2.66 2.59 2.49 2.81 2.70 2.61 2.55 2.45 2.77 2.66 2.58 2.51 2.41 2.74 2.63 2.54 2.48 2.38 2.71 2.60 2.51 2.45 2.35 2.57 2.55 2.53 2.51 2.49 3.32 2.92 2.69 2.53 2.42 3.23 2.84 2.61 2.45 2.34 3.15 2.76 2.53 2.37 2.25 3.07 2.68 2.45 2.29 2.18 3.00 2.60 2.37 2.21 2.10 2.49 2.46 2.44 2.42 2.40 2.42 2.40 2.37 2.36 2.34 2.33 2.27 2.25 2.18 2.17 2.10 2.09 2.02 2.01 1.94 2.42 2.38 2.34 2.31 2.28 2.28 2.23 2.19 2.16 2.12 2.01 1.96 1.92 1.88 1.84 2.32 2.25 2.30 2.23 2.27 2.20 2.25 2.18 2.24 2.16 2.10 2.07 2.05 2.03 2.01 1.81 1.78 1.76 1.73 1.71 2.16 2.08 1.99 1.91 1.83 1.93 1.84 1.75 1.66 1.62 1.51 1.39 1.25 1.00 2.09 2.00 1.92 1.83 1.75 1.57 Source: Abridged from M Merrington and C M Thompson, "Tables of percentage points of the inverted beta (F) distribution," Biometrika, Vol 33, 1943, p 73 By permission of the Biometrika trustees 609 610 APPENDIX B Table B-3: The F-Distribution The F-distribution is used in regression analysis to test two-sided hypotheses about more than one regression coefficient at a time To test the most typical joint hypothesis (a test of the overall significance of the regression): Ho: Ri = P2 = HA: H o is not true = PK = ° the computed F-value is compared with a critical F-value, found in Tables B-2 and B-3 The F-statistic has two types of degrees of freedom, one for the numerator (columns) and one for the denominator (rows) For the null and alternative hypotheses above, there are K numerator (the number of restrictions implied by the null hypothesis) and n — K — denominator degrees of freedom, where n is the number of observations and K is the number of explanatory variables in the equation This particular F-statistic is printed out by most computer regression programs For example, if K = and n = 30, there are numerator and 24 denominator degrees of freedom, and the critical F-value for a percent level of significance (Table B-3) is 3.90 A computed F-value greater than 3.90 would lead us to reject the null hypothesis and declare that the equation is statistically significant at the 99 percent level of confidence For more on the F-test, see Sections 5.5 and 7.7 STATISTICAL TABLES TABLE B-3 CRITICAL VALUES OF THE F-STATISTIC: PERCENT LEVEL OF SIGNIFICANCE y2 = Degrees of Freedom for Denominator vi = Degrees of Freedom for Numerator 10 12 20 co 4052 98.5 34.1 21.2 16.3 5000 99.0 30.8 18.0 13.3 5403 99.2 29.5 16.7 12.1 5625 99.2 28.7 16.0 11.4 5764 99.3 28.2 15.5 11.0 5859 99.3 27.9 15.2 10.7 5928 99.4 27.7 15.0 10.5 5982 99.4 27.5 14.8 10.3 6056 99.4 27.2 14.5 10.1 6106 99.4 271 14.4 9.89 6209 99.4 26.7 14.0 9.55 6366 99.5 26.1 13.5 9.02 10 13.7 12.2 11.3 10.6 10.0 10.9 9.55 8.65 8.02 7.56 9.78 9.15 8.45 7.85 7.59 7.01 6.99 6.42 6.55 5.99 8.75 7.46 6.63 6.06 5.64 8.47 7.19 6.37 5.80 5.39 8.26 6.99 6.18 5.61 5.20 8.10 7.87 7.72 7.40 6.88 6.84 6.62 6.47 6.16 5.65 6.03 5.81 5.67 5.36 4.86 5.47 5.26 5.11 4.81 4.31 5.06 4.85 4.71 4.41 3.91 11 12 13 14 15 9.65 9.33 9.07 8.86 8.68 7.21 6.93 6.70 6.51 6.36 6.22 5.67 5.32 5.07 5.95 5.41 5.06 4.82 5.74 5.21 4.86 4.62 5.56 5.04 4.70 4.46 5.42 4.89 4.56 4.32 4.89 4.64 4.44 4.28 4.14 4.74 4.50 4.30 4.14 16 17 18 19 20 8.53 6.23 8.40 6.11 8.29 6.01 8.19 5.93 8.10 5.85 5.29 5.19 5.09 5.01 4.94 4.77 4.67 4.58 4.50 4.43 4.44 4.20 4.34 4.10 4.25 4.01 4.17 3.94 4.10 3.87 4.03 3.89 3.93 3.79 3.84 3.71 3.77 3.63 3.70 3.56 3.69 3.59 3.51 3.43 3.37 3.55 3.46 3.37 3.30 3.23 3.26 2.75 3.16 2.65 3.08 2.57 3.00 2.49 2.94 2.42 21 22 8.02 7.95 7.88 7.82 7.77 4.87 4.82 4.76 4.72 4.68 4.37 4.31 4.26 4.22 4.18 4.04 3.99 3.94 3.90 3.86 3.64 3.51 3.31 3.59 3.45 3.26 3.54 3.41 3.21 3.50 3.36 3.17 3.46 3.32 3.13 3.17 3.12 3.07 3.03 2.99 2.88 2.83 2.78 2.74 2.70 2.36 2.31 2.26 2.21 2.17 2.98 2.84 2.80 2.66 2.63 2.50 2.47 2.34 2.32 2.18 2.55 2.37 2.20 2.03 1.88 2.01 1.80 1.60 1.38 1.00 23 24 25 30 40 60 120 co 5.78 5.72 5.66 5.61 5.57 7.56 5.39 4.51 4.02 7.31 5.18 4.31 3.83 7.08 4.98 4.13 3.65 6.85 4.79 3.95 3.48 6.63 4.61 3.78 3.32 3.81 3.76 3.71 3.67 3.63 4.54 4.40 4.10 4.30 4.16 3.86 4.10 3.96 3.66 3.94 3.80 3.51 4.00 3.80 3.67 3.37 3.70 3.47 3.30 3.17 3.51 3.29 3.12 2.99 3.34 3.12 2.95 2.82 3.17 2.96 2.79 2.66 3.02 2.80 2.64 2.51 3.60 3.36 3.17 3.00 2.87 Source: Abridged from M Merrington and C M Thompson, "Tables of percentage points of the invertecl beta (F) distribution," Biometrika, Vol 3, 1943, p 73 By permission of the Biometrika trustees 611 612 APPENDIX B Tables B4, B5, and B6: The Durbin—Watson d Statistic The Durbin-Watson d statistic is used to test for first-order serial correlation in the residuals First-order serial correlation is characterized by E t = Pet -1 + u t, where Et, is the error term found in the regression equation and ut is a classical (nonserially correlated) error term Since p = implies no serial correlation, and since most economic and business models imply positive serial correlation if any pure serial correlation exists, the typical hypotheses are: Hp:p^O HA: p> To test the null hypothesis of no positive serial correlation, the Durbin-Watson d statistic must be compared to two different critical d-values, d L and du found in the tables that follow, depending on the level of significance, the number of explanatory variables (k'), and the number of observations (n) For example, with two explanatory variables and 30 observations, the percent one-tailed critical values are d L = 1.07 and d u = 1.34, so any computed Durbin-Watson statistic less than 1.07 would lead to the rejection of the null hypothesis For computed DW d-values between 1.07 and 1.34, the test is inconclusive, and for values greater than 1.34, we can say that there is no evidence of positive serial correlation at the 99 percent level of confidence These ranges are illustrated in the diagram below: 99 Percent One-Sided Test of Ho p vs HA p > Test Inconclusive Reject Ho Do Not Reject Ho dL du 1.07 1.34 Two-sided tests are done similarly, with - d u and - d L being the critical DW d-values between and For more on this, see Chapter Tables B-5 and B-6 (for 2.5 and percent levels of significance in a one-sided test) go only up to five explanatory variables, so extrapolation for more variables (and interpolation for observations between listed points) is often in order TABLE B-4 CRITICAL VALUES OF THE DURBIN-WATSON TEST STATISTICS D i_ AND D u : PERCENT ONE-SIDED LEVEL OF SIGNIFICANCE (10 PERCENT TWO-SIDED LEVEL OF SIGNIFICANCE) n k'=1 k'=2 k'=3 k'=4 k'=5 k'=6 k'=7 d i_ d u d i d u d i_ d u d i_ d u d du d du d d u 15 16 17 18 19 20 1.08 11 13 16 1.18 1.20 1.36 1.37 1.38 1.39 1.40 1.41 0.81 0.86 0.90 0.93 0.97 1.00 1.75 1.73 1.71 1.69 1.68 1.68 0.69 0.73 0.78 0.82 0.86 0.89 1.97 1.93 1.90 1.87 1.85 1.83 0.56 0.62 0.66 0.71 0.75 0.79 2.21 2.15 2.10 2.06 2.02 1.99 0.34 0.40 0.45 0.50 0.55 0.60 2.73 2.62 2.54 2.46 2.40 2.34 21 1.22 1.42 1.13 1.54 22 1.24 1.43 1.15 1.54 23 1.26 1.44 17 1.54 24 1.27 1.45 1.19 1.55 25 1.29 1.45 1.21 1.55 1.03 1.05 1.08 10 1.12 1.67 1.66 1.66 1.66 1.66 0.93 0.96 0.99 1.01 1.04 1.81 1.80 1.79 1.78 1.77 0.83 0.86 0.90 0.93 0.95 1.96 0.73 2.12 0.64 1.94 0.77 2.09 0.68 1.92 0.80 2.06 0.72 1.90 0.84 2.04 0.75 1.89 0.87 2.01 0.78 2.29 2.25 2.21 2.17 26 27 28 29 30 1.55 1.56 1.56 1.56 1.57 1.14 1.16 1.18 1.20 1.21 1.65 1.65 1.65 1.65 1.65 1.06 1.08 1.10 1.12 14 1.76 1.76 1.75 1.74 1.74 0.98 1.00 1.03 1.05 1.07 1.88 1.86 1.85 1.84 1.83 0.90 0.93 0.95 0.98 1.00 1.99 1.97 1.96 1.94 1.93 0.82 0.85 0.87 0.90 0.93 2.12 2.09 2.07 2.05 2.03 31 1.36 1.50 32 1.37 1.50 33 1.38 1.51 34 1.39 1.51 35 1.40 1.52 1.30 1.57 1.31 1.57 1.32 1.58 1.33 1.58 1.34 1.58 1.23 1.24 1.26 1.27 1.28 1.65 1.65 1.65 1.65 1.65 1.16 1.18 19 1.21 1.22 1.74 1.73 1.73 1.73 1.73 1.09 1.11 13 1.14 16 1.83 1.82 1.81 1.81 1.80 1.02 1.04 1.06 1.08 1.10 1.92 1.91 1.90 1.89 1.88 0.95 0.97 0.99 1.02 1.03 2.02 2.00 1.99 1.98 1.97 36 37 38 39 40 1.41 1.42 1.43 1.43 1.44 1.52 1.53 1.54 1.54 1.54 1.35 1.36 1.37 1.38 1.39 1.59 1.59 1.59 1.60 1.60 1.30 1.31 1.32 1.33 1.34 1.65 1.66 1.66 1.66 1.66 1.24 1.25 1.26 1.27 1.29 1.73 1.72 1.72 1.72 1.72 1.18 1.19 1.20 1.22 1.23 1.80 1.80 1.79 1.79 1.79 1.11 1.13 1.15 16 1.18 1.88 1.05 1.96 1.87 1.07 1.95 1.86 1.09 1.94 1.86 10 1.93 1.85 1.12 1.93 45 50 55 60 65 70 75 1.48 1.50 1.53 1.55 1.57 1.58 1.60 1.57 1.59 1.60 1.62 1.63 1.64 1.65 1.43 1.46 1.49 1.51 1.54 1.55 1.57 1.62 1.63 1.64 1.65 1.66 1.67 1.68 1.38 1.42 1.45 1.48 1.50 1.53 1.54 1.67 1.67 1.68 1.69 1.70 1.70 1.71 1.34 1.38 1.41 1.44 1.47 1.49 1.52 1.72 1.72 1.72 1.73 1.73 1.74 1.74 1.29 1.34 1.37 1.41 1.44 1.46 1.49 1.78 1.77 77 1.77 1.24 1.29 1.33 1.37 1.77 1.40 1.77 1.43 1.77 1.46 1.84 1.82 1.81 1.81 1.81 1.80 1.80 1.19 1.25 1.29 1.34 1.37 1.40 1.43 1.90 1.88 1.86 1.85 1.84 1.84 1.83 80 85 90 95 100 1.61 1.62 1.63 1.64 1.65 1.66 1.67 1.68 1.69 1.69 1.59 1.60 1.61 1.62 1.63 1.69 1.70 1.70 1.71 1.72 1.56 1.58 1.59 1.60 1.61 1.72 72 1.73 1.73 1.74 1.53 1.55 1.57 1.58 1.59 1.74 1.75 1.75 1.75 1.76 1.51 1.53 1.54 1.56 1.57 1.77 1.77 1.78 1.78 1.78 1.80 1.80 1.80 1.80 80 1.45 1.47 1.49 1.51 1.53 1.83 1.83 1.83 1.83 1.83 1.46 1.47 1.48 1.48 1.49 1.22 1.24 1.26 1.27 1.28 0.45 0.50 0.55 0.60 0.65 0.69 2.47 2.39 2.32 2.26 2.21 2.16 1.54 1.54 1.54 1.53 1.53 1.54 1.30 1.32 1.33 1.34 1.35 0.95 0.98 1.02 1.05 1.07 1.10 1.48 1.50 1.52 1.54 1.55 2.14 Source: N E Savin and Kenneth J White "The Durbin-Watson Test for Serial Correlation with Extreme Sample Sizes or Many Regressors," Econometrica, November 1977, p 1994 Reprinted with permission Note: n = number of obse rv ations, k' = number of explanatory variables excluding the constant term We assume the equation contains a constant term and no lagged dependent variables (if so see Table B - 7) 613 TABLE B-5 CRITICAL VALUES OF THE DURBIN-WATSON TEST STATISTICS OF D i_ AND D u : 2.5 PERCENT ONE-SIDED LEVEL OF SIGNIFICANCE (5 PERCENT TWO-SIDED LEVEL OF SIGNIFICANCE) n k'=1 k' = k' = k' =4 k' =5 d i_ du d i_ du d i_ du d i_ du d i_ du 15 16 17 18 19 20 0.95 0.98 1.01 0.83 0.86 0.90 0.93 0.96 0.99 1.40 1.40 1.40 1.40 1.41 1.41 0.71 0.75 1.61 1.59 1.58 1.03 1.06 1.08 1.23 1.24 1.25 1.26 1.28 1.28 1.56 1.55 1.55 0.59 0.64 0.68 0.72 0.76 0.79 1.84 1.80 1.77 1.74 1.72 1.70 0.48 0.53 0.57 0.62 0.66 0.70 2.09 2.03 1.98 1.93 1.90 1.87 21 22 23 24 25 10 1.12 1.14 16 18 1.30 1.31 1.32 1.33 1.34 1.01 1.04 1.06 1.08 10 1.41 1.42 1.42 1.43 1.43 0.92 0.95 0.97 1.00 1.02 1.54 1.54 1.54 1.54 1.54 0.83 0.86 0.89 0.91 0.94 1.69 1.68 1.67 1.66 1.65 0.73 0.77 0.80 0.83 0.86 1.84 1.82 1.80 1.79 1.77 26 27 28 29 30 19 1.21 1.22 1.24 1.25 1.35 1.36 37 38 1.38 12 13 1.15 1.17 1.18 1.44 1.44 1.45 45 1.46 1.04 1.06 08 10 1.12 1.54 1.54 1.54 1.54 1.54 0.96 0.99 1.01 1.03 1.05 1.65 1.64 64 63 1.63 0.88 0.91 0.93 0.96 0.98 1.76 1.75 1.74 1.73 1.73 31 32 33 34 35 1.26 1.27 1.28 1.29 1.30 1.39 1.40 1.41 1.41 1.42 1.20 1.21 1.22 1.24 1.25 1.47 1.47 1.48 1.48 1.48 13 1.15 16 1.17 1.19 1.55 1.55 1.55 1.55 1.55 1.07 1.08 10 12 1.13 1.63 1.63 1.63 1.63 1.63 1.00 1.02 1.04 1.06 1.07 1.72 1.71 1.71 1.70 1.70 36 37 38 39 40 1.31 1.32 1.33 1.34 1.35 1.43 1.43 1.44 1.49 1.49 1.44 1.45 1.26 1.27 1.28 1.29 1.30 1.50 1.50 1.51 1.20 1.21 1.23 1.24 1.25 1.56 1.56 1.56 1.56 1.57 15 1.16 17 1.19 1.20 1.63 1.62 1.62 1.63 1.63 1.09 1.10 1.12 13 15 1.70 1.70 1.70 1.69 1.69 45 50 55 60 65 70 75 39 1.42 1.45 1.47 1.49 1.51 1.53 48 1.50 1.52 1.54 1.55 1.57 1.58 1.34 1.38 41 1.44 1.46 1.48 1.50 53 1.54 1.56 1.57 1.59 1.60 1.61 30 1.34 1.37 1.40 1.43 1.45 1.47 1.58 1.59 60 1.61 1.62 1.63 64 1.25 1.30 1.33 1.37 1.40 1.42 1.45 63 1.64 1.64 1.65 1.66 1.66 1.67 1.21 1.26 30 1.33 1.36 1.39 42 1.69 1.69 69 69 1.69 1.70 70 80 85 90 95 100 1.54 1.56 1.57 1.58 1.59 59 60 1.61 1.62 1.63 1.52 1.53 1.55 1.56 1.57 62 63 1.64 1.65 1.65 1.49 1.51 1.53 1.54 1.55 1.65 1.65 1.66 1.67 1.67 1.47 1.49 1.50 1.52 1.53 67 68 1.69 1.69 1.70 1.44 1.46 1.48 1.50 1.51 1.70 1.71 1.71 1.71 1.72 0.79 0.82 0.86 0.89 Source: J Durbin and G S Watson, "Testing for Serial Correlation in Least Squares Regression," Biometrika, Vol 38, 1951, pp 159 - 171 Reprinted with permission of the Biometrika trustees Note: n = number of observations, k' = number of explanatory variables excluding the constant term It is assumed that the equation contains a constant term and no lagged dependent variables (if T a kIP B-7) 614 TABLE B-6 CRITICAL VALUES OF THE DURBIN-WATSON TEST STATISTICS D AND D u : PERCENT ONE-SIDED LEVEL OF SIGNIFICANCE (2 PERCENT TWO-SIDED LEVEL OF SIGNIFICANCE) n k'=1 k' = k'=3 k' = k' = d i du di_ du d i du d i du d i du 15 16 17 18 19 20 0.81 0.84 0.87 0.90 0.93 0.95 1.07 1.09 10 1.12 13 15 0.70 0.74 0.77 0.80 0.83 0.86 1.25 1.25 1.25 1.26 1.26 1.27 0.59 0.63 0.67 0.71 0.74 0.77 1.46 1.44 1.43 1.42 1.41 1.41 0.49 0.53 0.57 0.61 0.65 0.68 1.70 1.66 1.63 1.60 1.58 1.57 0.39 0.44 0.48 0.52 0.56 0.60 1.96 1.90 1.85 1.80 1.77 1.74 21 22 23 24 25 0.97 1.00 1.02 1.04 1.05 16 1.17 1.19 1.20 1.21 0.89 0.91 0.94 0.96 0.98 1.27 1.28 1.29 1.30 1.30 0.80 0.83 0.86 0.88 0.90 1.41 1.40 1.40 1.41 1.41 0.72 0.75 0.77 0.80 0.83 1.55 1.54 1.53 1.53 1.52 0.63 0.66 0.70 0.72 0.75 1.71 1.69 1.67 1.66 1.65 26 27 28 29 30 1.07 1.09 12 13 1.22 1.23 1.24 1.25 1.26 1.00 1.02 1.04 1.05 1.07 1.31 1.32 1.32 1.33 1.34 0.93 0.95 0.97 0.99 1.01 1.41 1.41 1.41 1.42 1.42 0.85 0.88 0.90 0.92 0.94 1.52 1.51 1.51 1.51 1.51 0.78 0.81 0.83 0.85 0.88 1.64 1.63 1.62 1.61 1.61 31 32 33 34 35 1.15 1.16 1.17 1.18 1.19 1.27 1.28 1.29 1.30 1.31 1.08 10 1.11 1.13 1.14 1.34 1.35 1.36 1.36 1.37 1.02 1.04 1.05 1.07 1.08 1.42 1.43 1.43 1.43 1.44 0.96 0.98 1.00 1.01 1.03 1.51 1.51 1.51 1.51 1.51 0.90 0.92 0.94 0.95 0.97 1.60 1.60 1.59 1.59 1.59 36 37 38 39 40 1.21 1.22 1.23 1.24 1.25 1.32 1.32 1.33 1.34 1.34 1.15 1.16 18 19 1.20 1.38 1.38 1.39 1.39 1.40 1.10 1.11 1.12 1.14 1.15 1.44 1.45 1.45 1.45 1.46 1.04 1.06 1.07 1.09 1.10 1.51 1.51 1.52 1.52 1.52 0.99 1.00 1.02 1.03 1.05 1.59 1.59 1.58 1.58 1.58 45 50 55 60 65 70 75 1.29 1.32 1.36 1.38 1.41 1.43 1.45 1.38 1.40 1.43 1.45 1.47 1.49 1.50 1.24 28 1.32 1.35 1.38 1.40 1.42 1.42 1.45 1.47 1.48 1.50 1.52 1.53 1.20 1.24 1.28 1.32 1.35 1.37 1.39 1.48 1.49 1.51 1.52 1.53 1.55 1.56 1.16 1.20 1.25 1.28 1.31 1.34 1.37 1.53 54 1.55 1.56 1.57 1.58 1.59 1.11 16 1.21 1.25 1.28 1.31 1.34 1.58 1.59 1.59 1.60 1.61 1.61 1.62 80 85 90 95 100 1.47 1.48 1.50 1.51 1.52 1.52 1.53 1.54 1.55 1.56 1.44 1.46 1.47 1.49 1.50 1.54 1.55 1.56 1.57 1.58 1.42 1.43 1.45 1.47 1.48 1.57 1.58 1.59 1.60 1.60 1.39 1.41 1.43 1.45 1.46 1.60 1.60 1.61 1.62 1.63 1.36 39 1.41 1.42 1.44 1.62 63 1.64 1.64 1.65 10 Source and Note: See Table B-5 615 616 APPENDIX B Table B-7: The Normal Distribution The normal distribution is usually assumed for the error term in a regression equation Table B-7 indicates the probability that a randomly drawn number from the standardized normal distribution (mean = and variance = 1) will be greater than or equal to the number identified in the side tabs, called Z For a normally distributed variable E with mean µ and variance Q 2, Z = (€ - µ)/Q The row tab gives Z to the first decimal place, and the column tab adds the second decimal place of Z The normal distribution is referred to infrequently in the text, but it does come in handy in a number of advanced settings For instance, testing for serial correlation when there is a lagged dependent variable in the equation (distributed lags) is done with a normally distributed statistic, Durbin's h statistic: h = (1 - 0.5DW) V n/(1 - n s)2,) where DW is the Durbin-Watson d statistic, n is the number of observations, and s,2, is the estimated variance of the estimated coefficient of the lagged dependent variable (Yt_ ) The h statistic is asymptotically distributed as a standard normal variable To test a one-sided null hypothesis of no positive serial correlation: Ho : p calculate h and compare it to a critical h value for the desired level of significance For a one-sided 2.5 percent test, for example, the critical h value is 1.96 as shown in the accompanying graph If we observed a computed h higher than 1.96, we would reject the null hypothesis of no positive serial correlation at the 97.5 percent level of confidence ... for a given model is a theory-based skill that is often referred to as the "a rt " of econometrics There are many alternative approaches to quantifying the same equation, and each approach may... purchases some flowers by no means implies that the bell causes purchases! If events A and B are related statistically, it may be that A causes B, that B causes A, that some omitted factor causes... Such is hardly the case in the real estate market Consequently, an impo rt ant element of every housing purchase is an appraisal of the market value of the house, and many real estate appraisers

Ngày đăng: 17/01/2020, 13:53

Từ khóa liên quan

Mục lục

  • USING ECONOMETRICS: A PRACTICAL GUIDE, 4TH ED.

  • Part I: The Basic Regression Model

    • Chapter 1. An Overview of Regression Analysis

      • 1.1 What Is Econometrics?

        • 1.1.1 Uses of Econometrics

        • 1.1.2 Alternative Econometric Approaches

        • 1.2 What Is Regression Analysis?

          • 1.2.1 Dependent Variables, Independent Variables, and Causality

          • 1.2.2 Single-Equation Linear Models

          • 1.2.3 The Stochastic Error Term

          • 1.2.4 Extending the Notation

          • 1.3 The Estimated Regression Equation

          • 1.4 A Simple Example of Regression Analysis

          • 1.5 Using Regression to Explain Housing Prices

          • 1.6 Summary

          • Exercises

          • Chapter 2. Ordinary Least Squares

            • 2.1 Estimating Single-Independent-Variable Models with OLS

              • 2.1.1 Why Use Ordinary Least Squares?

              • 2.1.2 How Does OLS Work?

              • 2.1.3 Total, Explained, and Residual Sums of Squares

              • 2.1.4 An Illustration of OLS Estimation

              • 2.2 Estimating Multivariate Regression Models with OLS

                • 2.2.1 The Meaning of Multivariate Regression Coefficients

                • 2.2.2 OLS Estimation of Multivariate Regression Models

                • 2.2.3 An Example of a Multivariate Regression Model

                • 2.3 Evaluating the Quality of a Regression Equation

Tài liệu cùng người dùng

  • Đang cập nhật ...

Tài liệu liên quan