Chapter 03_Discrete choice analysis_Binary outcome models tài liệu, giáo án, bài giảng , luận văn, luận án, đồ án, bài t...
Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models Chapter DISCRETE CHOICE ANALYSIS: BINARY OUTCOME MODELS I INTRODUCTION: The simplest of the model in which the dependent variable is discrete value is the model with y is binary Discrete choice model: Model in which the dependent variable assumes discrete values Example: 1 if person i is employed in labor force yi = 0 otherwise Regardless the definition of y, it is traditional to refer to y = as a “success” and y = as a “failure” Basic types of discrete values: a) Dichotomous or binary: These take on a value of one or zero, depending on which of two possible results occur b) Polychotomons variables: These take on a discrete number, greater than two, of possible values/ non-categoring: y = {number of patent issued to a company during a year} c) Unordered Variables: These are variables for which there is no natural ranking of the alternatives Example: for a sample of commuters, we define a variable: Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models 1 2 yi = 3 4 if if if if person i is a lawyer person i is a teacher person i is a doctor person i is a plumer 1 2 yi = 3 4 if if if if person i drives to work person i takes a bus person i takes a train otherwise We can define these dependent variables in any order desired → unordered categorical variables d) Ordered variables: With these variables, outcome have a natural ranking Examples: 1 if person i is in poor health yi = 2 if person i is in good health 3 if person i is in excellent health 1 2 yi = 3 4 if if if if person i spent less than $1,000 person i spent $1,000 − $2,000 person i spent $2,000 − $4000 person i spent more than $4,000 A special case of an ordered variable is a “sequential variables” This occurs when second event is dependent on the first event, the third event is dependent on the previous two event, … 1 2 yi = 3 4 if if if if person i has not completed high school person i has completed high school , not college person i has college degree, not higher degree person i has a professional degree In marketing research, one often considers attitudes of preference that are measured on a scale 1, 2, 3, 4, for instance Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II 1 2 yi = 4 5 II Chapter 3: Discrete choice analysis: Binary outcome models if intensely dislike if moderately dislike if neutral if moderately like if intensely like THE PROBABILITY MODELS: We assume that there is a latent variables y * such that yi* = X i β + ε i yi* is the additional utility that individual i would get by choosing yi = rather than yi = we not observe yi* , but we observe variable y, which takes on value of or according to the rule: 1 if y * > yi = 0 otherwise Prob ( yi = ) = Prob ( yi* >0) = Prob ( X i β + ε i >0) = Prob ( ε i > − X i β ) = – Prob ( ε i < − X i β ) = – F( − X i β ) = F( X i β ) Where F is the cumulative distribution function of ε i Prob ( yi = ) = Prob ( yi* < 0) = Prob ( X i β + ε i < 0) = Prob ( ε i < − X i β )= F (− X i β ) = - F( X i β ) The likelihood for the sample is the product of the probability of each observation Then = ∏ F (− X β ) ∏ [1 − F (− X β )] i = i yi 0= i yi i The functional form of F will depend on the assumptions made about ε i Probit Model: If the cumulative distribution of ε i is the normal distribution, i.e ε i ~ N(0, σ ) Then we have: the Probit Model Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models Xβ ε = Prob ( ε i < − X i β ) = Prob i < i σ σ F (− X i β ) X β Xβ = Φ − i = − Φ i σ σ X β − F ( − X i β ) = Φ i σ → Likelihood function y 1− yi i X i β Xiβ n Xiβ X i β = Φ ∏ ∏ 1 − Φ ∏ Φ 1 − Φ σ σ σ i yi = i1 σ = i yi = = Where: X β → Φ i σ Xiβ = z2 σ −2 e dz 2π ∫ −∞ The standard normal probability distribution function log-likelihood function ( σ) ( σ) = ln= β L β X i β X i β + (1 − yi ) ln 1 − Φ σ σ ∑ y ln Φ i i Notice that L β ≤ Because ln Φ (.) ≤ and ln[1 − Φ (.)] ≤ σ Another important note of the likelihood function is that the parameter β and σ always appear together, only the ratio β matters → convenient to normalize σ to be one → σ so we can talk just about β First derivatives ∂L( β ) = ∂β n ∑y i =1 i ϕ(Xiβ ) ϕ(Xiβ ) − (1 − yi ) − Φ( X i β ) Φ( X i β ) The Logit Model If the cumulative distribution of ε i is the logistic, we have the logit model, with the probability density function f (ε ) = Nam T Hoang UNE Business School eε eε , cdf: F ( ε ) = + eε [1 + eε ]2 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models → F (− X i β ) = exp(− X i β ) = + exp(− X i β ) + exp( X i β ) − F (− X i β ) = exp( X i β ) + exp( X i β ) Likelihood function: 1− yi yi exp( X i β ) exp( X i β ) 1 = (β ) ∏ = ∏ ∏ exp( X i β ) =i + exp( X i β ) i yi 0= += + exp( X i β ) i yi + exp( X i β ) n n = exp( β ∑ X i yi ) n i =1 ∏ [1 + exp( X β )] i =1 i Log-likelihood function: n exp( X i β ) n + ∑ (1 − yi ) ln L( β ) = ∑ yi ln i =1 + exp( X i β ) i =1 + exp( X i β ) With derivative: ∂L = ∂β ∑X i exp( X i β ) yi − + exp( X i β ) (marginal effect) Solving the maximum likelihood function: Denote φ (.) is the density function of the standard normal For the probit model: n n i =1 i =1 L = ∑ yi ln Φ ( X i β ) + ∑ (1 − yi ) ln[1 − Φ ( X i β )] S (β = ) ∂L = ∂β n [ yi − Φ ( X i β )] ∑ Φ( X β )[1 − Φ( X β )]ϕ ( X β ) X i =1 i i i i The ML estimator βˆ ML can be obtained as a solution of the equation S ( β ) = These equation are nonlinear in β , thus we have to solve them by an iterative procedure The Information matrix: Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models ∂2 L n [ϕ ( X i β )]2 = I (β ) = E− X i X i′ ∑ ∂β∂β ′ i =1 Φ ( X i β )[1 − Φ ( X i β )] We start with some initial value of β , say β , and complete the value S ( β ) and I ( β ) → new estimate of β by the method of scoring For the logit model: n n i =1 i =1 L = β ∑ X i yi −∑ ln[1 + exp( X i β )] n n exp( X i β ) ∂L S (β ) = = −∑ xi + ∑ X i yi = ∂β + exp( X i β ) = i 1= i These equation are nonlinear → we have to use the Newton – Raphson method to solve the equation Information δ 2L n exp( X i β ) = ∑ I ( β ) = E − X i X i′ ′ δβδ β β [ exp( )] + X i =1 i → starting with some initial value of β , say β , … The linear probability model: Assume: yi = X i β + ε i = β1 + β X 2i + β X 3i + + β k X ki + ε i Where yi takes values or and X ji are non – stochastic → E (ui ) = E ( yi ) = β1 + β X 2i + + β k X ki E ( yi ) = 0.P( yi = 0) + 1.P( yi = 1) = P( yi = 1) E (ε i2 ) = E[ yi − E ( yi )]2 = E ( yi2 ) − [ E ( yi )]2 = E ( yi ) − [ E ( yi )]2 = E ( yi )[1 − E ( yi )] = P( yi = 1)[1 − P( yi = 1)] → vary with i → heteroskedasticity Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models Problems of the Linear Probability Model: Heteroskedasticity → OLS will not be efficient Some cases that yi > and yi < → difficulties of interpretation (We cannot constrain Xβ to the (0, 1) interval) In many cases E ( yi X i ) can be outside the limit (0, 1) III PROBIT vs LOGIT: Why we choose logistic distribution? → Simplicity: the equation of the logistic CDF is very simple, while normal CDF involves an unevaluated integral (for multivariate this will be important) Additional: f (X ) = eX (1 + e X ) eX 1+ eX F(X ) = F (ε i < X i β )= F ( X i β )= Xiβ ∫ f ( x)dx −∞ CDF of logistic distribution F(x) Graph Distribution: lim F ( x) = x → +∞ F ( x) = xlim → −∞ f ( x) = ex (1 + e x ) F(.) is continuous & monotonous strictly x1 < x2 → f ( x1 ) < f ( x2 ) - Logistic distribution: is almost the same as normal distribution - It’s symmetric Normal distribution: F ( x) = f ( x) = − ( X −µ )2 e 2σ σ 2π x ∫ f ( x)dx −∞ Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models x2 −2 Standard normal: f ( x) = e 2π F ( x) = Φ ( x) = x ∫ −∞ IV x2 −2 e dx 2π ESTIMATION AND INFERENCE IN BINARY CHOICE MODEL: Likelihood function: Estimation of binary choice models is usually based on the method of maximum likelihood Each observation is treated as a single draw from a Bernoulli distribution, suppose F is symmetric distribution = ∏ F (− X β ) ∏ [1 − F (− X β )=] ∏ [1 − F ( X β )] ∏ F ( X β ) i = i yi 0= i yi = n ∏ [ F ( X β )] i i =1 = log = L yi i = i yi i = i yi i [1 − F ( X i β )]1− yi n ∑ {y ln F ( X β ) + (1 − y ) ln[1 − F ( X β )]} i =1 i i i i The likelihood equations are: n y f − fi ∂L = ∑ i i + (1 − y i ) X i = ∂β i =1 Fi (1 − Fi ) Where f i is the density: fi = dF ( X i β ) = f (Xiβ ) d(Xiβ ) Likelihood equation are: n y f (Xiβ ) − f (Xiβ ) ∂L = ∑ i + (1 − y i ) X i′ = ∂β i =1 F ( X i β ) (1 − F ( X i β )) ( k×1) ( k×1) ( k ×1) Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models where f ( X i β ) = dF ( X i β ) d(Xiβ ) The likelihood equation will be non-linear and require a numerical solution For the logit model: n e Xiβ ∂L = ∑ yi − Xi = ∂β i =1 (1 + e X i β ) ( k×1) ( k×1) ( k ×1) 1 1 If X i contain a constant term (1) X 1′ = 1 then 1 e Xiβ Xiβ i =1 + e n n ∑ yi = ∑ i =1 → The average of the predicted probability must equal the proportion of ones in the sample For the normal distribution: L = ∑ ln[1 − Φ ( X i β )] + ∑ ln Φ ( X i β ) yi =0 yi =1 φ(Xiβ ) − φ(Xiβ ) ∂L Xi + ∑ X i =∑ ∂β yi =0 − Φ ( X i β ) ( n×1) yi =1 Φ ( X i β ) ( n×1) X ′i is the column i of matrix X Let qi = yi − then n q φ(Xiβ ) ∂L = ∑ i X i = ∑ λi X i = ∂β i =1 Φ ( X i β ) λi The second derivatives for the logit model are simple n ∂2L e Xiβ e Xiβ 1 − = −∑ H = Xiβ Xiβ (log it ) ∂β ∂β ′ i =1 + e 1+ e X i X ′i Hessian matrix H is always negative definite, so the log – likelihood is globally concave For the probit: Nam T Hoang UNE Business School University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models n ∂2L = ∑ − λi (λi + X i β )X i X ′i ∂β∂β ′ i =1 H = ( probit ) The matrix is also negative definite for all value of β L(probit) is also globally concave Newton – Raphson method for calculating the MLE: Suppose we need to find max L(θ ), θ ( k ×1) ∂L(θ ) → U (θ ) = ∂θ ( k ×1) (Score function) ( k ×1) If the L(θ ) is concave, maximum likelihood will be obtained by solving: U (θˆ) = ( k ×1) ( k ×1) ˆ Consider expanding the score function evaluated at the MLE θ around a trial value θ o using a first order Taylor expansion, ∂U (θ o ) ˆ (θ − θ o ) U (θˆ) = U (θ o )+ ∂θ o ( k ×1) ( k ×1) ( k ×k ) ( k ×1) The Hessian matrix of the L(θ ) is: H (θ ) = ( k ×1) → H (θ o ) = ∂ L(θ ) ∂U (θ ) = ' ∂θ∂θ ( k × k ) ∂θ ∂U (θ o ) ∂θ o Set U (θˆ) = then U (θ o ) + ∂U (θ o ) ˆ (θ − θ o ) = = U (θ o ) + H (θ o )(θˆ − θ o ) = ∂θ o → H (θ o ).θˆ = H (θ o )θ o − U (θ o ) = → θˆ = θ o − H −1 (θ o ) U (θ o ) = ( k ×1) Nam T Hoang UNE Business School ( k ×1) ( k ×k ) ( k ×1) 10 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models This result provides the basis for an iterative approach for computing the MLE known as the Newton-Raphson method Give a trial value θ o - Use the equation to get an improved estimate and repeat the process with the new estimator - Stop when the differences between successive estimates are sufficiently close to zero (or elements of U (θˆ) are sufficiently close to zero) This procedure tends to converges quickly if the log-likelihood function L(θ ) is concave and if the starting value is reasonably close to the optimal value We can replace the Hessian matrix by the information matrix θˆ = θ o + I −1 (θ o )U (θ o ) ∂2L I (θ o ) = − E ' ∂θ∂θ Margianl effects • After estimating β , we can get estimated value of the probability that the i th observation is equal to ( yi = 1) Logit: Pr( yi = X i ) = exp( X i βˆ ) (i = 1, , n) + exp( X βˆ ) i Probit: Pr( yi = X i ) = ˆ Φ( X i β ) • eπ Xiβ ∫ e − Z2 dZ (i = 1, , n) −∞ The coefficients from the probit and logit models are difficult to interpret because they measure the change in the unobservable yi* associated with a change in one of the explanatory variables • A more useful measure is what we call the marginal effects: ME j = ∂ Pr( yi = X i ) ∂xij = ∂F ( X i β ) ∂xij = F ' ( X i β )β j = φ ( X i β )β j ME j = ϕ ( X i β ) β j Nam T Hoang UNE Business School 11 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models change unit of x j change of the probability of person i choose yi = • For: ME j = ∂ Pr( yi = X i ) ∂xij = eXiβ [1 + e ] Xiβ βj change unit of x j the probability of person i choose yi = change F ( X i βˆ ) = Fˆ = f ( X i βˆ ) βˆ ' ∂Fˆ ∂Fˆ VarCov( βˆ ) VarCov( Fˆ ) = ∂βˆ ˆ β ∂ An partial effects Average Partial Effects: ∂ Pr( yi = X ) APE = E x ∂x In practice: APE j = n ∑ f ( X i βˆ )βˆ j n i =1 γ k = APEk = ( j = 1, k ) ∂ Pr( yi = X i ) n ' 1 n = ∑ F ( X i β ) β k = ∑ γ k ( xi ) ∑ n ∂xik n i =1 n i =1 Let γ ij = f ( X i βˆ ) βˆ j marginal effect of xij on probability of person i takes the action Then APE j (Average Partial Effects) of xij is ( j = 1, , k ) APE j = n n ˆ )β = γ = γ ( β f X ∑ i j n∑ ij j n i =1 i =1 If xij changes one unit the average probability of an individual takes the action will change APE j = n ∑ f ( X i βˆ )β j n i=1 () [ ] Notes: V = VarCov βˆ = I ( βˆ ) Var (γ j ) = Nam T Hoang UNE Business School −1 ∂ L( βˆ ) = −E ˆ ˆ' ∂β∂β −1 1 n (γ ij − γ j ) ∑ n n − i =1 12 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models n n ∑∑ f ( X i βˆ )V [ f ( X i βˆ )]' n =i =j Var ( γ ) = ( k ×1) V is the estimated covariance matrix of βˆ Or Var ( γ ) = f ( X βˆ )V f ( X βˆ ) ' ( k ×1\ γ γ 2 . γ = . . γ k f ( Xβ ) = • n ∑ f ( Xiβ ) n i =1 Report estimation result for probit and logit: Variable coefficients (SE) marginal effects ( X ) (SF) APE (SE) Variable coefficients (SE) Average Partial Effects Hypothesis test: Wald test: For a set of restriction: Rβ = q , the statistic is ( ){ ' W = Rβˆ − q RVarCov ( βˆ ) R ' [ ] VarCov( βˆ ) = I ( βˆ ) −1 Likelihood Ratio test: [ LR = −2 ln Lˆ R − ln Lˆu } ( Rβˆ − q) ~ χ −1 ∂ L( βˆ ) = −E ˆ ˆ' ∂β∂β [J ] −1 ] EX: Test that all the slope coefficients in the probit of logit model are zero the restricted log-likelihood is the same for both probit and logit models Lo = n[P ln P + (1 − P) ln(1 − P)] P is the proportion of ( yi = 1) Lagrange Multiplier test: Nam T Hoang UNE Business School 13 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models LM = G ( βˆ )VarCov ( βˆ ) G ' ( βˆ ) ~ χ[2J ] (1× k ) ( k ×k ) ( k ×1) Where G ( βˆ ) = F ' ( Xβˆ ) evaluated at the restricted parameter vector of βˆ (1×k ) (1×k ) Specification tests for Binary Choice Models: We consider two important specification problem effect of omitted variables and the effects of heteroskedasticity In the classical model: Y = X 1β1 + X β + ε If we omit X E[ βˆ1 ] = β1 + ( X 1' X ) −1 X 1' X β Unless X and X are orthogonal or β = , βˆ1 is a biased estimator, but still consistent In the context of a binary choice model, we have: a) If x2 is omitted from a model containing x1 , x2 p lim βˆ1 = c1β1 + c2 β Where c1 , c2 are complicated function of unknown parameters the coefficient on the included variable will be inconsistent (trouble) b) If the disturbances in the regression are heteroskedasticity, then the maximum likelihood estimators are inconsistent and the covariance matrix is inappropriate (trouble) The second result is particularly troublesome because the probit and logit model is most often with micro data, which are frequently heteroskedasticity Test the omitted variables: H o : y * = X β1 + ε H : y * = X β1 + X β + ε So the test is of null hypothesis that β = Nam T Hoang UNE Business School 14 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models Test the Heteroskedasticity: Assume that the heteroskedasticity takes the form: [ ] Var (ε ) = e Zγ → σ = e Zγ Z is subset of X y * = Xβ + ε [ ] Var (ε X ) = e Zγ → ∂ Pr( yi = X ) Xβ β − ( Xβ )γ k = φ Zγ k Zγ ∂xk e e The log – likelihood is: = ln n Xiβ X i β + (1 − yi ) ln 1 − F e Z γ Z ∑ y ln F e γ i =1 i We can use the LM test construct at restricted γ = Measuring goodness of fit: There are several ways to measure the goodness of fit a) Percentage correctly predicted: For each i, computer the predicted probability that yi = If Pr( yi = X i ) > → yi = Pr( yi = X i ) < → yi = The percentage of times the predicted yi matches the actual yi the Percentage Correctly predicted Nam T Hoang UNE Business School 15 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models We can compute the percentage correctly predicted for each outcome y = and y = The overall percentage correctly predicted is a weighted average of the two, with the weights being the fractions of zero and one outcomes b) Pseudo R-squared L ln £ = Pseudo R − squared = 1− 1− Lo ln£ o Lo is the value of the log – likelihood function in the model with only intercept (hypothesis is that all β = ) L is the value of the log – likelihood function of the estimated model If all the slope coefficient are zero, then Psendo R-squared is zero L < Lo (because L is negative and L > Lo ) then we always have: < Pseudo R − squared < (Lo = n[P ln P + (1 − P) ln(1 − P)]) V BINARY CHOICE MODELS FOR PANEL DATA: yit* = X it β + ε it (1×k ) ( k ×1) (1×1) y = if and it yit = if yit* > yit* ≤ i = 1,2, , n t = 1,2, , Ti Random Effects Models: Specifies: ε it = vit + ui → yit* = X it β + vit + ui yit = if yit* > and otherwise Where vit & ui are independent random variables with E (vit X ) = , Cov(vit , v j X ) = Var (vit X ) = If i = j and t = s , otherwise Nam T Hoang UNE Business School 16 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models E (ui X ) = , Cov(ui , u j X ) = Var (ui X ) = σ u2 if i = j , otherwise Cov(vit , u j X ) ≈ for all i, t, j → E (ε it X ) = Var (ε it X ) = σ v2 + σ u2 = + σ u2 Corr (ε it , ε is X ) = ρ = σ u2 + σ u2 ρ → σ u2 = 1− ρ In the cross–section case: P( yi X i ) = − X i β (U i ) ∫ f (ε )dε i if yi = and i − ∞ ( Li ) +∞ (U i ) ∫ f (ε )dε = i if yi = i − X i β ( Li ) • The contribution of the group i to the likelihood would be the joint probability of all Ti observations Li = P( yi1 , , yiTi X ) = f (ε i1 , ε i , , ε iTi ) = U iTi U i ∫ ∫ f (ε LiTi Li +∞ Ti ∫ ∏ f (ε −∞ t =1 it i1 , ε i , , ε iTi )dε i1dε i ε iTi ui ) f (ui )dui Because joint density: f (ε i1 , ε i , , ε iTi , ui ) = f (ε i1 , ε i , , ε iTi ui ) f (ui ) So: + ∞ Ti U iTi Li = P[ yi1 , , yiTi X ] = ∫ ∏ ∫ f (ε it ui )dε it f (ui )dui − ∞ t =1 LiTi Li = P( yi1 , , yiTi X ) = +∞ Ti Pr ob(Yit = yit X it β + ui ) f (ui )dui ∫−∞∏ t =1 Likelihood function: n = ∏ Li i =1 n L = ln = ln ∏ Li = ln L1 + ln L2 + + ln Ln i =1 Nam T Hoang UNE Business School 17 University of New England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models Assume that ui is normally distributed using MLE to estimate β Fixed Effects Models: The fixed effects model is: i = 1,2, , n t = 1,2, , Ti yit* = α i dit + X it β + ε it yit = if yit* > yit = otherwise Where d it is a dummy variable that takes the value one for individual i and zero otherwise Ti n ∑∑ ln P( y = L ln= it =i =t α i + X it β ) P() is the probability of the observed outcome: For logit: eαi + X it β = Fit + eαi + X it β prob( = yit 1= X it ) Likelihood function: = ∏∏ ( F ) i it t y it (1 − Fit )1− yit Reading: 15.8.2, 15.8.3 Pooled Models: Suppose the model is: i = 1,2, , n P( yit = X it ) = F ( X it β ) t = 1,2, , Ti Log – likelihood function: L = ln = n Ti ∑∑ { y =i =t it log F ( X it β ) + (1 − yit ) log[1 − F ( X it β )]} Matlab programming work: IQ 0.045 ME β = 0.06 FE 0.04 ε ~ N (0,1) Nam T Hoang UNE Business School 18 University of New England ... 4 5 II Chapter 3: Discrete choice analysis: Binary outcome models if intensely dislike if moderately dislike if neutral if moderately like if intensely like THE PROBABILITY MODELS: We assume... - Part II Chapter 3: Discrete choice analysis: Binary outcome models x2 −2 Standard normal: f ( x) = e 2π F ( x) = Φ ( x) = x ∫ −∞ IV x2 −2 e dx 2π ESTIMATION AND INFERENCE IN BINARY CHOICE MODEL:... England Advanced Econometrics - Part II Chapter 3: Discrete choice analysis: Binary outcome models We can compute the percentage correctly predicted for each outcome y = and y = The overall percentage