Multivariate probit model

In statistics and econometrics, the multivariate probit model is a generalization of the probit model used to estimate several correlated binary outcomes jointly. For example, if it is believed that the decisions of sending at least one child to public school and that of voting in favor of a school budget are correlated (both decisions are binary), then the multivariate probit model would be appropriate for jointly predicting these two choices on an individual-specific basis.

Example: bivariate probit

In the ordinary probit model, there is only one binary dependent variable Y and so only one latent variable Y^* is used. In contrast, in the bivariate probit model there are two binary dependent variables Y_1 and Y_2, so there are two latent variables:  Y^*_1 and  Y^*_2 . It is assumed that each observed variable takes on the value 1 if and only if its underlying continuous latent variable takes on a positive value:


Y_1 = \begin{cases} 1 & \text{if }Y^*_1>0, \\
0 & \text{otherwise},
\end{cases}

Y_2 = \begin{cases}
1 & \text{if }Y^*_2>0, \\
0 & \text{otherwise},
\end{cases}

with


\begin{cases}
Y_1^* = X_1\beta_1+\varepsilon_1 \\
Y_2^* = X_2\beta_2+\varepsilon_2
\end{cases}

and


\begin{bmatrix}
\varepsilon_1\\
\varepsilon_2
\end{bmatrix}
\mid X
\sim \mathcal{N}
\left(
\begin{bmatrix}
0\\
0
\end{bmatrix},
\begin{bmatrix}
1&\rho\\
\rho&1
\end{bmatrix}
\right)

Fitting the bivariate probit model involves estimating the values of \beta_1,\ \beta_2, and  \rho . To do so, the likelihood of the model has to be maximized. This likelihood is


\begin{align}
L(\beta_1,\beta_2) =\Big( \prod & P(Y_1=1,Y_2=1\mid\beta_1,\beta_2)^{Y_1Y_2} P(Y_1=0,Y_2=1\mid\beta_1,\beta_2)^{(1-Y_1)Y_2} \\[8pt]
& {}\qquad P(Y_1=1,Y_2=0\mid\beta_1,\beta_2)^{Y_1(1-Y_2)}
P(Y_1=0,Y_2=0\mid\beta_1,\beta_2)^{(1-Y_1)(1-Y_2)} \Big)
\end{align}

Substituting the latent variables Y_1^* and Y_2^* in the probability functions and taking logs gives


\begin{align}
\sum & \Big( Y_1Y_2 \ln P(\varepsilon_1>-X_1\beta_1,\varepsilon_2>-X_2\beta_2) \\[4pt]
& {}\quad{}+(1-Y_1)Y_2\ln P(\varepsilon_1<-X_1\beta_1,\varepsilon_2>-X_2\beta_2) \\[4pt]
& {}\quad{}+Y_1(1-Y_2)\ln P(\varepsilon_1>-X_1\beta_1,\varepsilon_2<-X_2\beta_2) \\[4pt]
& {}\quad{}+(1-Y_1)(1-Y_2)\ln P(\varepsilon_1<-X_1\beta_1,\varepsilon_2<-X_2\beta_2) \Big).
\end{align}

After some rewriting, the log-likelihood function becomes:


\begin{align}
\sum & \Big ( Y_1Y_2\ln \Phi(X_1\beta_1,X_2\beta_2,\rho) \\[4pt]
& {}\quad{} + (1-Y_1)Y_2\ln \Phi(-X_1\beta_1,X_2\beta_2,-\rho) \\[4pt]
& {}\quad{} + Y_1(1-Y_2)\ln \Phi(X_1\beta_1,-X_2\beta_2,-\rho) \\[4pt]
& {}\quad{} +(1-Y_1)(1-Y_2)\ln \Phi(-X_1\beta_1,-X_2\beta_2,\rho) \Big).
\end{align}

Note that  \Phi is the cumulative distribution function of the bivariate normal distribution.  Y_1 and  Y_2 in the log-likelihood function are observed variables being equal to one or zero.

Further reading

Greene, William H., Econometric Analysis, seventh edition, Prentice-Hall, 2012.

This article is issued from Wikipedia - version of the Saturday, November 29, 2014. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.