Samiksha Jaiswal (Editor)

Linear probability model

Updated on
Edit
Like
Comment
Share on FacebookTweet on TwitterShare on LinkedInShare on Reddit

In statistics, a linear probability model is a special case of a binomial regression model. Here the dependent variable for each observation takes values which are either 0 or 1. The probability of observing a 0 or 1 in any one case is treated as depending on one or more explanatory variables. For the "linear probability model", this relationship is a particularly simple one, and allows the model to be fitted by simple linear regression.

The model assumes that, for a binary outcome (Bernoulli trial), Y , and its associated vector of explanatory variables, X ,

Pr ( Y = 1 | X = x ) = x β .

For this model,

E [ Y | X ] = Pr ( Y = 1 | X ) = x β ,

and hence the vector of parameters β can be estimated using least squares. This method of fitting would be inefficient, and can be improved by adopting an iterative scheme based on weighted least squares, in which the model from the previous iteration is used to supply estimates of the conditional variances, Var ( Y | X = x ) , which would vary between observations. This approach can be related to fitting the model by maximum likelihood.

A drawback of this model is that, unless restrictions are placed on β , the estimated coefficients can imply probabilities outside the unit interval [ 0 , 1 ] . For this reason, models such as the logit model or the probit model are more commonly used.

References

Linear probability model Wikipedia