Presentation
A simple approach to compute generalized residuals for nonlinear models
Arnab Bhattacharjee
Friday 12th September
Session
In models where the relationship between the outcome and the error term is linear, a residual can be computed by simply plugging-in the estimated coefficients and computing the difference between observed and predicted values of the outcome variable. These residuals can then be used for many different purposes, for example: (a) evaluating assumptions of orthogonality of errors (like, fixed and random effects); (b) examining the entire shape of the error distribution; and (c) computation and inference on externalities such as network effects. However, this simple approach does not work when the model is nonlinear in outcomes and errors. Here, different context-specific generalized residuals have been proposed, each having different properties for specific models. Note that, for the canonical linear or nonlinear Gaussian regression model, the above construction is simply a scaled version of the partial derivative of the log-likelihood contribution of an individual observation with respect to the outcome variable. This suggests a general construction of generalized regression by perturbing the outcome variable and computing contrasts. This approach is closely related to Huber's influence function and can be routinely computed using Stata for example and also parallelized for large datasets. We propose this general construction of generalized residuals and evaluate its use in several contexts: (a) quantile regression and evaluation of conditional quantiles at the tails (for example, growth at risk); (b) computing errors distributions (for example, binary regression and random effects models); and (c) computing network externalities in discrete choice and duration models. This delivers a unified approach with promising findings.
Validate your login
Assinar em
Criar nova conta