Handbook of Regression Analysis With Applications in R. Samprit Chatterjee

Чтение книги онлайн.

Читать онлайн книгу Handbook of Regression Analysis With Applications in R - Samprit Chatterjee страница 23

Handbook of Regression Analysis With Applications in R - Samprit  Chatterjee

Скачать книгу

predictors in the model by images, rather than by images (which means that using images will tend to lead to more complex models than using images will). This suggests another model selection rule:

      1 Choose the model that minimizes . In case of tied values, the simplest model (smallest ) would be chosen. In these data, this rule implies choosing .

      An additional operational rule for the use of images has been suggested. When a particular model contains all of the necessary predictors, the residual mean square for the model should be roughly equal to images. Since the model that includes all of the predictors should also include all of the necessary ones, images should also be roughly equal to images. This implies that if a model includes all of the necessary predictors, then

equation

      This suggests the following model selection rule:

      1 Choose the simplest model such that or smaller. In these data, this rule implies choosing .

      A weakness of the images criterion is that its value depends on the largest set of candidate predictors (through images), which means that adding predictors that provide no predictive power to the set of candidate models can change the choice of best model. A general approach that avoids this is through the use of statistical information. A detailed discussion of the determination of information measures is beyond the scope of this book, but Burnham and Anderson (2002) provides extensive discussion of the topic. The Akaike Information Criterion images, introduced by Akaike (1973),

      1 Choose the model that minimizes . In case of tied values, the simplest model (smallest ) would be chosen. In these data, this rule implies choosing , although the value for is virtually identical to that of . Note that the overall level of the values is not meaningful, and should not be compared to values or values for other data sets; it is only the value for a model for a given data set relative to the values of others for that data set that matter.

      images, images, and images have the desirable property that they are efficient model selection criteria. This means that in the (realistic) situation where the set of candidate models does not include the “true” model (that is, a good model is just viewed as a useful approximation to reality), as the sample gets larger the error obtained in making predictions using the model chosen using these criteria becomes indistinguishable from the error obtained using the best possible model among all candidate models. That is, in this large‐sample predictive sense, it is as if the best approximation was known to the data analyst. Another well‐known criterion, the Bayesian Information Criterion images [which substitutes images for images in (2.2)], does not have this property, but is instead a consistent criterion. Such a criterion has the property that if the “true” model is in fact among the candidate models the criterion will select that model with probability approaching images as the sample size increases. Thus, images is a more natural criterion to use if the goal is to identify the “true” predictors with nonzero slopes (which of course presumes that there are such things as “true” predictors in a “true” model). images will generally choose simpler models than images because of its stronger penalty (images for images), and a version images that adjusts

Скачать книгу