Minimum Description Length Model Selection Criteria for Generalized Linear Models

نویسندگان

  • Mark Hansen
  • Bin Yu
چکیده

This paper derives several model selection criteria for generalized linear models (GLMs) following the principle of Minimum Description Length (MDL). We focus our attention on the mixture form of MDL. Normal or normal-inverse gamma distributions are used to construct the mixtures, depending on whether or not we choose to account for possible over-dispersion in the data. For the latter, we use Efron's (1986) double exponential family characterization of GLMs as the likelihood. Standard Laplace approximations are employed to derive computationally tractable selection rules. Each of the criteria we construct have adaptive penalties on model complexity, either explicitly or implicitly. Theoretical results for the normal linear model, and a set of simulations for logistic regression illustrate that mixture MDL can \bridge" AIC and BIC in the sense that it can mimic the performance of either criterion, depending on which is best for the situation at hand.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

The MDL model choice for linear regression

In this talk, we discuss the principle of Minimum Description Length (MDL) for problems of statistical modeling. By viewing models as a means of providing statistical descriptions of observed data, the comparison between competing models is based on the stochastic complexity (SC) of each description. The Normalized Maximum Likelihood (NML) form of the SC (Rissanen 1996) contains a component tha...

متن کامل

Model Selection and the Principle of Minimum Description Length

This paper reviews the principle of Minimum Description Length (MDL) for problems of model selection. By viewing statistical modeling as a means of generating descriptions of observed data, the MDL framework discriminates between competing models based on the complexity of each description. This approach began with Kolmogorov’s theory of algorithmic complexity, matured in the literature on info...

متن کامل

Model Selection using Information Theory and the MDL Principle ∗

Information theory offers a coherent, intuitive view of model selection. This perspective arises from thinking of a statistical model as a code, an algorithm for compressing data into a sequence of bits. The description length is the length of this code for the data plus the length of a description of the model itself. The length of the code for the data measures the fit of the model to the dat...

متن کامل

Minimum Message Length Ridge Regression for Generalized Linear Models

This paper introduces an information theoretic model selection and ridge parameter estimation criterion for generalized linear models based on the minimum message length principle. The criterion is highly general in nature, and handles a range of target distributions, including the normal, binomial, Poisson, geometric and gamma distributions. Estimation of the regression parameters, the ridge h...

متن کامل

Stochastic Complexity and Its Applications in Signal Processing: Some New Results

We use the linear model (LM) as a vehicle to illustrate the relationship between the generalized likelihood ratio test and two detection strategies which are derived from the Minimum Description Length principle. The reason for choosing the LM is twofold: (i) it is frequently used in signal processing; (ii) some derivations are simpler and the results are more intuitive than those obtained for ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2002