id ndltd-OhioLink-oai-etd.ohiolink.edu-bgsu1594910831256966
record_format oai_dc
collection NDLTD
language English
sources NDLTD
topic Statistics
linear mixed models
variable selection and estimation in linear mixed models
non-convex optimization
approximated information criterion
spellingShingle Statistics
linear mixed models
variable selection and estimation in linear mixed models
non-convex optimization
approximated information criterion
Atutey, Olivia Abena
Linear Mixed Model Selection via Minimum Approximated Information Criterion
author Atutey, Olivia Abena
author_facet Atutey, Olivia Abena
author_sort Atutey, Olivia Abena
title Linear Mixed Model Selection via Minimum Approximated Information Criterion
title_short Linear Mixed Model Selection via Minimum Approximated Information Criterion
title_full Linear Mixed Model Selection via Minimum Approximated Information Criterion
title_fullStr Linear Mixed Model Selection via Minimum Approximated Information Criterion
title_full_unstemmed Linear Mixed Model Selection via Minimum Approximated Information Criterion
title_sort linear mixed model selection via minimum approximated information criterion
publisher Bowling Green State University / OhioLINK
publishDate 2020
url http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1594910831256966
work_keys_str_mv AT atuteyoliviaabena linearmixedmodelselectionviaminimumapproximatedinformationcriterion
_version_ 1719457363640975360
spelling ndltd-OhioLink-oai-etd.ohiolink.edu-bgsu15949108312569662021-08-03T07:15:42Z Linear Mixed Model Selection via Minimum Approximated Information Criterion Atutey, Olivia Abena Statistics linear mixed models variable selection and estimation in linear mixed models non-convex optimization approximated information criterion The analyses of correlated, repeated measures, or multilevel data with a Gaussian response are often based on models known as the linear mixed models (LMMs). LMMs are modeled using both fixed effects and random effects. The random intercepts (RI) and random intercepts and slopes (RIS) models are two exceptional cases from the linear mixed models that are taken into consideration. Our primary focus in this dissertation is to propose an approach for simultaneous selection and estimation of fixed effects only in LMMs. This dissertation, inspired by recent research of methods and criteria for model selection, aims to extend a variable selection procedure referred to as minimum approximated information criterion (MIC) of Su et al. (2018). Our contribution presents further use of the MIC for variable selection and sparse estimation in LMMs. Thus, we design a penalized log-likelihood procedure referred to as the minimum approximated information criterion for LMMs (lmmMAIC), which is used to find a parsimonious model that better generalizes data with a group structure. Our proposed lmmMAIC method enforces variable selection and sparse estimation simultaneously by adding a penalty term to the negative log-likelihood of the linear mixed model. The method differs from existing regularized methods mainly due to the penalty parameter and the penalty function.With regards to the penalty function, the lmmMAIC mimics the traditional Bayesian information criterion (BIC)-based best subset selection (BSS) method but requires a continuous or smooth approximation to the L<sub>0</sub> norm penalty of BSS. In this context, lmmMAIC performs sparse estimation by optimizing an approximated information criterion, which substantially requires approximating that L<sub>0</sub> norm penalty of BSS with a continuous unit dent function. A unit dent function, motivated by bump functions called mollifiers (Friedrichs, 1944), is an even continuous function with a [0, 1] range. Among several unit dent functions, incorporating a hyperbolic tangent function is most preferred. The approximation changes the discrete nature of the L<sub>0</sub> norm penalty of BSS to a continuous or smooth one making our method less computationally expensive. Besides, the hyperbolic tangent function has a simple form and it is much easier to compute its derivatives. This shrinkage-based method fits a linear mixed model containing all <i>p</i> predictors instead of comparing and selecting a correct sub-model across 2<sup>p</sup> candidate models. On this account, the lmmMAIC is feasible for high-dimensional data. The replacement, however, does not enforce sparsity since the hyperbolic tangent function is not singular at its origin. To better handle this issue, a reparameterization trick of the regression coefficients is needed to achieve sparsity.For a finite number of parameters, numerical investigations demonstrated by Shi and Tsai (2002) prove that traditional information criterion (IC)-based procedure like BIC can consistently identify a model. Following these suggestions of consistent variable selection and computational efficiency, we maintain the BIC fixed penalty parameter. Thus, our newly proposed procedure is free of using the frequently applied practices such as generalized cross validation (GCV) in selecting an optimal penalty parameter for our penalized likelihood framework. The lmmMAIC enjoys less computational time compared to other regularization methods.We formulate the lmmMAIC procedure as a smooth optimization problem and seek to solve for the fixed effects parameters by minimizing the penalized log-likelihood function. The implementation of the lmmMAIC involves an initial step of using the simulated annealing algorithm to obtain estimates. We proceed using these estimates as starting values by applying the modified Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm until convergence. After this step, we plug estimates obtained from the modified BFGS into the reparameterized hyperbolic tangent function to obtain our fixed effects estimates. Alternatively, the optimization of the penalized log-likelihood can be solved using generalized simulation annealing.Our research explores the performance and asymptotic properties of the lmmMAIC method by conducting extensive simulation studies using different model settings. The numerical results of our simulations for our proposed variable selection and estimation method are compared to other standard LMMs shrinkage-based methods such as Lasso, ridge, and elastic net. The results provide evidence that lmmMAIC is more consistent and efficient than the existing shrinkage-based methods under study. Furthermore, two applications with real-life examples are illustrated to evaluate the effectiveness of the lmmMAIC method. 2020-08-06 English text Bowling Green State University / OhioLINK http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1594910831256966 http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1594910831256966 unrestricted This thesis or dissertation is protected by copyright: all rights reserved. It may not be copied or redistributed beyond the terms of applicable copyright laws.