Estimation of discriminant analysis error rate for high dimensional data
Methodologies for data reduction, modeling, and classification of grouped response curves are explored. In particular, the thesis focuses on the analysis of a collection of highly correlated, highly dimensional response-curve data of spectral reflectance curves of wood surface features. In the analy...
Main Author: | |
---|---|
Other Authors: | |
Language: | en_US |
Published: |
2012
|
Subjects: | |
Online Access: | http://hdl.handle.net/1957/35903 |
id |
ndltd-ORGSU-oai-ir.library.oregonstate.edu-1957-35903 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-ORGSU-oai-ir.library.oregonstate.edu-1957-359032012-12-29T03:14:03ZEstimation of discriminant analysis error rate for high dimensional dataLebow, Patricia K.Discriminant analysisMethodologies for data reduction, modeling, and classification of grouped response curves are explored. In particular, the thesis focuses on the analysis of a collection of highly correlated, highly dimensional response-curve data of spectral reflectance curves of wood surface features. In the analysis, questions about the application of cross-validation estimation of discriminant function error rates for data that has been previously transformed by principal component analysis arise. Performing cross-validation requires re-calculating the principal component transformation and discriminant functions of the training sets, a very lengthy process. A more efficient approach of carrying out the cross-validation calculations, plus the alternative of estimating error rates without the re-calculation of the principal component decomposition, are studied to address questions about the cross-validation procedure. If populations are assumed to have common covariance structures, the pooled covariance matrix can be decomposed for the principal component transformation. The leave-one-out cross-validation procedure results in a rank-one update in the pooled covariance matrix for each observation left out. Algorithms have been developed for calculating the updated eigenstructure under rank-one updates and they can be applied to the orthogonal decomposition of the pooled covariance matrix. Use of these algorithms results in much faster computation of error rates, especially when the number of variables is large. The bias and variance of an estimator that performs leave-one-out cross-validation directly on the principal component scores (without re-computation of the principal component transformation for each observation) is also investigated.Graduation date: 1993Butler, David A.Thomas, David R.2012-12-28T22:11:33Z2012-12-28T22:11:33Z1992-10-231992-10-23Thesis/Dissertationhttp://hdl.handle.net/1957/35903en_US |
collection |
NDLTD |
language |
en_US |
sources |
NDLTD |
topic |
Discriminant analysis |
spellingShingle |
Discriminant analysis Lebow, Patricia K. Estimation of discriminant analysis error rate for high dimensional data |
description |
Methodologies for data reduction, modeling, and classification of grouped
response curves are explored. In particular, the thesis focuses on the analysis of
a collection of highly correlated, highly dimensional response-curve data of
spectral reflectance curves of wood surface features.
In the analysis, questions about the application of cross-validation
estimation of discriminant function error rates for data that has been previously
transformed by principal component analysis arise. Performing cross-validation
requires re-calculating the principal component transformation and discriminant
functions of the training sets, a very lengthy process. A more efficient approach
of carrying out the cross-validation calculations, plus the alternative of
estimating error rates without the re-calculation of the principal component
decomposition, are studied to address questions about the cross-validation
procedure.
If populations are assumed to have common covariance structures, the
pooled covariance matrix can be decomposed for the principal component
transformation. The leave-one-out cross-validation procedure results in a rank-one
update in the pooled covariance matrix for each observation left out.
Algorithms have been developed for calculating the updated eigenstructure
under rank-one updates and they can be applied to the orthogonal
decomposition of the pooled covariance matrix. Use of these algorithms results
in much faster computation of error rates, especially when the number of
variables is large.
The bias and variance of an estimator that performs leave-one-out cross-validation
directly on the principal component scores (without re-computation
of the principal component transformation for each observation) is also
investigated. === Graduation date: 1993 |
author2 |
Butler, David A. |
author_facet |
Butler, David A. Lebow, Patricia K. |
author |
Lebow, Patricia K. |
author_sort |
Lebow, Patricia K. |
title |
Estimation of discriminant analysis error rate for high dimensional data |
title_short |
Estimation of discriminant analysis error rate for high dimensional data |
title_full |
Estimation of discriminant analysis error rate for high dimensional data |
title_fullStr |
Estimation of discriminant analysis error rate for high dimensional data |
title_full_unstemmed |
Estimation of discriminant analysis error rate for high dimensional data |
title_sort |
estimation of discriminant analysis error rate for high dimensional data |
publishDate |
2012 |
url |
http://hdl.handle.net/1957/35903 |
work_keys_str_mv |
AT lebowpatriciak estimationofdiscriminantanalysiserrorrateforhighdimensionaldata |
_version_ |
1716393845408661504 |