Direct Optimization for Classification with Boosting

Bibliographic Details
Main Author: Zhai, Shaodan
Language:English
Published: Wright State University / OhioLINK 2015
Subjects:
Online Access:http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665
id ndltd-OhioLink-oai-etd.ohiolink.edu-wright1453001665
record_format oai_dc
spelling ndltd-OhioLink-oai-etd.ohiolink.edu-wright14530016652021-08-03T06:34:51Z Direct Optimization for Classification with Boosting Zhai, Shaodan Computer Science computer science Boosting, as one of the state-of-the-art classification approaches, is widely used in the industry for a broad range of problems. The existing boosting methods often formulate classification tasks as a convex optimization problem by using surrogates of performance measures. While the convex surrogates are computationally efficient to globally optimize, they are sensitive to outliers and inconsistent under some conditions. On the other hand, boosting's success can be ascribed to maximizing the margins, but few boosting approaches are designed to directly maximize the margin. In this research, we design novel boosting algorithms that directly optimize non-convex performance measures, including the empirical classification error and margin functions, without resorting to any surrogates or approximations. We first applied this approach on binary classification, and then extended this idea to more complicated classification problems, including multi-class classification, semi-supervised classification, and multi-label classification. These extensions are non-trivial, where we have to mathematically re-formulate the optimization problem: defining new objectives and designing new algorithms that depend on the specific learning tasks. Moreover, we showed good theoretical properties of the optimization objectives, which explains why we define these objectives and how we design algorithms to efficiently optimize them. Finally, we showed experimentally that the proposed approaches display competitive or better results than state-of-the-art convex relaxation boosting methods, and they perform especially well on noisy cases. 2015 English text Wright State University / OhioLINK http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665 http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665 unrestricted This thesis or dissertation is protected by copyright: all rights reserved. It may not be copied or redistributed beyond the terms of applicable copyright laws.
collection NDLTD
language English
sources NDLTD
topic Computer Science
computer science
spellingShingle Computer Science
computer science
Zhai, Shaodan
Direct Optimization for Classification with Boosting
author Zhai, Shaodan
author_facet Zhai, Shaodan
author_sort Zhai, Shaodan
title Direct Optimization for Classification with Boosting
title_short Direct Optimization for Classification with Boosting
title_full Direct Optimization for Classification with Boosting
title_fullStr Direct Optimization for Classification with Boosting
title_full_unstemmed Direct Optimization for Classification with Boosting
title_sort direct optimization for classification with boosting
publisher Wright State University / OhioLINK
publishDate 2015
url http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665
work_keys_str_mv AT zhaishaodan directoptimizationforclassificationwithboosting
_version_ 1719439874157707264