Direct Optimization for Classification with Boosting
Main Author: | |
---|---|
Language: | English |
Published: |
Wright State University / OhioLINK
2015
|
Subjects: | |
Online Access: | http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665 |
id |
ndltd-OhioLink-oai-etd.ohiolink.edu-wright1453001665 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-OhioLink-oai-etd.ohiolink.edu-wright14530016652021-08-03T06:34:51Z Direct Optimization for Classification with Boosting Zhai, Shaodan Computer Science computer science Boosting, as one of the state-of-the-art classification approaches, is widely used in the industry for a broad range of problems. The existing boosting methods often formulate classification tasks as a convex optimization problem by using surrogates of performance measures. While the convex surrogates are computationally efficient to globally optimize, they are sensitive to outliers and inconsistent under some conditions. On the other hand, boosting's success can be ascribed to maximizing the margins, but few boosting approaches are designed to directly maximize the margin. In this research, we design novel boosting algorithms that directly optimize non-convex performance measures, including the empirical classification error and margin functions, without resorting to any surrogates or approximations. We first applied this approach on binary classification, and then extended this idea to more complicated classification problems, including multi-class classification, semi-supervised classification, and multi-label classification. These extensions are non-trivial, where we have to mathematically re-formulate the optimization problem: defining new objectives and designing new algorithms that depend on the specific learning tasks. Moreover, we showed good theoretical properties of the optimization objectives, which explains why we define these objectives and how we design algorithms to efficiently optimize them. Finally, we showed experimentally that the proposed approaches display competitive or better results than state-of-the-art convex relaxation boosting methods, and they perform especially well on noisy cases. 2015 English text Wright State University / OhioLINK http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665 http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665 unrestricted This thesis or dissertation is protected by copyright: all rights reserved. It may not be copied or redistributed beyond the terms of applicable copyright laws. |
collection |
NDLTD |
language |
English |
sources |
NDLTD |
topic |
Computer Science computer science |
spellingShingle |
Computer Science computer science Zhai, Shaodan Direct Optimization for Classification with Boosting |
author |
Zhai, Shaodan |
author_facet |
Zhai, Shaodan |
author_sort |
Zhai, Shaodan |
title |
Direct Optimization for Classification with Boosting |
title_short |
Direct Optimization for Classification with Boosting |
title_full |
Direct Optimization for Classification with Boosting |
title_fullStr |
Direct Optimization for Classification with Boosting |
title_full_unstemmed |
Direct Optimization for Classification with Boosting |
title_sort |
direct optimization for classification with boosting |
publisher |
Wright State University / OhioLINK |
publishDate |
2015 |
url |
http://rave.ohiolink.edu/etdc/view?acc_num=wright1453001665 |
work_keys_str_mv |
AT zhaishaodan directoptimizationforclassificationwithboosting |
_version_ |
1719439874157707264 |