Detecting outliers using the Least Square Regression Line that is forced through an observation

碩士 === 逢甲大學 === 統計與精算所 === 95 === Cook''s Distance are commonly used to detect outliers. Although only has one outlier its diagnostic effect is very good, but if has more than one outlier, it easily has the situation because of masking effect which sentences by mistake. This article propos...

Full description

Bibliographic Details
Main Authors: Ming-tan Hsieh, 謝明潭
Other Authors: Jung-Pin Wu
Format: Others
Language:zh-TW
Published: 2007
Online Access:http://ndltd.ncl.edu.tw/handle/44467562856032893098
id ndltd-TW-095FCU05336019
record_format oai_dc
spelling ndltd-TW-095FCU053360192015-10-13T11:31:40Z http://ndltd.ncl.edu.tw/handle/44467562856032893098 Detecting outliers using the Least Square Regression Line that is forced through an observation 利用強迫通過觀察值的最小平方迴歸線診斷異常值 Ming-tan Hsieh 謝明潭 碩士 逢甲大學 統計與精算所 95 Cook''s Distance are commonly used to detect outliers. Although only has one outlier its diagnostic effect is very good, but if has more than one outlier, it easily has the situation because of masking effect which sentences by mistake. This article proposes new method of detect outlier. We use Least Square Method and Lagrange Multiplier Method to solve the Least Square Regression Line that is forced through an observation, and solve the Least Square Regression Line that is deleted above observation. Then calculates the angle of two straight lines, the angle is used to judge this observation whether is outlier. It''s not easy to calculate the sampling distribution of angle, therefore we use bootstrap to simulate its sampling distribution so as to estimate the p value of angle. If p value smaller than the probability of type I error α, this observation is recognized for outlier. The new technique that we suggested is compared with several traditional diagnosis method (Cook''s Distance, H, DFFITS, DFBETAS and COVRATIO) penetrates Monte Carlo simulation. Using Positive and False Positive to ponder the quality of diagnosis method. Positive is the ratio that the true outlier not be found. False Positive is the ratio that a good observation was mistaken for the outlier. Jung-Pin Wu 吳榮彬 2007 學位論文 ; thesis 60 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 逢甲大學 === 統計與精算所 === 95 === Cook''s Distance are commonly used to detect outliers. Although only has one outlier its diagnostic effect is very good, but if has more than one outlier, it easily has the situation because of masking effect which sentences by mistake. This article proposes new method of detect outlier. We use Least Square Method and Lagrange Multiplier Method to solve the Least Square Regression Line that is forced through an observation, and solve the Least Square Regression Line that is deleted above observation. Then calculates the angle of two straight lines, the angle is used to judge this observation whether is outlier. It''s not easy to calculate the sampling distribution of angle, therefore we use bootstrap to simulate its sampling distribution so as to estimate the p value of angle. If p value smaller than the probability of type I error α, this observation is recognized for outlier. The new technique that we suggested is compared with several traditional diagnosis method (Cook''s Distance, H, DFFITS, DFBETAS and COVRATIO) penetrates Monte Carlo simulation. Using Positive and False Positive to ponder the quality of diagnosis method. Positive is the ratio that the true outlier not be found. False Positive is the ratio that a good observation was mistaken for the outlier.
author2 Jung-Pin Wu
author_facet Jung-Pin Wu
Ming-tan Hsieh
謝明潭
author Ming-tan Hsieh
謝明潭
spellingShingle Ming-tan Hsieh
謝明潭
Detecting outliers using the Least Square Regression Line that is forced through an observation
author_sort Ming-tan Hsieh
title Detecting outliers using the Least Square Regression Line that is forced through an observation
title_short Detecting outliers using the Least Square Regression Line that is forced through an observation
title_full Detecting outliers using the Least Square Regression Line that is forced through an observation
title_fullStr Detecting outliers using the Least Square Regression Line that is forced through an observation
title_full_unstemmed Detecting outliers using the Least Square Regression Line that is forced through an observation
title_sort detecting outliers using the least square regression line that is forced through an observation
publishDate 2007
url http://ndltd.ncl.edu.tw/handle/44467562856032893098
work_keys_str_mv AT mingtanhsieh detectingoutliersusingtheleastsquareregressionlinethatisforcedthroughanobservation
AT xièmíngtán detectingoutliersusingtheleastsquareregressionlinethatisforcedthroughanobservation
AT mingtanhsieh lìyòngqiángpòtōngguòguāncházhídezuìxiǎopíngfānghuíguīxiànzhěnduànyìchángzhí
AT xièmíngtán lìyòngqiángpòtōngguòguāncházhídezuìxiǎopíngfānghuíguīxiànzhěnduànyìchángzhí
_version_ 1716845272780242944