Detecting outliers using the Least Square Regression Line that is forced through an observation
碩士 === 逢甲大學 === 統計與精算所 === 95 === Cook''s Distance are commonly used to detect outliers. Although only has one outlier its diagnostic effect is very good, but if has more than one outlier, it easily has the situation because of masking effect which sentences by mistake. This article propos...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2007
|
Online Access: | http://ndltd.ncl.edu.tw/handle/44467562856032893098 |
id |
ndltd-TW-095FCU05336019 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-095FCU053360192015-10-13T11:31:40Z http://ndltd.ncl.edu.tw/handle/44467562856032893098 Detecting outliers using the Least Square Regression Line that is forced through an observation 利用強迫通過觀察值的最小平方迴歸線診斷異常值 Ming-tan Hsieh 謝明潭 碩士 逢甲大學 統計與精算所 95 Cook''s Distance are commonly used to detect outliers. Although only has one outlier its diagnostic effect is very good, but if has more than one outlier, it easily has the situation because of masking effect which sentences by mistake. This article proposes new method of detect outlier. We use Least Square Method and Lagrange Multiplier Method to solve the Least Square Regression Line that is forced through an observation, and solve the Least Square Regression Line that is deleted above observation. Then calculates the angle of two straight lines, the angle is used to judge this observation whether is outlier. It''s not easy to calculate the sampling distribution of angle, therefore we use bootstrap to simulate its sampling distribution so as to estimate the p value of angle. If p value smaller than the probability of type I error α, this observation is recognized for outlier. The new technique that we suggested is compared with several traditional diagnosis method (Cook''s Distance, H, DFFITS, DFBETAS and COVRATIO) penetrates Monte Carlo simulation. Using Positive and False Positive to ponder the quality of diagnosis method. Positive is the ratio that the true outlier not be found. False Positive is the ratio that a good observation was mistaken for the outlier. Jung-Pin Wu 吳榮彬 2007 學位論文 ; thesis 60 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 逢甲大學 === 統計與精算所 === 95 === Cook''s Distance are commonly used to detect outliers.
Although only has one outlier its diagnostic effect is very good,
but if has more than one outlier,
it easily has the situation because of masking effect which sentences by mistake.
This article proposes new method of detect outlier.
We use Least Square Method and Lagrange Multiplier Method to solve the Least Square Regression Line that is forced through an observation,
and solve the Least Square Regression Line that is deleted above observation.
Then calculates the angle of two straight lines,
the angle is used to judge this observation whether is outlier.
It''s not easy to calculate the sampling distribution of angle,
therefore we use bootstrap to simulate its sampling distribution so as to estimate the p value of angle.
If p value smaller than the probability of type I error α,
this observation is recognized for outlier.
The new technique that we suggested is compared with several traditional diagnosis method (Cook''s Distance, H, DFFITS, DFBETAS and COVRATIO) penetrates Monte Carlo simulation.
Using Positive and False Positive to ponder the quality of diagnosis method.
Positive is the ratio that the true outlier not be found.
False Positive is the ratio that a good observation was mistaken for the outlier.
|
author2 |
Jung-Pin Wu |
author_facet |
Jung-Pin Wu Ming-tan Hsieh 謝明潭 |
author |
Ming-tan Hsieh 謝明潭 |
spellingShingle |
Ming-tan Hsieh 謝明潭 Detecting outliers using the Least Square Regression Line that is forced through an observation |
author_sort |
Ming-tan Hsieh |
title |
Detecting outliers using the Least Square Regression Line that is forced through an observation |
title_short |
Detecting outliers using the Least Square Regression Line that is forced through an observation |
title_full |
Detecting outliers using the Least Square Regression Line that is forced through an observation |
title_fullStr |
Detecting outliers using the Least Square Regression Line that is forced through an observation |
title_full_unstemmed |
Detecting outliers using the Least Square Regression Line that is forced through an observation |
title_sort |
detecting outliers using the least square regression line that is forced through an observation |
publishDate |
2007 |
url |
http://ndltd.ncl.edu.tw/handle/44467562856032893098 |
work_keys_str_mv |
AT mingtanhsieh detectingoutliersusingtheleastsquareregressionlinethatisforcedthroughanobservation AT xièmíngtán detectingoutliersusingtheleastsquareregressionlinethatisforcedthroughanobservation AT mingtanhsieh lìyòngqiángpòtōngguòguāncházhídezuìxiǎopíngfānghuíguīxiànzhěnduànyìchángzhí AT xièmíngtán lìyòngqiángpòtōngguòguāncházhídezuìxiǎopíngfānghuíguīxiànzhěnduànyìchángzhí |
_version_ |
1716845272780242944 |