On statistical approaches to climate change analysis

Evidence for a human contribution to climatic changes during the past century is accumulating rapidly. Given the strength of the evidence, it seems natural to ask whether forcing projections can be used to forecast climate change. A Bayesian method for post-processing forced climate model simula...

Full description

Bibliographic Details
Main Author: Lee, Terry Chun Kit
Other Authors: Tsao, Min
Language:English
en
Published: 2008
Subjects:
Online Access:http://hdl.handle.net/1828/877
id ndltd-uvic.ca-oai-dspace.library.uvic.ca-1828-877
record_format oai_dc
spelling ndltd-uvic.ca-oai-dspace.library.uvic.ca-1828-8772015-01-29T16:50:27Z On statistical approaches to climate change analysis Lee, Terry Chun Kit Tsao, Min Zwiers, Francis Climate change Climate prediction Paleoclimate reconstruction State-space model Kalman filter and smoother Maximum likelihood estimation Asymptotic distribution UVic Subject Index::Sciences and Engineering::Mathematics::Mathematical statistics Evidence for a human contribution to climatic changes during the past century is accumulating rapidly. Given the strength of the evidence, it seems natural to ask whether forcing projections can be used to forecast climate change. A Bayesian method for post-processing forced climate model simulations that produces probabilistic hindcasts of inter-decadal temperature changes on large spatial scales is proposed. Hindcasts produced for the last two decades of the 20th century are shown to be skillful. The suggestion that skillful decadal forecasts can be produced on large regional scales by exploiting the response to anthropogenic forcing provides additional evidence that anthropogenic change in the composition of the atmosphere has influenced our climate. In the absence of large negative volcanic forcing on the climate system (which cannot presently be forecast), the global mean temperature for the decade 2000-2009 is predicted to lie above the 1970-1999 normal with probability 0.94. The global mean temperature anomaly for this decade relative to 1970-1999 is predicted to be 0.35C (5-95% confidence range: 0.21C-0.48C). Reconstruction of temperature variability of the past centuries using climate proxy data can also provide important information on the role of anthropogenic forcing in the observed 20th century warming. A state-space model approach that allows incorporation of additional non-temperature information, such as the estimated response to external forcing, to reconstruct historical temperature is proposed. An advantage of this approach is that it permits simultaneous reconstruction and detection analysis as well as future projection. A difficulty in using this approach is that estimation of several unknown state-space model parameters is required. To take advantage of the data structure in the reconstruction problem, the existing parameter estimation approach is modified, resulting in two new estimation approaches. The competing estimation approaches are compared based on theoretical grounds and through simulation studies. The two new estimation approaches generally perform better than the existing approach. A number of studies have attempted to reconstruct hemispheric mean temperature for the past millennium from proxy climate indicators. Different statistical methods are used in these studies and it therefore seems natural to ask which method is more reliable. An empirical comparison between the different reconstruction methods is considered using both climate model data and real-world paleoclimate proxy data. The proposed state-space model approach and the RegEM method generally perform better than their competitors when reconstructing interannual variations in Northern Hemispheric mean surface air temperature. On the other hand, a variety of methods are seen to perform well when reconstructing decadal temperature variability. The similarity in performance provides evidence that the difference between many real-world reconstructions is more likely to be due to the choice of the proxy series, or the use of difference target seasons or latitudes, than to the choice of statistical method. 2008-04-21T16:40:24Z 2008-04-21T16:40:24Z 2008 2008-04-21T16:40:24Z Thesis http://hdl.handle.net/1828/877 Lee, T.C.K., F. Zwiers, X. Zhang and M. Tsao, 2006: Evidence of decadal climate prediction skill resulting from changes in anthropogenic forcing. Journal of climate, 19, 5305-5318. Lee, T.C.K., F. Zwiers and M. Tsao, 2008: Evaluation of proxy-based millennial reconstruction methods. Climate Dynamics, in press. English en Available to the World Wide Web
collection NDLTD
language English
en
sources NDLTD
topic Climate change
Climate prediction
Paleoclimate reconstruction
State-space model
Kalman filter and smoother
Maximum likelihood estimation
Asymptotic distribution
UVic Subject Index::Sciences and Engineering::Mathematics::Mathematical statistics
spellingShingle Climate change
Climate prediction
Paleoclimate reconstruction
State-space model
Kalman filter and smoother
Maximum likelihood estimation
Asymptotic distribution
UVic Subject Index::Sciences and Engineering::Mathematics::Mathematical statistics
Lee, Terry Chun Kit
On statistical approaches to climate change analysis
description Evidence for a human contribution to climatic changes during the past century is accumulating rapidly. Given the strength of the evidence, it seems natural to ask whether forcing projections can be used to forecast climate change. A Bayesian method for post-processing forced climate model simulations that produces probabilistic hindcasts of inter-decadal temperature changes on large spatial scales is proposed. Hindcasts produced for the last two decades of the 20th century are shown to be skillful. The suggestion that skillful decadal forecasts can be produced on large regional scales by exploiting the response to anthropogenic forcing provides additional evidence that anthropogenic change in the composition of the atmosphere has influenced our climate. In the absence of large negative volcanic forcing on the climate system (which cannot presently be forecast), the global mean temperature for the decade 2000-2009 is predicted to lie above the 1970-1999 normal with probability 0.94. The global mean temperature anomaly for this decade relative to 1970-1999 is predicted to be 0.35C (5-95% confidence range: 0.21C-0.48C). Reconstruction of temperature variability of the past centuries using climate proxy data can also provide important information on the role of anthropogenic forcing in the observed 20th century warming. A state-space model approach that allows incorporation of additional non-temperature information, such as the estimated response to external forcing, to reconstruct historical temperature is proposed. An advantage of this approach is that it permits simultaneous reconstruction and detection analysis as well as future projection. A difficulty in using this approach is that estimation of several unknown state-space model parameters is required. To take advantage of the data structure in the reconstruction problem, the existing parameter estimation approach is modified, resulting in two new estimation approaches. The competing estimation approaches are compared based on theoretical grounds and through simulation studies. The two new estimation approaches generally perform better than the existing approach. A number of studies have attempted to reconstruct hemispheric mean temperature for the past millennium from proxy climate indicators. Different statistical methods are used in these studies and it therefore seems natural to ask which method is more reliable. An empirical comparison between the different reconstruction methods is considered using both climate model data and real-world paleoclimate proxy data. The proposed state-space model approach and the RegEM method generally perform better than their competitors when reconstructing interannual variations in Northern Hemispheric mean surface air temperature. On the other hand, a variety of methods are seen to perform well when reconstructing decadal temperature variability. The similarity in performance provides evidence that the difference between many real-world reconstructions is more likely to be due to the choice of the proxy series, or the use of difference target seasons or latitudes, than to the choice of statistical method.
author2 Tsao, Min
author_facet Tsao, Min
Lee, Terry Chun Kit
author Lee, Terry Chun Kit
author_sort Lee, Terry Chun Kit
title On statistical approaches to climate change analysis
title_short On statistical approaches to climate change analysis
title_full On statistical approaches to climate change analysis
title_fullStr On statistical approaches to climate change analysis
title_full_unstemmed On statistical approaches to climate change analysis
title_sort on statistical approaches to climate change analysis
publishDate 2008
url http://hdl.handle.net/1828/877
work_keys_str_mv AT leeterrychunkit onstatisticalapproachestoclimatechangeanalysis
_version_ 1716728960645070848