Loading…

Identifying, attributing, and overcoming common data quality issues of manned station observations

ABSTRACT In situ climatological observations are essential for studies related to climate trends and extreme events. However, in many regions of the globe, observational records are affected by a large number of data quality issues. Assessing and controlling the quality of such datasets is an import...

Full description

Saved in:
Bibliographic Details
Published in:International journal of climatology 2017-09, Vol.37 (11), p.4131-4145
Main Authors: Hunziker, Stefan, Gubler, Stefanie, Calle, Juan, Moreno, Isabel, Andrade, Marcos, Velarde, Fernando, Ticona, Laura, Carrasco, Gualberto, Castellón, Yaruska, Oria, Clara, Croci‐Maspoli, Mischa, Konzelmann, Thomas, Rohrer, Mario, Brönnimann, Stefan
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:ABSTRACT In situ climatological observations are essential for studies related to climate trends and extreme events. However, in many regions of the globe, observational records are affected by a large number of data quality issues. Assessing and controlling the quality of such datasets is an important, often overlooked aspect of climate research. Besides analysing the measurement data, metadata are important for a comprehensive data quality assessment. However, metadata are often missing, but may partly be reconstructed by suitable actions such as station inspections. This study identifies and attributes the most important common data quality issues in Bolivian and Peruvian temperature and precipitation datasets. The same or similar errors are found in many other predominantly manned station networks worldwide. A large fraction of these issues can be traced back to measurement errors by the observers. Therefore, the most effective way to prevent errors is to strengthen the training of observers and to establish a near real‐time quality control (QC) procedure. Many common data quality issues are hardly detected by usual QC approaches. Data visualization, however, is an effective tool to identify and attribute those issues, and therefore enables data users to potentially correct errors and to decide which purposes are not affected by specific problems. The resulting increase in usable station records is particularly important in areas where station networks are sparse. In such networks, adequate selection and treatment of time series based on a comprehensive QC procedure may contribute to improving data homogeneity more than statistical data homogenization methods.
ISSN:0899-8418
1097-0088
DOI:10.1002/joc.5037