INTERNATIONAL JOURNAL OF CLIMATOLOGY Int. J. Climatol. (2014) Published online in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/joc.4216 Skill of CMIP5 climate models in reproducing 20th century basic climate features in Central America Hugo G. Hidalgoa,b* and Eric J. Alfaroa,b,c a School of Physics, University of Costa Rica, San Pedro, Costa Rica Center for Geophysical Research, University of Costa Rica, San Pedro, Costa Rica c Center for Research in Marine Sciences and Limnology, University of Costa Rica, San Pedro, Costa Rica b ABSTRACT: A total of 107 climate runs from 48 Coupled Model Inter-comparison Project 5 (CMIP5) general circulation models (GCMs) were evaluated for their ability to skillfully reproduce basic characteristics of late 20th century climate over Central America. The models were ranked according to metrics that take into consideration the mean and standard deviation of precipitation (pr) and surface temperature (tas), as well as the El Niño-Southern Oscillation (ENSO)-pr teleconnection. Verification was performed by comparing model runs to observations and a reanalysis dataset. Based on the rankings, the best 13 models were further evaluated. Not surprisingly, the models showed better skill at reproducing mean tas patterns throughout the year. The skill is generally low for mean pr patterns, except for some models during March, April, and May. With a few exceptions, the skill was low for reproducing the observed monthly standard deviation patterns for both pr and tas. The ENSO-pr teleconnection was better simulated in the best 13 model runs compared to the sea-surface temperature global pattern characteristic of ENSO which showed low skill. The Inter-tropical Convergence Zone (ITCZ) appeared better modeled in July than in January. In January, there were instances of a double ITCZ pattern. Some models skillfully reproduced the seasonal distribution of the Caribbean Low-Level Jet index (CLLJ). More detailed research evaluating the specific performance of the models on a variety of time-scales and using parameters relevant to these and other climatic features of Central America is needed. This study facilitates a pre-selection of models that may be useful for this task. KEY WORDS climate model; ITCZ; Caribbean Low-Level Jet; GCM; Central America; skill Received 12 March 2014; Revised 2 October 2014; Accepted 24 October 2014 1. Introduction General circulation models (GCMs) are the main tool used by the Intergovernmental Panel on Climate Change (IPCC) for projecting climate impacts at the end of the century (IPCC, 2013). GCMs are based on mathematical representations of the physical laws governing the earth’s climate. Due to limitations in the modeling of small-scale processes and other deficiencies, GCMs cannot accurately reproduce certain features of regional and global climate. It is therefore often necessary to rank them according to their ability to reproduce observed climatic characteristics of a region, especially when evaluating large sets (around 100) of model runs, which can be prohibitively expensive if they are used for dynamical downscaling (Hidalgo and Alfaro, 2012b). The task of selecting models is not trivial as the types and numbers of metrics used for ranking greatly influences the results (Brekke et al., 2008; Hidalgo and Alfaro, 2012b). For example, a model could be very good at reproducing the general pattern of the El Niño-Southern Oscillation (ENSO), but deficient at reproducing the monthly * Correspondence to: H. G. Hidalgo, School of Physics, University of Costa Rica, San Pedro, Costa Rica. E-mail: [email protected] © 2014 Royal Meteorological Society means or standard deviations of precipitation (pr) or surface temperature (tas) over the area of interest, and vice versa. Although some biases, e.g. in the monthly means and standard deviations of pr and tas patterns can be bias-corrected using statistical methods (Hidalgo et al., 2013), other factors that affect the model’s skill in representing important climatic phenomena (e.g. ENSO) cannot be corrected. Model selection, therefore, always implicitly requires compromise. Pre-selection of the most skillful GCM runs based on basic climatic characteristics is useful for determining an initial set of ‘best’ model runs for further analysis (see for Sheffield et al., 2013; Neelin et al., 2013). In this study, 107 available GCM runs from 48 different models are ranked using basic climatic characteristics, as a precursor for future studies evaluating the use of the same models over Central America. That is, in this study, emphasis is placed on evaluating the ability of the GCM runs to reproduce basic but desirable historical climatic characteristics of Central America, namely (1) the monthly mean patterns of pr and tas, (2) the standard deviations of pr and tas, and (3) the ENSO-pr teleconnection patterns. These three metrics are used because they are amongst the most important climatic sources of variability in Central America. H. G. HIDALGO AND E. J. ALFARO It is recognized that (1) metrics representing other climatic features important to the region, e.g. the Inter-Tropical Convergence Zone (ITCZ), the Caribbean Low-Level Jet (CLLJ) (Amador, 1998; 2008), and the Mid-summer Drought (MSD) (Magaña et al., 1999) could have been included as a part of the evaluation of model skill, and (2) depending on the type and number of metrics used different rankings could result (Brekke et al., 2008). The objective of this study, however, is not to present a full evaluation of the models’ abilities to reproduce most (or all) climatic features relevant to the region at various time-scales. Rather, this study seeks to provide an initial ranking of a large quantity of model runs based on a limited and manageable number of basic climatic characteristics, which can then support (by aiding pre-selection) future and more detailed modeling studies in the region. As an example of what is possible, however, the general shape of the ITCZ, the climatology of the CLLJ and the ENSO-PDO global sea-surface temperature (SSTs) patterns are also evaluated qualitatively for the most skillful 13 models determined from the ranking [note that the Pacific Decadal Oscillation (PDO) is considered here as part of the ENSO pattern, given that both are closely related (Polade et al., 2013)]. In this study, the analysis is performed on runs from 48 Coupled Model Inter-Comparison Project, Phase 5 (CMIP5; Taylor et al., 2012b) fully coupled GCMs. Previous studies have determined the skill of CMIP3 (Meehl et al., 2007) GCMs in reproducing relevant climatic features of the region. For example, Pierce et al. (2008, 2009) show that the CMIP3 models reproduce with limitations, ENSO and the PDO (Mantua et al., 1997). Other evaluations can be found in the studies by Rauscher et al. (2011), Hirota et al. (2011) Delworth et al. (2012) and Liu et al. (2012) for the ITCZ; Rauscher et al. (2011) for the MSD; Martin and Schumacher (2011) for the CLLJ; and in the study by Jiang et al.’s (2012) examination of intra-seasonal variability. Hidalgo and Alfaro (2012b) also showed that over the Eastern Tropical Pacific, most of 30 CMIP3 GCM runs evaluated had significant biases in the monthly mean and standard deviation of monthly pr and tas patterns with respect to the National Center for Environmental Prediction/National Center for Atmospheric Research (NCEP/NCAR) Reanalysis (hereafter the Reanalysis; Kalnay et al., 1996). This suggests that statistical bias-correction of raw CMIP3 GCM pr and tas data is a necessary step when using these data in climate change studies involving statistical downscaling techniques (Hidalgo and Alfaro, 2012b). Knutti and Sedláček (2013) evaluated the robustness and uncertainties in the projections of CMIP5 models. They found similarities between CMIP3 and CMIP5 models in the projected pr change and in the ‘robustness’ of the projections, suggesting little improvement between the two generations of models. There is in general a lack of improved model convergence in CMIP5 compared to CMIP3. Knutti and Sedláček (2013), however, also found that potential improvements in pr projections are larger in the Tropics, including Central America. Notwithstanding © 2014 Royal Meteorological Society deficiencies, Knutti and Sedláček (2013) suggest that the incorporation of improved understandings of certain processes (i.e. incorporating bold assumptions or previously ignored factors) inspires confidence that the CMIP5 models capture more of the relevant processes. Other relevant CMIP5 evaluations include: • Kim et al.’s (2012) assessment of the CMIP5 decadal hindcast/forecast simulations of seven state-of-the art ocean–atmosphere coupled models. All the models showed high prediction skill for tas over the Indian, North Atlantic and western Pacific Oceans where the externally forced component and low-frequency climate variability is dominant (Kim et al., 2012). The Atlantic Multi-Decadal Oscillation (AMO; Enfield et al., 2001) is also predicted in the models with significant skill, while the PDO shows relatively low predictive skill (Kim et al., 2012). • Sheffield et al.’s (2013) evaluation of 17 CMIP5 models in terms of their skill in reproducing selected regional climate processes relevant to North American climate, including cool season western Atlantic cyclones, the North American monsoon, the United States Great Plains low-level jet and Arctic Sea Ice. No particular model stood out as performing better for all analyses, although some models performed much better for sets of metrics (Sheffield et al., 2013). • Neelin et al.’s (2013) report on the skill of CMIP5 models in representing California precipitation. The paper appears in volume 26 of the Journal of Climate which was devoted to studies of CMIP5 models’ representation of North American climate. Although some of the studies (e.g. Sheffield et al., 2013) included Central America, the predominant focus is on climatic processes of the mid-latitudes. • Polade et al.’s (2013) study of the teleconnections between SSTs in the Pacific (north of 30∘ S) and North American pr using Singular-Value Decomposition (SVD) analysis. In general, an improvement in the skill of the teleconnection was found in the CMIP5 models, compared to the CMIP3. Increased spatial resolution in the new generation of models and better physics contributed to the improvement (Polade et al., 2013). • Langenbrunner and Neelin’s (2013) study of the skill of CMIP5 AMIP (Atmospheric Model Inter-Comparison Project; Gates et al., 1999) models in reproducing the patterns of global ENSO teleconnections for 1979–2005. In AMIP5 models, SSTs are prescribed and only the atmospheric component of climate is simulated. In regions of strong signal (equatorial South America, the western equatorial Pacific, and a southern section of South America) there was little improvement in reproducing the amplitude and spatial correlation metrics of the pr teleconnection for CMIP5 versus CMIP3 models. However, other aspects, such as the amplitude of the pr response (root mean-square deviation over each region) were reasonably captured by the mean of the amplitudes of the individual models, in contrast with the multi-model ensemble mean. Although Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA the simulation of ENSO teleconnections is fairly challenging, the skill of individual models in reproducing the teleconnection signal amplitude (assessed using the mean of individual model amplitudes as opposed to the multi-model ensemble or MME) suggests that similar measures are trustworthy in the case of climate change studies. Sign agreement plots also proved to be a useful tool for such studies (Langenbrunner and Neelin, 2013). • Flato et al.’s (2013) summary of CMIP5 model evaluations. None of the above studies specifically focused on Central America or the Intra-Americas Sea. Some CMIP5 projection results for Central America and the Caribbean can be found in the study by Bindoff et al. (2013). Central America is a region of complex topography influenced by many climatic mechanisms, such as ENSO, the migration of the ITCZ (Sachs et al., 2009), AMO, PDO, the CLLJ, the MSD and others. The climate variability at seasonal and inter-annual scales, in particular those aspects associated with ENSO, have significant socioeconomic impacts on the countries of Central America (Waylen et al., 1996; IPCC, 1997; Hidalgo and Alfaro, 2012a, 2012b; Hidalgo et al., 2013). Central America has also been identified as a climate change ‘hot-spot’ (Giorgi, 2006). In this study, the focus is on the skill of the CMIP5 GCM model runs in reproducing basic climatic characteristics of Central America. 2. Characteristics of Central American climate High mountains divide Central America into two main climate regions – the Pacific and Caribbean slopes, which are lee and windward, respectively, of the North Atlantic trade winds which are the dominant wind regime (Maldonado et al., 2013). The rain over the Pacific slope has a bimodal annual cycle, with a maximum in May–June and another in August–September–October, separated by the MSD in July (Magaña et al., 1999). A first maximum of deep convective activity (and hence precipitation) occurs when surrounding SSTs exceed 29 ∘ C (around May). During July and August, the SSTs cool by ∼1 ∘ C due to diminished downwelling solar radiation and stronger easterly winds associated with the CLLJ, leading to a decrease in deep convective activity (Amador et al., 2006). The decreased convection facilitates increased downwelling solar radiation and a slight increase in SSTs (∼28.5 ∘ C), so that by the end of August through early September convection is again enhanced and a second maximum of precipitation occurs. Alfaro (2002) notes that a different climate regime dominates at stations located on the Caribbean Coast of Honduras, Costa Rica, and Panama. There are similarly two rainfall maxima but they occur in July and November–December, with the latter maximum being wetter than the first. There are also two minima in March and September–October, with the first being drier than the second. © 2014 Royal Meteorological Society The annual cycle of air surface temperature in Central America is tropical, predominantly maritime, with small annual changes, and dependent on cloud cover and altitude (Taylor and Alfaro, 2005). The dominant annual cycle (excepting the Atlantic coasts of Honduras and northern Nicaragua) is monsoonal, with highest temperatures occurring just before the summer rains. Temperatures are at their lowest during January, largely due to the cooling effect of the strong trade winds. Maximum temperatures occur during April and are associated with a decrease in trade wind strength, less cloud cover and higher values of solar radiation (Amador et al., 2006). There is a temperature minimum in July that coincides with the onset of the MSD (or in Spanish ‘veranillo’ or ‘canícula’). During this period, the trade winds briefly increase in intensity, the subtropical ridge over the Caribbean intensifies, and a second minimum and maximum in cloud cover and radiation respectively occur (Alfaro, 2000). The CLLJ, also known as the Intra-Americas Low-Level Jet, is another dominant climatic feature of the region. Its annual cycle is characterized by two wind maxima near 925 hPa in July and January–February (Amador, 2008). During summer, winds in excess of 13 m s−1 dominate the central Caribbean Sea (15∘ N, 75∘ W) and extend upward to 700 hPa. In comparison, the winter component of the jet is compressed below 850 hPa, with values of up 10 m s−1 , and with strong vertical wind shear which is unfavorable for convection. The summer peak generally starts to develop in early June just after the onset of the rainy season, reaches a maximum in July, before weakening in early September. During September to early November, the trades are relatively weak, vertical wind shear over the Caribbean is reduced, hurricane activity peaks, and rainfall spreads across most of the Intra-Americas Sea. In late November and early December, the trades again increase in strength, cold surges from mid and high latitudes reach the Tropics, and the second maximum of wind appears over the Caribbean Sea. 3. Data Simulations from 107 runs from 48 different CMIP5 GCMs for the historical period 1979–1999 and for ensembles from initializations r1i1p1, r2i1p1 and r3i1p1 were obtained from the Centro Agronómico Tropical de Investigación y Enseñanza (CATIE) located in the city of Turrialba, Costa Rica, and from the Earth System Grid Federation (http://pcmdi9.llnl.gov/esgf-web-fe/), a data server from the Program for Climate Model Diagnostics and Inter-comparison (PCMDI) at the Lawrence Livermore National Laboratory of the United States. The data consist of monthly pr, tas and zonal and meridional 925 hPa wind speeds (u, v). Not all the models have wind data and therefore much of the analysis performed was based on the pr and tas data only. The wind data were used in an analysis of the climatology of the CLLJ index, which is based in u only and at the level of the jet core (925 hPa). Therefore, the v data and the u data at other levels were not used. Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO In Table 1, the characteristics of the runs are shown. There is at least one run available for each model corresponding to run r1i1p1. In some cases, additional runs corresponding to r2i1p1 and r3i1p1 were also available. Detailed explanation of the difference between the runs can be found in the study by Taylor et al. (2012a). In summary, the runs examined correspond to equally likely outcomes for a particular simulation (i.e., they typically differ only by being started from equally realistic initial conditions). Historical runs initialized from different times of a control run are identified by ‘r1’, ‘r2’, ‘r3’, etc. The other letters in the ensemble identifier distinguish between initializations of models with different methods (‘i1’, ‘i2’, ‘i3’, etc.) and different perturbed physics (‘p1’, ‘p2’, ‘p3’, etc.). As previously mentioned, only runs with initializations 1, 2 and 3 were selected, while the methodof-initialization and perturbed physics were fixed to i1p1. The choice of runs is in part dictated by the unavailability of historical (and matching climate change) runs for initializations beyond r3 and for other methods and physics. Monthly pr, tas and u data from the Reanalysis dataset for 1979–1999 are compared with the data from the GCMs. In particular, tas Reanalysis patterns are used as verification for the tas patterns from the GCM runs. Where necessary, units were converted to match those from the models, and the GCM data interpolated to the 2.5 × 2.5 degree resolution of the Reanalysis dataset using the nearest grid-point method. Although it is recognized that surface air temperature requires correction for elevation over land regions, especially if they are to be used for impacts analysis, no correction is carried out in this study. The purpose of this work is to explore and determine all possible sources of error, including those caused by different grid resolutions between models. The Reanalysis data were provided by the National Oceanographic and Atmospheric Administration, Office of Oceanic and Atmospheric Research, Earth System Research Laboratory, Physical Sciences Division (NOAA/OAR/ESRL PSD) in Boulder, Colorado, USA, and accessed from their website at http://www.esrl.noaa.gov/psd. Monthly pr data from 1979 to 1999 were also obtained from the Global Precipitation Climatology Project (GPCP; Adler et al., 2003), version 2.2, provided by NOAA/OAR/ESRL PSD. The GPCP dataset combines satellite and rain-gauge sources at a spatial resolution of 2.5 × 2.5 degrees. The GCMs’ data were converted to the resolution of the GPCP for comparison. Global monthly SST data (Smith and Reynolds, 2004) version 3b (hereinafter Reynolds), from 1854 to the present were obtained from NOAA/OAR/ESRL PSD. The data are available at 2 × 2 degrees resolution and were used for verification of ENSO SST patterns and indices simulated by the GCMs. 4. Methods The first part of the analysis compares observed and modeled climatic patterns from 1979 to 1999 (the common © 2014 Royal Meteorological Society period of the models and GPCP data) using an index of skill. The skill index compares observed and modeled patterns of 5 metrics shown in Table 2. Initially, 12 pr and 12 tas mean monthly patterns, and 12 pr and 12 tas standard deviation monthly patterns are calculated over a target area bounded by coordinates 4∘ N and 20∘ N, and 95∘ W and 75∘ W. The 1979–1999 period is used for all pr and tas metrics. Twenty pairs (modeled and observed) of patterns corresponding to the ENSO-pr teleconnection metrics are also calculated. The procedure for obtaining each ENSO-pr teleconnection pattern was as follows: • The Niño1.2 time series (see location of the averaging area in Figure 1) was calculated from the observed NOAA/OAR/ESRL PSD SST data for DJF for each year from 1979 to 1999. • The precipitation time series was calculated from the observed GPCP data for each grid point in the target area (4∘ –20∘ N, 95∘ –75∘ W) for DJF for each year from 1979 to 1999. • Temporal correlation coefficients between the Niño1.2 time series and the precipitation time series for all the grid points in the target area were calculated. This gave a two-dimensional distribution of correlation coefficients over the target area (i.e. the teleconnection pattern). • The process was repeated for DJF Niño1.2 and DJF precipitation time series derived for each CMIP5 model runs. • The process was repeated for the Niño3, Niño3.4 and Niño4 regions during DJF. • The process was repeated for MAM, JJA, SON and annual averages of the ENSO indexes and precipitation data. The metrics in Table 2 partially differ from those used by Hidalgo and Alfaro (2012b), as some of the metrics in the latter study could not be calculated due to the shorter period of record in this study (a limitation imposed by the availability of GPCP data). The short period of record also limits the determination of AMO-pr teleconnection patterns given the multi-decadal nature of the AMO. Additionally, the ENSO-pr teleconnection metric was not used by Hidalgo and Alfaro (2012b), but is incorporated in this study because of the importance of ENSO (second only to the annual cycle) as an influencing climatic mechanism for Central America. Following the study by Pierce et al. (2009), the degree of similarity between any two climate patterns (e.g. between the same metric calculated from both GPCP data and a GCM simulation) was calculated using a Skill Score (SS) defined by: [ ( )]2 [ ]2 Sm m−o 2 − (1) SS = rm,o − rm,o − So So In Equation (1), rm,o is the Pearson’s (ordinary) spatial correlation between modeled (i.e. GCM) and ‘observed’ (e.g. GPCP) patterns, and Sm and So are the sample spatial standard deviations for the modeled and observed patterns, respectively. The m and o correspond to the spatial average Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Table 1. CMIP5 runs used in the study. No. Model Modeling center or group 1 access1_0 2 3 access1_3 bcc_csm1_1_m 4 5 bcc_csm1_1 bnu_esm 6 7 8 9 cancm4 canesm2 ccsm4 cesm1_bgc 10 11 12 13 14 15 16 17 cesm1_cam5_1_fv2 cesm1_cam5 cesm1_fastchem cesm1_waccm cmcc_cesm cmcc_cm cmcc_cms cnrm_cm5 18 csiro_mk3_6_0 19 20 ec_earth fgoals_g2 21 fgoals_s2 22 23 24 25 26 27 28 29 30 31 fio_esm gfdl_cm2p1 gfdl_cm3 gfdl_esm2g gfdl_esm2m giss_e2_h_cc giss_e2_h giss_e2_r_cc giss_e2_r hadcm3 32 33 34 35 36 37 38 39 hadgem2_ao hadgem2_cc hadgem2_es inmcm4 ipsl_cm5a_lr ipsl_cm5a_mr ipsl_cm5b_lr miroc4h 40 41 42 43 44 45 46 47 48 miroc5 miroc_esm_chem miroc_esm mpi_esm_lr mpi_esm_mr mpi_esm_p mri_cgcm3 noresm1_m noresm1_me Commonwealth Scientific and Industrial Research Organization and Bureau of Meteorology (CSIRO-BOM) CSIRO-BOM Beijing Climate Center, China Meteorological Administration (BCC) BCC College of Global Change and Earth System Science, Beijing Normal University (BNU) Canadian Centre for Climate Modelling and Analysis (CCCMA) CCMA CCMA Centro Euro-Mediterraneo per I Cambiamenti Climatici (CMCC) CMCC CMCC CMCC CMCC CMCC CMCC CMCC Centre National de Recherches Meteorologiques / Centre Europeen de Recherche et Formation Avancees en Calcul Scientifique (CNRM) Commonwealth Scientific and Industrial Research Organisation in collaboration with the Queensland Climate Change Centre of Excellence (CSIRO-QCCE) EC-EARTH consortium LASG, Institute of Atmospheric Physics, Chinese Academy of Sciences; and CESS, Tsinghua University LASG, Institute of Atmospheric Physics, Chinese Academy of Sciences The First Institute of Oceanography, SOA, China Geophysical Fluid Dynamics Laboratory (GFDL) GFDL GFDL GFDL GFDL GFDL GFDL GFDL Met Office Hadley Centre (HadCM) additional HadGEM2-ES realizations contributed by Instituto Nacional de Pesquisas Espaciais HadCM-HADGM HadCM-HADGM HadCM-HADGM Institute for Numerical Mathematics (INM) Institut Pierre-Simon Laplace (IPSL) IPSL IPSL Atmosphere and Ocean Research Institute (The University of Tokyo), National Institute for Environmental Studies, and Japan Agency for Marine-Earth Science and Technology. (MIROC) MIROC MIROC MIROC Max Planck Institute for Meteorology (MPI-M) MPI-M MPI-M Meteorological Research Institute (MRI) Norwegian Climate Centre (NCC) NCC © 2014 Royal Meteorological Society Runs 1 2 3 3 1 1 3 3 1 3 3 3 2 1 1 1 3 3 2 3 3 3 3 3 3 3 1 3 1 3 3 1 1 3 1 3 1 1 3 3 1 3 3 3 2 3 3 1 Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO are the Conditional and Unconditional Biases respectively (see Pierce et al., 2009). Note that SS not only reflects correlation coherence between the patterns but also incorporates biases in its calculation. Once the SSs were calculated, for each model and for each type of metric they were combined using the Euclidean distance (Δ) from their optimum solution SS = 1, such that the lower the Δ, the better is the model in reproducing observed patterns. An example of the formula used to calculate Δ for metric 1 is: √ √ ( )2 ( )2 pr √ 1 − SSpr + 1 − SSFeb mean √ Jan mean Δ=√ (3) √ ( )2 pr + … + 1 − SSAnnual mean Table 2. Metrics used in the study. No. Name of the metric Description 1 Precipitation monthly means 2 Surface temperature monthly means Precipitation monthly standard deviations Surface temperature monthly standard deviations ENSO teleconnections Seasonal monthly means of pr Seasonal monthly means of tas Seasonal monthly std. devs. of pr 3 4 5 Symbols Seasonal monthly std. devs. of tas Correlations between SST (or tas as surrogate) in ENSO regions and precipitation (over the region of interest) for DJF, MAM, JJA, SON, and annual of the modeled and observed climate patterns, respectively. SS varies from −∞ (no skill) to 1 (perfect match between the patterns). Zero SS values correspond to cases in which the mean of the observations is reproduced correctly by the model in a certain region, but only as a featureless uniform pattern (Pierce et al., 2009). The right-hand side of Equation (1) is composed of three squared terms, and therefore SS can also be expressed as: SS = RHO − CBIAS − UBIAS (2) RHO is the square of the spatial correlation between the observed and modeled patterns, and CBIAS and UBIAS In the equation, Δ is the Euclidean distance between the SSs of all patterns in this particular metric (see again definition of the metrics in Table 2) with the optimum vector of (1, 1, 1, … 1). Therefore, Δ is computed for each model and metric. In the example represented by Equation (3), metric 1 (mean pr patterns) has 13 SSs – one for each of the monthly mean patterns plus one associated with the annual average patterns. Sorting the Δs from low to high for all models, and for that particular metric, results in the partial rankings for metric 1. The procedure is then repeated for all metrics of Table 2 and the results are Δ ranks (partial ranks) for each individual metric. The ranks for all the metrics are added to produce a sum of rankings, and the Final rank is produced by ranking the sum of rankings (Table 3). Finally, the 13 best models are compared qualitatively and quantitatively in terms of how good they reproduce some characteristics of the ITCZ, the seasonal cycle of the CLLJ and the ENSO global loading pattern associated with the first principal component (PC) of the annual average of SST limited to 60o S and 60o N. For the models, tas is used over the oceans as a surrogate for SST. 5. Results 5.1. SS ranking Although the partial ranks were constructed using all the SSs of each individual type of metric (as shown in Tables 2 30°N 20°N 10°N 0° 10°S 160°E 180°W 160°W 140°W 120°W 100°W 80°W 60°W Figure 1. Location of the target area (solid thick line over Central America). Also shown are the ENSO SST regions: Niño1.2 (dashed line), Niño3 (dark grey region), Niño4 (light fray region) and Niño3.4 (solid thick line over the ocean partially covering Niño3 and Niño4). © 2014 Royal Meteorological Society Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Table 3. Partial and final rankings of the GCM model runs using different metrics and variables. Numbers in parenthesis indicate run number. Model cesm1_cam5(1) cesm1_cam5(3) cnrm_cm5(3) cnrm_cm5(1) cesm1_cam5_1_fv2(3) cnrm_cm5(2) cesm1_cam5(2) mpi_esm_p(1) cmcc_cms(1) cesm1_fastchem(3) cesm1_fastchem(2) mpi_esm_p(2) ccsm4(3) ec_earth(2) mpi_esm_lr(1) mri_cgcm3(2) mpi_esm_mr(3) ccsm4(1) cesm1_cam5_1_fv2(2) ccsm4(2) mpi_esm_lr(3) mpi_esm_lr(2) cesm1_fastchem(1) mpi_esm_mr(2) ec_earth(1) cesm1_bgc(1) giss_e2_r(3) miroc4h(1) miroc5(3) miroc5(1) inmcm4(1) hadgem2_cc(1) giss_e2_r_cc(1) giss_e2_r(2) miroc4h(2) miroc4h(3) giss_e2_r(1) cesm1_cam5_1_fv2(1) giss_e2_h(1) mpi_esm_mr(1) hadcm3(1) fgoals_g2(1) gfdl_esm2g(2) fio_esm(2) fgoals_g2(3) giss_e2_h(3) gfdl_cm3(2) miroc5(2) giss_e2_h(2) bcc_csm1_1(3) mri_cgcm3(1) noresm1_me(1) cmcc_cm(1) fgoals_g2(2) hadgem2_es(2) gfdl_esm2g(1) gfdl_cm3(1) access1_3(1) noresm1_m(3) fio_esm(1) hadgem2_es(1) cesm1_waccm(2) Metric 1 (pr) Metric 2 (tas) Metric 3 (pr) Metric 4 (tas) Metric 5 (ENSO) Sum of ranking Final rankings 9 5 1 2 12 3 4 19 7 24 31 13 42 6 18 29 65 34 14 30 28 25 32 33 8 55 41 78 15 10 56 63 45 39 80 77 36 16 59 62 20 73 27 46 67 54 68 11 57 48 37 83 23 61 71 49 66 26 87 52 74 35 17 13 18 14 20 16 15 4 21 25 24 3 27 75 5 40 10 30 22 26 1 2 28 11 77 31 92 44 35 34 37 7 95 94 45 54 93 23 99 6 55 66 85 49 67 96 61 33 97 65 41 59 29 63 9 90 56 62 52 48 12 36 18 3 15 26 16 32 28 11 1 38 50 49 39 6 37 57 70 47 19 27 34 44 42 64 25 54 8 24 12 2 75 104 23 13 36 33 17 10 21 63 30 9 35 14 5 46 56 43 29 96 51 67 89 4 106 31 52 97 77 22 102 41 4 10 51 36 6 43 19 60 66 17 2 57 11 48 87 29 39 31 58 8 61 68 3 74 64 1 12 5 86 95 33 32 24 37 7 28 21 90 46 88 98 26 59 35 27 34 23 106 56 45 41 13 65 50 63 84 54 78 9 70 40 80 6 33 1 9 42 3 43 17 27 25 24 14 22 7 8 11 2 46 76 99 68 54 89 13 23 58 48 51 61 77 19 16 38 49 64 41 69 98 12 20 39 72 40 103 81 18 45 60 15 5 94 52 70 100 30 28 55 21 59 96 63 101 54 64 86 87 96 97 109 111 122 129 131 136 141 142 155 166 186 188 189 190 192 193 194 195 197 199 201 202 209 218 220 222 225 232 232 233 236 237 237 239 242 246 246 247 247 248 253 253 254 259 264 274 276 278 279 282 283 284 284 288 291 293 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 © 2014 Royal Meteorological Society Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO Table 3. Continued Model fio_esm(3) cesm1_waccm(1) bcc_csm1_1(2) noresm1_m(2) access1_0(1) hadgem2_ao(1) gfdl_cm3(3) hadcm3(3) noresm1_m(1) cmcc_cesm(1) fgoals_s2(1) mri_cgcm3(3) gfdl_esm2g(3) hadgem2_es(3) hadcm3(2) access1_3(2) miroc_esm_chem(1) bnu_esm(1) canesm2(1) cancm4(2) ipsl_cm5a_lr(2) cancm4(1) giss_e2_h_cc(1) gfdl_cm2p1(3) ipsl_cm5a_lr(3) fgoals_s2(3) bcc_csm1_1_m(3) miroc_esm(1) miroc_esm(3) gfdl_cm2p1(1) bcc_csm1_1(1) bcc_csm1_1_m(2) ipsl_cm5a_lr(1) gfdl_cm2p1(2) canesm2(2) ipsl_cm5a_mr(1) canesm2(3) gfdl_esm2m(1) ipsl_cm5b_lr(1) miroc_esm(2) fgoals_s2(2) csiro_mk3_6_0(3) csiro_mk3_6_0(1) bcc_csm1_1_m(1) csiro_mk3_6_0(2) Metric 1 (pr) Metric 2 (tas) Metric 3 (pr) Metric 4 (tas) Metric 5 (ENSO) Sum of ranking Final rankings 58 38 64 81 99 98 60 22 85 17 102 47 43 75 21 69 76 90 84 100 93 97 44 40 88 104 92 79 72 51 70 91 95 53 96 101 86 50 89 82 103 106 107 94 105 53 39 69 46 19 32 60 70 58 107 100 42 89 8 71 64 104 74 47 57 72 43 102 87 68 98 83 103 106 86 73 81 76 88 51 78 50 91 38 105 101 79 80 82 84 20 40 94 69 107 101 45 83 68 7 65 86 48 105 81 98 62 60 72 80 87 78 58 59 82 74 93 76 61 55 92 88 84 53 85 90 79 73 100 71 66 99 103 91 95 83 81 30 14 49 71 62 101 15 102 16 52 76 47 99 73 85 25 89 69 18 93 77 105 22 20 53 75 97 103 67 55 44 96 91 42 92 94 72 79 38 107 100 82 104 82 102 44 91 32 4 80 31 84 78 34 90 65 86 56 29 10 104 62 57 93 53 83 74 106 71 47 36 37 79 73 67 88 97 66 85 92 95 107 75 105 26 35 87 50 296 300 301 301 306 306 307 307 310 311 317 317 321 321 328 333 337 353 354 363 363 364 364 365 366 367 368 369 373 374 375 382 387 387 389 396 399 403 406 412 413 417 425 436 438 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 and 3), it is important to have a general idea of which types of metrics showed the highest deviations from the observed values and during which months of the year. Histograms of the (1 − SS)2 terms for a few examples of each different type of metrics are shown in Figures 2–4. For pr (Figure 2), the mean observed patterns are generally reproduced with more skill [lower (1 − SS)2 ] for January than for July. In fact, there were a few runs that performed very poorly during July compared to the majority of the simulations. The standard deviations of pr were generally also better simulated during January compared to July, and there were more high-value (1 − SS)2 outliers during July. In terms of tas (Figure 3), there was generally less contrast in the skill between January and July, but there © 2014 Royal Meteorological Society were more high-value (1 − SS)2 outliers during July. In terms of the ENSO-pr teleconnection, the skill during all seasons was similar, except for MAM which (in general) showed considerably less skill. The annual teleconnection also showed slightly better skill than for the individual seasons. In Table 3, the 107 runs are shown in order of increasing Final ranks. The partial rankings for each metric are also included. The cesm1_cam5 model runs (1, 2 and 3) consistently ranked at position 7 or higher in the Final ranks. This model performed relatively well (top 10 ranks) at reproducing the mean patterns of precipitation (Metric 1), and its other partial ranks were also relatively good (equal or better than 28, i.e. in the top third of all model Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Figure 2. Histograms of (1 − SS)2 corresponding to the Skill Score (SS) between the modelled and observed mean and std. dev. precipitation patterns for January and July. The 107 runs from Table 1 are considered in this figure. Figure 3. Histograms of (1 − SS)2 corresponding to the Skill Score (SS) between the modelled and observed mean and std. dev. temperature patterns for January and July. The 107 runs from Table 1 are considered in this figure. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO Figure 4. Histograms of (1 − SS)2 corresponding to the Skill Score (SS) between the modelled and observed mean and std. dev. correlation patterns of Niño3.4 region SST average and Central American precipitation for different seasons. The 107 runs from Table 1 are considered in this figure. runs), with the exception of the ranks of runs 3 and 2 for Metric 3 (ENSO) which are slightly higher (33 and 43, respectively). The three runs of the cnrm_cm5 model also performed relatively well (ranked equal or better than 6) in the Final rank. In general, the cnrm_cm5 model performed better in terms of the ENSO teleconnection pattern (Metric 5) than the cesm1_cam5 model, but showed higher ranks for reproducing standard deviations of tas patterns (Metric 4). There are models represented with just one run in the top 25 rankings, but it is remarkable to note models with at least two runs in the top 25 positions, such as cesm1_fastchem, ccsm4 and mpi_esm_lr. Among the models with the poorest performance that showed at least two runs in the bottom 25 positions were csiro_mk_6_0, miroc_esm, canesm2, fgoals_s2, some of the ipsl_cm5 runs and others. 5.2. Characteristics of the ‘best’ 13 runs The best 13 runs using the Final ranking were selected and their skill in reproducing characteristics of the 1979–1999 © 2014 Royal Meteorological Society climate of the region examined. In Figure 5, the SSs distribution of these 13 runs for metrics 1 through 4 (from Table 2) are shown. Not surprisingly, there was generally better skill in reproducing mean monthly tas patterns than mean monthly pr patterns, but both variables showed poor results with respect to standard deviation patterns. The runs showed median positive SSs when reproducing mean pr patterns of February, March, April and May, but low skill for all other months (including negative SS values), suggesting the presence of large biases or low spatial inter-correlation between observed and modeled patterns [Figure 5(a)]. The skill in reproducing pr standard deviations was also generally low [Figure 5(c)], although in February and March the median SS values were above zero, suggesting some skill during those months. In terms of tas, the median SSs for metric 2 (Figure 5(b)) were always above zero, suggesting significant skill in reproducing the observed patterns. However, for the standard deviation of tas [Figure 5(d)] the results for some months (especially September and October) suggested low skill. Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Figure 5. Boxplots of the distribution of the SSs for the best 13 models shown in Table 3. The upper two subplots correspond to the SS calculated between the mean modelled and observed patterns for pr and tas (metrics 1 and 2 of Table 2), while the bottom correspond to the standard deviation patterns (metrics 3 and 4 of Table 2). The boxes represent the 25th and 75th percentiles, the median is shown inside the boxes, the whiskers extend 1.5 the interquartile range or to the extent of the data, and the outliers are shown with the asterisk symbol. Figure 6. Boxplots of the distribution of the SSs for the best 13 models shown in Table 3. The upper two subplots correspond to the SS calculated between the mean modelled and observed patterns for the ENSO teleconnections (metric 5 of Table 2) for different ENSO indexes. The boxes represent the 25th and 75th percentiles, the median is shown inside the boxes, the whiskers extend 1.5 the interquartile range or to the extent of the data, and the outliers are shown with the asterisk symbol. With respect to ENSO-pr teleconnections (Figure 6), there was generally less skill during MAM for all ENSO indicators. However, in the case of annual averages, the SS medians for the best 13 models were always above zero for the four ENSO indicators evaluated. Positive median © 2014 Royal Meteorological Society skill was found in at least one of the four ENSO indicators during JJA and SON, while the SS medians during DJF and MAM were always below zero for all indicators. The strongest ENSO-pr teleconnection is known to be located off the Pacific coast of Central America (Waylen et al., Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO 1996) which has a rainy season during May to November and a dry season from December to April (Alfaro, 2002; Taylor and Alfaro, 2005). It is noteworthy then, that the highest ENSO skill of the GCM runs occurs mainly during the wettest seasons of the year in this climate regime. In Figures 7 and 8, mean pr patterns for the GCPC, the Reanalysis and the best 13 GCM models are shown for January and July, respectively (1979–1999). The area of interest over which the SSs were calculated is shown as a red square in Figures 7(a) and 8(a). The SS values inside the subplots indicate that the GCMs generally performed as well or better than the Reanalysis in reproducing mean GCPC pr patterns in the area of interest during January and July. In fact, all of the first 13 models showed top (highest?) ranks in terms of metric 1 (Table 3). Although there is no direct comparison between the methodologies employed in this study and that used in the study by Sheffield et al. (2013), some similarities and differences are noted between results obtained for models common to both studies. For example, the runs corresponding to the crnm_cm5 model are highly ranked in Table 3 for the partial rank corresponding to metric 1 (mean pr patterns). In the study by Sheffield et al. (2013), the model showed low pr biases compared to the models they analyzed for December–January–February (DJF), but significant pr biases for June–July–August (JJA). Another difference is that in the study by Sheffield et al.’s (2013) study, the csiro_mk3.6.0 model showed relatively low pr biases in both DJF and JJA, while in this study the model did not perform well (Table 3). We do not think that the mentioned differences are solely explained by the different methodologies used. They may also result from differences in the definition of the regions over which the models were evaluated. It should be mentioned, however, that the mean pr SSs found for the best 13 models during January and July (Figures 7 and 8) were generally low and in many cases negative, possibly due to the large biases in the means detected in the GCMs (Hidalgo and Alfaro, 2012b). Fortunately, some of these biases can be partially corrected through bias-correction procedures (Maurer and Hidalgo, 2008; Maurer et al., 2010; Hidalgo et al., 2013). However, it is desirable to use models requiring the smallest amount of correction and hence an evaluation of the models’ skill such as is being done in this study is useful. Although the area of interest is smaller than the domain shown in the maps of Figures 7–16, the larger domain facilitates a qualitative evaluation of the shape of the ITCZ for the months shown. The plots show that the ITCZ over the eastern Pacific Ocean was stronger and more defined in July than in January and that the GCMs capture the shape of the pr maximum over the Pacific Ocean with some skill (interestingly, the Reanalysis appears to be missing this feature, but it should be remembered that the Reanalysis precipitation is modeled and not observed). From Figures 7 and 8, there was no clear evidence of a double ITCZ, a problem that was common to GCMs in the previous generation of models (Hirota et al., 2011; Liu et al., 2012). To investigate this further, the difference © 2014 Royal Meteorological Society maps (modeled − observed) of the pr patterns for the best 13 models were plotted (not shown). In these plots, there was evidence of a double ITCZ in January in the models ranked 1 to 7 and 11. During July, although there are errors present in the model, the patterns do not resemble a double ITCZ (not shown). The problems with the double ITCZ in these best 13 models appear to occur in the months when the ITCZ is the weakest, and the pr over Central America is the lowest. In terms of the pr standard deviations (Figures 9 and 10), there was better skill in the GCMs than in the Reanalysis in January over the region of interest. In July, some models performed better while others did worse than the Reanalysis. In general, there was great dissimilarity (over the entire plot domain) between the large-scale GPCP pr patterns and those from the Reanalysis or GCMs. It is evident that of all the metrics examined, the standard deviations are the most difficult patterns to be replicated by the models. For tas means (Figures 11 and 12), the SSs were very high and positive during January and July. The large-scale patterns were also well replicated, including the contrast between land and sea in some regions of the maps. The tas standard deviations (Figures 13 and 14) were not well reproduced during January for the area of interest or the large-scale, Interestingly, however, some of the GCM runs (cesm1_cam5(1), cesm1_cam5(3), cesm1_cam5_1_f2, cesm1_cam5(2), cesm1_fastchem(3), cesm1_fastchem(2) and ccsm4(3)) showed significantly positive skill in reproducing the standard deviations of the July tas in the region of interest. This was consistent with Figure 5(d) which showed that the best skill for standard deviations of tas was found during the month of July, and the worst during October. With respect to the ENSO annual teleconnection using Niño3.4, all GCM runs were better than the Reanalysis at reproducing the observed ENSO-pr teleconnection pattern contained in the GPCP/Reynolds data for the area of interest (Figure 15). Overall, the SSs were all positive and most of them relatively high. The large-scale correlation patterns (considering the entire map domain) generally showed different amplitudes compared to the observations (Figure 15), though in many model runs the patterns had similar shapes. The patterns over the region of interest showed very good coherence between the observed and modeled patterns in most of the cases. In order to examine the ability of the models in reproducing global ENSO patterns, the loading patterns of the first un-rotated PC of annual SST (or tas as surrogate) over the global oceans inside 60o S and 60o N were computed from the Reynolds data, the Reanalysis and the models (Figure 16). The variances explained by these PCs are shown in Table 4. On average, the first PCs of the GCM runs explained around 23% of the variance, while in the observations this number was close to 21%. The GCM runs performed poorly in reproducing the global loading pattern with respect to the Reanalysis, as this latter database includes assimilated tas observations (in contrast to the GCM runs that are entirely modeled). The closest match Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Figure 7. January mean precipitation (mm month−1 ) for GPCP, Reanalysis and the best 13 GCM runs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the GPCP subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. The Reanalysis was included for reference (i.e. not ranked with the GCMs). Numbers in the subfigures represent the Skill Score for the seasonal patterns in the area of interest. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO Figure 8. July mean precipitation (mm month−1 ) for GPCP, Reanalysis and the best 13 GCM runs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the GPCP subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. The Reanalysis was included for reference (i.e. not ranked with the GCMs). Numbers in the subfigures represent the Skill Score for the seasonal patterns in the area of interest. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) (k) (l) (m) (n) (o) Figure 9. January precipitation standard deviation (mm month−1 ) for GPCP, Reanalysis and the best 13 GCM runs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the GPCP subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. The Reanalysis was included for reference (i.e. not ranked with the GCMs). Numbers in the subfigures represent the Skill Score for the seasonal patterns in the area of interest. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO Figure 10. July precipitation standard deviation (mm month−1 ) for GPCP, Reanalysis and the best 13 GCM runs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the GPCP subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. The Reanalysis was included for reference (i.e. not ranked with the GCMs). Numbers in the subfigures represent the Skill Score for the seasonal patterns in the area of interest. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Figure 11. January mean temperature (∘ C) for Reanalysis and the best 13 GCMs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the REAN subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. Numbers in the subfigures represent the Skill Score for the seasonal patterns. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO Figure 12. July mean temperature (∘ C) for Reanalysis and the best 13 GCMs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the REAN subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. Numbers in the subfigures represent the Skill Score for the seasonal patterns. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Figure 13. January temperature standard deviation (∘ C) for Reanalysis and the best 13 GCMs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the REAN subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. Numbers in the subfigures represent the Skill Score for the seasonal patterns. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO Figure 14. July temperature standard deviation (∘ C) for Reanalysis and the best 13 GCMs ordered from lowest to highest Final ranking (see text). Numbers in parenthesis indicate the run number. The square in the REAN subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. Numbers in the subfigures represent the Skill Score for the seasonal patterns. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Figure 15. Annual correlations between precipitation and Niño3.4 for GPCP/Reynolds, Reanalysis and the best 13 GCMs runs ordered from lowest to highest delta total (see text). Numbers in parenthesis indicate the run number. The square in the GPCP/Reynolds subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. The Reanalysis was included for reference (i.e. not ranked with the GCMs). The numbers in the figures are the Skill Scores between observed and modeled patterns. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO Figure 16. Annual correlations between precipitation and Niño3.4 for annual temperature loading maps of the first principal component of SST (or tas) over the oceans from 60o S to 60o N. GCMs runs ordered from lowest to highest delta total (see text). Numbers in parenthesis indicate the run number. The square in the Reynolds subfigure represents the area of interest where comparisons were made between observed and modeled patterns for selecting the best models. The Reanalysis was included for reference (i.e. not ranked with the GCMs). The numbers inside the maps are the Skill Scores between the observed pattern (using Reynolds data) and each model for the global domain. © 2014 Royal Meteorological Society Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Table 4. Variance explained by first principal component of SST or tas over the global ocean below 60o S and 60o N, for 13 models of Figure 7. Model Reynolds REAN cesm1_cam5(1) cesm1_cam5(3) cnrm_cm5(3) cnrm_cm5(1) cesm1_cam5_1_fv2(3) cnrcm_cm5(2) cesm1_cam5(2) mpi_esm_p(1) cmcc_cms(1) cesm1_fastchem(3) cesm1_fastchem(2) mpi_esm_p(2) ccsm4(3) Variance explained (%) 20.69 20.03 20.32 26.90 18.73 18.44 22.53 16.63 27.38 19.94 28.56 27.29 21.36 24.94 24.28 Figure 17. Seasonal averages of the CLLJ index for the Reanalysis and selected models with available wind data. See text for definition of the CLLJ index. Data for model cesm1_cam5_fv(2) were not available. The darkest line correspond to the Reanalysis results. with observations from the subset of 13 best models was cesm1_fastchem(2) which had a SS of −0.81 compared to 0.57 for the Reanalysis). A few of the best GCMs had wind speed data. In Figure 17, the seasonal cycles of the CLLJ index (Amador, 1998, 2008) are shown. The CLLJ index was constructed by averaging the zonal wind speed at 925 hPa from the Reanalysis and the GCMs in a region bounded by latitudes 7.5∘ N to 12.5∘ N, and longitudes 85∘ W and 75∘ W. Note that a stronger jet is associated with a more negative CLLJ index, as the sign convention of the wind data is negative for easterly winds (the prevailing direction). Although the sample size is too small to reach robust conclusions, it is evident that at least some of the models were very skillful at reproducing the seasonality of the CLLJ. The CLLJ seasonal cycles of the three cnrm_cm5 runs were very similar to those of the Reanalysis data (Figure 17), while other models had significant negative biases. These models suggest themselves as good possibilities for further future evaluation of other parameters (e.g. atmospheric circulations, humidity patterns and others) to determine if they too will display such promising results. © 2014 Royal Meteorological Society 6. Summary and conclusions The skill of 107 GCMs runs in reproducing climatic patterns of Central America as observed in the GCPC, Reynolds and Reanalysis data was measured using a SS index. Results indicated that for the metrics evaluated, the models had greatest difficulty in accurately reproducing the monthly standard deviation pr and tas patterns over the region of interest. With some exceptions, the standard deviations for pr and tas, were reproduced with low skill, even in the best models. Reproducing precipitation mean monthly patterns was also challenging for the models. Only a few of the best 13 runs have high SSs for reproducing mean pr patterns. For tas, the models, however, show significant skill. It should be remembered that biases play an important role in producing low SSs for the mean and standard deviations of pr and tas. Many of these biases can be reduced significantly using bias-correction procedures, which could make these runs useful for climate change studies (Maurer and Hidalgo, 2008; Maurer et al., 2010; Hidalgo and Alfaro, 2012b; Hidalgo et al., 2013). However, as previously noted, it is desirable to identify and use the less biased models, and this study helps with this task. Other features are more difficult to fix when they are not simulated well, such as the ENSO teleconnection patterns and ENSO characteristic SST patterns, ITCZ patterns and the seasonal cycles of the CLLJ index. It is also difficult to determine the causes of the differences between runs/models, as they all have different physics, initializations and/or spatial resolutions. The Final ranking of the models showed that differences in the initialization of the models (resulting in different runs of a particular model) are not as important when determining its performance, compared to the difference between models’ structure. That is, generally the multiple runs of a particular model have similar ranks, grouped in the top, middle or bottom of the list. Finally, features such the shapes and amplitudes of the ITCZ patterns in the GCMs are in some cases reproduced with more skill than the Reanalysis. However, the Reanalysis is much better in reproducing the ENSO SST patterns than the GCM runs. The assimilated temperature data in the Reanalysis may be playing an important role in this regard. The seasonality of the CLLJ index was reproduced with great skill by some of the models/runs, however, the sample is too small to reach conclusive statements. The most promising models can/should be studied in more detail using other variables and time-scales to validate their usefulness for reproducing climatic features of the region. Acknowledgements This work was partially funded by the International Climate Initiative (ICI) of the German Federal Ministry for the Environment, Nature Conservation, Building and Nuclear Safety (BMUB), as part of the CASCADE project (‘Ecosystem-based Adaptation for Smallholder Subsistence and Coffee Farming Communities in Central Int. J. Climatol. (2014) H. G. HIDALGO AND E. J. ALFARO America’). The German Federal Ministry for the Environment, Nature Conservation, Building and Nuclear Safety (BMUB) supports this initiative on the basis of a decision adopted by the German Bundestag. Also by projects 805-B3-413, 805-A9-224, 808-A9-180, 805-A9-532, 805-B3-600 and 808-B0-654, from the Center for Geophysical Research (CIGEFI) and the Center for Research in Marine Sciences and Limnology (CIMAR) of the University of Costa Rica (UCR). Thanks to the logistics support provided by the School of Physics of UCR. Thanks also to Dr. Pablo Imbach from CATIE for supplying much of the original raw GCM data. The authors thank Ricardo Herrera and Andrés Jiménez who formatted the data and collaborated in the calculation of some of the indexes. Thanks also to two anonymous reviewers whose comments considerably enhanced the quality and structure of this paper. Thanks to Michael Taylor and David Enfield for revising the manuscript. References Adler RF, Huffman GJ, Chang A, Ferraro R, Xie P, Janowiak J, Rudolf B, Schneider U, Curtis S, Bolvin D, Gruber A, Susskind J, Arkin P. 2003. The Version 2 Global Precipitation Climatology Project (GPCP) monthly precipitation analysis (1979–present). J. Hydrometeorol. 4: 1147–1167. Alfaro EJ. 2000. Response of air surface temperatures over Central America to oceanic climate variability indices. Tóp. Meteor. Oceanog. 7(2): 63–72. Alfaro E. 2002. Some characteristics of the annual precipitation cycle in Central America and their relationships with its surrounding tropical oceans. Tóp. Meteor. Oceanog. 9(2): 88–103. Amador JA. 1998. A climatic feature of the tropical Americas: the trade wind easterly jet. Tóp. Meteor. Oceanog. 5: 91–102. Amador JA. 2008. The intra-Americas sea low-level jet, overview and future research. Ann. N. Y. Acad. Sci. 1146: 153–188. Amador JA, Alfaro EJ, Lizano OG, Magaña VO. 2006. Atmospheric forcing of the eastern tropical Pacific: a review. Prog. Oceanogr. 69: 101–142. Bindoff NL, Stott PA, AchutaRao KM, Allen MR, Gillett N, Gutzler D, Hansingo K, Hegerl G, Hu Y, Jain S, Mokhov II, Overland J, Perlwitz J, Sebbari R, Zhang X. 2013. Detection and attribution of climate change: from global to regional. In Climate Change 2013: The Physical Science Basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change, Stocker TF, Qin D, Plattner G-K, Tignor M, Allen SK, Boschung J, Nauels A, Xia Y, Bex V, Midgley PM (eds). Cambridge University Press: Cambridge, UK and New York, NY. Brekke LD, Dettinger MD, Maurer EP, Anderson M. 2008. Significance of model credibility in estimating climate projection distributions for regional hydroclimatological risk assessments. Clim. Change 89: 371–394. Delworth TL, Rosati A, Anderson W, Adcroft AJ, Balaji V, Benson R, Dixon K, Griffies SM, Stephen M, Lee HC, Pacanowski RC, Vecchi GA, Wittenberg AT, Zeng FR, Zhang R. 2012. Simulated climate and climate change in the GFDL CM2.5 high-resolution coupled climate model. J. Clim. 25: 2755–2781. Enfield DB, Mestas-Nuñez AM, Trimble PJ. 2001. The Atlantic multidecadal oscillation and its relation to rainfall and river flows in the continental U.S. Geophys. Res. Lett. 28: 2077–2080. Flato G, Marotzke J, Abiodun B, Braconnot P, Chou SC, Collins W, Cox P, Driouech F, Emori S, Eyring V, Forest C, Gleckler P, Guilyardi E, Jakob C, Kattsov V, Reason C, Rummukainen M. 2013. Evaluation of climate models. In Climate Change 2013: The Physical Science Basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change, Stocker TF, Qin D, Plattner G-K, Tignor M, Allen SK, Boschung J, Nauels A, Xia Y, Bex V, Midgley PM (eds). Cambridge University Press: Cambridge, UK and New York, NY. © 2014 Royal Meteorological Society Gates WL, Boyle JS, Covey C, Dease CG, Doutriaux CM, Drach RS, Fiorino M, Gleckler PJ, Hnilo JJ, Marlais SM, Phillips TJ, Potter GL, Santer BD, Sperber KR, Taylor KE, Williams DN. 1999. An Overview of the Results of the Atmospheric Model Intercomparison Project (AMIP I). Bull. Am. Meteorol. Soc. 80: 29–55. Giorgi F. 2006. Climate change hot-spots. Geophys. Res. Lett. 33: 8707, doi: 10.1029/2006GL025734. Hidalgo HG, Alfaro EJ. 2012a. Some physical and socio-economical aspects of climate change in Central America. Prog. Phys. Geogr. 36: 379–399. Hidalgo HG, Alfaro EJ. 2012b. Global model selection for evaluation of climate change projections in the eastern tropical Pacific seascape. Int. J. Trop. Biol. Conserv. 60: 67–81. Hidalgo HG, Amador JA, Alfaro EJ, Quesada B. 2013. Hydrological climate change projections for Central America. J. Hydrol. 495: 94–112. Hirota N, Takayabu YN, Watanabe M, Kimoto M. 2011. Precipitation reproducibility over tropical oceans and its relationship to the double ITCZ problem in CMIP3 and MIROC5 climate models. J. Clim. 24: 4859–4873. IPCC. 1997. The Regional Impacts of Climate Change: An Assessment of Vulnerability. Regional Vulnerability to Global Climate Change, Watson RT, Zinyowera MC, Moss RH (eds). Cambridge University Press: Cambridge, UK. IPCC. 2013. Climate Change 2013: The Physical Science Basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change, Stocker TF, Qin D, Plattner G-K, Tignor M, Allen SK, Boschung J, Nauels A, Xia Y, Bex V, Midgley PM (eds). Cambridge University Press: Cambridge, UK and New York, NY. Jiang X, Waliser DE, Kim D, Zhao M, Sperber KR, Stern WF, Shubert SD, Zhang GJ, Wang W, Khairoutdinov M, Neale RB, Lee M. 2012. Simulation of the intraseasonal variability over the Eastern Pacific ITCZ in climate models. Clim. Dyn. 39: 617–636. Kalnay E, Kanamitsu M, Kistler R, Collins W, Deaven D, Gandin L, Iredell M, Saha S, White G, Woollen J, Zhu Y, Leetmaa A, Reynolds R. 1996. The NCEP/NCAR 40-year reanalysis project. Bull. Am. Meteorol. Soc. 77: 437–471. Kim HM, Webster PJ, Curry JA. 2012. Evaluation of short-term climate change prediction in multi-model CMIP5 decadal hindcasts. Geophys. Res. Lett. 39: L10701, doi: 10.1029/2012GL051644. Knutti R, Sedláček J. 2013. Robustness and uncertainties in the new CMIP5 climate model projections. Nat. Clim. Chang. 3: 368–373. Langenbrunner B, Neelin JD. 2013. Analyzing ENSO teleconnections in CMIP models as a measure of model fidelity in simulating precipitation. J. Clim. 26: 4431–4446. Liu HL, Zhang MH, Lin WY. 2012. An investigation of the initial development of the double-ITCZ warm SST biases in the CCSM. J. Clim. 25: 140–155. Magaña V, Amador J, Medina S. 1999. The midsummer drought over Mexico and Central America. J. Clim. 12: 1577–1588. Maldonado T, Alfaro E, Fallas B, Alvarado L. 2013. Seasonal prediction of extreme precipitation events and frequency of rainy days over Costa Rica, Central America, using Canonical Correlation Analysis. Adv. Geosci. 33: 41–52. Mantua NJ, Hare SR, Zhang Y, Wallace JM, Francis RC. 1997. A Pacific interdecadal climate oscillation with impacts on salmon production. Bull. Am. Meteorol. Soc. 78: 1069–1079. Martin ER, Schumacher C. 2011. The Caribbean low-level jet and its relationship with precipitation in IPCC AR4 models. J. Clim. 24: 5935–5950. Maurer EP, Hidalgo HG. 2008. Utility of daily vs. monthly large-scale climate data: an intercomparison of two statistical downscaling methods. Hydrol. Earth Syst. Sci. 12: 551–563. Maurer EP, Hidalgo HG, Das T, Dettinger MD, Cayan DR. 2010. The utility of daily large-scale climate data in the assessment of climate change impacts on daily streamflow in California. Hydrol. Earth Syst. Sci. 14: 1125–1138, doi: 10.5194/hess-14-1125-2010. Meehl GA, Covey C, Delworth T, Latif M, McAvaney B, Mitchel JFB, Stouffer RJ, Taylor KE. 2007. The WCRP CMIP3 Multimodel Dataset: A New Era in Climate Change Research. Bull. Am. Meteorol. Soc. 88: 1383–1394, doi: 10.1175/BAMS-88-9-1383. Neelin JD, Langenbrunner B, Meyerson JE, Hall A, Berg N. 2013. California winter precipitation change under global warming in the coupled model intercomparison project phase 5 ensemble. J. Clim. 26: 6238–6256, doi: 10.1175/JCLID1200514.1. Pierce DW, Barnett TP, Hidalgo HG, Das T, Bonfils C, Sander B, Bala G, Dettinger M, Cayan D, Mirin A. 2008. Attribution of declining western US snowpack to human effects. J. Clim. 21: 6425–6444. Int. J. Climatol. (2014) SKILL OF CMIP5 CLIMATE MODELS IN CENTRAL AMERICA Pierce DW, Barnett TP, Santer BD, Gleckler PJ. 2009. Selecting global climate models for regional climate change studies. Proc. Natl Acad. Sci. USA. 106: 8441–8446. Polade SD, Gershunov A, Cayan DR, Dettinger MD, Pierce DW. 2013. Natural climate variability and teleconnections to precipitation over the Pacific-North American region in CMIP3 and CMIP5 models. Geophys. Res. Lett. 40: 2296–2301, doi: 10.1002/grl.50491. Rauscher SA, Kucharski F, Enfield DB. 2011. The role of regional SST warming variations in the drying of Meso-America in future climate projections. J. Clim. 24: 2003–2016, doi: 10.1175/2010 JCLI3536.1. Sachs JP, Sachse D, Smittenberg RH, Zjang Z, Battisti DS, Golubic S. 2009. Southward movement of the Pacific intertropical convergence zone AD 1400–1850. Nat. Geosci. 2: 519–525, doi: 10.1038/NGEO554. © 2014 Royal Meteorological Society Sheffield J et al. 2013. North American climate in CMIP5 experiments, part I: evaluation of historical simulations of continental and regional climatology. J. Clim. 26: 9209–9245, doi: 10.1175/JCLID1200592.1. Smith TM, Reynolds RW. 2004. Improved extended reconstruction of SST (1854–1997). J. Clim. 17: 2466–2477. Taylor M, Alfaro E. 2005. Climate of Central America and the Caribbean. In Encyclopedia of World Climatology, Oliver JE (ed). Springer: Amsterdam, 183–189. Taylor KE, Balaji V, Hankin S, Juckes M, Lawrence B, Pascoe S. 2012a. CMIP5 and AR5 data reference syntax (DRS) and controlled vocabularies. http://cmip.llnl.gov/cmip5/ (accessed 6 January 2013). Taylor KE, Stouffer RJ, Meehl GA. 2012b. A summary of the CMIP5 experiment design. Bull. Am. Meteorol. Soc. 93: 485–498. Waylen PR, Quesada ME, Caviedes CN. 1996. Interannual variability on monthly rainfall in Costa Rica. J. Clim. 9: 2606–2613. Int. J. Climatol. (2014)
© Copyright 2026 Paperzz