Review of MWR 2112 (BS) Hines “Development and Testing of Polar WRF: Part I: Ice Sheet Meteorology” Reviewer A Recommendation: Accept with major revisions as outlined below. General Comments: The paper presents a study examining the performance of the Weather Research and Forecasting (WRF) mesoscale model over ice sheets in Greenland during December 2002 as the first step towards development of the so-called “Polar WRF”. As opposed to another recent paper I reviewed that dealt with high latitude applications of the WRF model, this paper was generally disappointing, in that arguments and conclusions strike me as a bit weak and leave me with many questions, which are addressed below. Some additional work is needed to better substantiate the conclusions reached. Also, many regional climate models have been applied to this area at much greater horizontal resolution than 40km, suggesting important needs for parameterization development to properly simulate processes at a hierarchy of scales. It seems a stretch to suggest that tests at 40km are a good indication of the results that regional climate modelers will achieve at the much higher resolutions they will want to run WRF at to examine some of the key unresolved questions that have been brought to the fore by SHEBA, FIRE-ACE and other recent field efforts. Specific Comments: 1. Page 3, section 1a. One wonders if the section heading a) is really needed since there is no b section. 2. Page 3, section 1a, paragraph 1: Second sentence is awkward grammatically, specifically the clause starting with “maintained”. Suggest a minor rewrite of the sentence to correct this. 3. Page 3, section 1a, bottom of page: Here this seems to directly refer to the AMPS forecasts…why not just state that Polar MM5 is utilized as the basis for AMPS? 4. Pages 4-7: The initial impression I had, upon reading the Polar MM5 section of the Introduction , was that the Polar MM5 modifications had been translated to Polar WRF. This is not at all clear from most of the Polar WRF discussion (save the land surface discussion) which reads mostly like an overview description of standard WRF options than anything else. Were modifications made to any of the other packages? 5. Pages 5-6, paragraph that is split over these two pages. Simply say: “The Advanced Research WRF (ARW) is a modular non-hydrostatic model…” and omit the first full sentence on page 6. 6. Page 6, first full paragraph. 28 vertical levels, with a top at 10 hPa and supposedly “high resolution” in the PBL (though more than the lowest two levels is not explicitly specified), suggests rather coarse resolution somewhere in the vertical column and leads one to wonder if some important processes in the mid and upper troposphere are properly represented. I am not sure of the need to cover so much of the lower stratosphere (~ 200 – 10 hPa) , which is generally only represented by satellite-retrieved information in most operational analyses. Some justification of the need for such a high model top would be welcome. 7. Page 7, bottom and all occurrences thereafter. “NOAH” is an acronym and should be represented as such. Also, why was the 5-layer thermal diffusion land surface model not tested given the period of interest is December and subsoil water is likely frozen, making the water phase changes represented in the other land models less significant? 8. Page 8, ending of section 2: The treatment of snowpack and heat transfer at the snow surface is at best interesting, at worst it may violate inherent assumptions in the NOAH model. What is the rationale for this approach, which comes across as rather ad hoc? At minimum a discussion of exactly what parameters and/or formulations within NOAH are impacted and how.. is appropriate. 9. Page 9: The crux of my resolution related comment above is here. While I understand a desire to have comparative tests with Bromwich et al (2001), I think tests at higher resolutions more likely to be utilized by regional climate modelers in an age of increased computer power at least relevant to attempt and could provide much useful information on the utility of this particular WRF configuration. Also see next comment. 10. Page 9: What is the resolution of the Ekholm (1996) digital elevation dataset? The author notes that 40 km resolution is not really sufficient to resolve the steepest margins of the ice sheet…how much spatial area is involved here and could it be a significant impact on the results. Again, at least one higher resolution experiment would clarify many of these issues. 11. Page 9: “Some surface energy balance terms are available…” Which ones? Shouldn’t verification be as consistent as possible across the very limited station base available (more on that later)? 12. Page 9, bottom: Many in the land surface modeling community feel that, for longer term climate/regional climate simulations, longer spin-up times (on the order of weeks) may be needed to truly reach an initial state that will be free of model shocks. This point is still debated to varying degrees, so some justification (maybe a sentence that is more definitive…ie, saying that Bromwich et al (2001) tested various spin up times and found 24 hours to be sufficient) is needed. 13. Page 10, primary paragraph. OK, this paragraph floored me. Unless verification is very carefully conducted with the disparate grid resolutions, we could well have an apples versus oranges situation w/r/t verification. It is not appropriate to assume equivalence between model quantities on grids of different resolutions as a unique indicator of relative performance. If anything, all that could be represented is the difference in grid resolution. The situation is further muddied by the need to time-average results and then therefore the observations. Many sins can be covered by such averaging processes, and it lowers my confidence of the significance of the results obtained. 14. Page 11, second paragraph. “…the optional WRF physics” makes it sound as if WRF can be run with no physics (which in some sense it can but that is not the point here).. Better to say “the WRF physics options”. 15. Page 12: Some discussion of how the interpolated values for Swiss Camp are derived is in order, given the sharp terrain slope, which, as was pointed out earlier, may not be well-represented at the 40km resolution. Such interpolation usually needs to be conducted very carefully as to not introduce bias, since standard relationships for reduction to 2- or 10-m values often can produce overor undershoot of values. Numerous papers related to similar issues with the Tibetan Plateau are in the literature. 16. Page 12 and hereafter. While averaging of the error statistics to produce a general picture of performance is often done, it can often hide a multitude of sins, especially for such a small number of stations. Are these results statistically significant? Can statistically significant verification even be done with such a small verifying dataset? Significance testing, for individual stations over the month-long period as well as the aggregate, is needed for there to be real confidence in the validation. 17. Page 13. I question whether the results for one point (Swiss Camp) are really sufficient basis to eliminate certain configurations. And the question also arises as to whether the simulation of the katabatic flow regime, important as it may be , and as represented by the wind speed simulations at a single point, is the most important criteria for disregarding the other simulations. Given that the large domain covers a much larger area than this point realistically represents from a meteorological point of view, this may be throwing out the baby with the bath water. 18. Page 13, discussion of Figure 6. Of course absolute correlations will likely be high, as the computation implicitly includes the impact of the entire column as opposed to the small variations. Anomaly correlations would provide a truer sense of the performance of model in simulating pressure field variations. 19. Page 13. Similarity of correlations does not rigorously equate to independence of the temporal structure of the errors. Independence should be more rigorously tested if claimed. 20. Page 13. 2nd sentence from the bottom of the page. This is a curious point. Since the MM5 also comes equipped with the RRTM longwave radiation package, would simulations using Polar MM5-RRTM produce similar results to the WRF configuration? If so, does that lessen the uniqueness of the improvements to the simulation via the use of WRF (and thus partially negating the rationale for the development of Polar WRF). A comparison run here would be very illuminating. 21. Page 14. I understand what the author is trying to do here, but the fact that the Reanalysis vertical profiles perform worst near the surface leads one to suspect that any comparison of PBL vertical structure between WRF and NCEP-2 is at best tenuous. And how much does the resolution of the NCEP-2 reanalysis play into this assessment? The sentence at the end (which continues onto page 15) adds nothing to the discussion at hand and should be omitted. 22. Page 15, first full paragraph. Grammatical error, 2nd sentence: “…the western Greenland slopes Swiss Camp….” 23. Page 15. Are the 10-m wind speeds for the Polar MM5 and Polar WRF both computed directly as output fields, or are they obtained by interpolation, and if the latter, is the interpolation algorithm the same? Otherwise the interpolation algorithm could be introducing relative bias into the comparison. 24. Page 15, bottom paragraph. “Polar WRF simulates a larger speed by about 2 m/s in the upper troposphere at Swiss Camp”. I see a larger speed all the way down to 850 hPa, which is hardly the upper troposphere. 25. Page 16, section 5. It seems very odd to focus so much attention on winter verification at Swiss Camp while the summer verification is for a different station (Summit). What is the rationale for this? Are we just being shown the “good” results, carefully selected to give the best possible impression of WRF? 26. Page 17. Top paragraph discussing radiation. This is interesting, but what is the physical process that is poorly simulated in Polar MM5…are the cloud bases too high (thus colder clouds and less downwelling IR)….is cloud coverage (# hrs/day) too little? Poor track forecasts of synoptic systems (leading to one or the other of the previous possibilities) It is hard to determine where a parameterization might be failing if the details are not examined. Did you utilize available satellite imagery to determine if the cloud coverage had spatial/temporal errors? 27. Page 17. How is the diurnal cycle smoothed? It could play a role in the results and interpretation, particularly the point at the bottom of the page talking about the damping of the diurnal cycle in the WRF runs. 28. Page 17, bottom of 1st full paragraph. Why not show what the differences in the cloud fields (even averaged ones) are? That way you can make conclusions instead of merely speculating. 29. Page 18, 4th line. “We shall use this opportunity…” sounds a bit like the author is giving a lecture. It is overly grandiose for a written article. 30. Page 18, last line, 1st full paragraph. Awkward. Better: “…simulations will employ the Goddard shortwave scheme in response to the results shown in Figure 13”. 31. Page 18, last paragraph through page 19. “…initial subsurface temperature….monthly average skin temperature from an earlier WRF run”. How much earlier? What is the justification? This is an approach that is more suited to operations rather than a rigorous model performance research study. What is the exact formula used? All of this certainly seems a lot like “tuning” and given it is only applied based upon the results at a single point, is not likely to be generally valid. Is radiation allowed to penetrate the snowpack as indeed happens to some extent in the real world? NOAH would likely not be designed to include this under the assumption that radiation strikes a ground surface and does not penetrate to any depth. 32. Page 20. Given the importance of the cloud mixing ratios to this discussion, inclusion of ice, graupel and rain may still be significant from a radiative point of view. If need be, they should be plotted on a separate graph. 33. Page 20: “the boundary layer cloud…not adversely impacting the incident shortwave radiation” Direct, diffuse or total? This seems like an odd result. 34. Page 20: It may be an exaggeration to say that Greenland has a pristine atmosphere….I believe…though I may be wrong…in saying that Arctic haze has occasionally spread as far as Greenland, not to mention the occasional, though very infrequent influx of volcanic ash from nearby Iceland. More importantly in this same paragraph, why do you want to encourage moisture loss from the atmosphere? You do not state that this is necessarily a deficiency in any of the simulations. What is the physical basis for this? 35. Page 21. OK, now Swiss Camp comes back into the validation…why? 36. Page 22, top. Based upon Figure 16, you can only claim that Polar MM5 better simulates the summer surface layer , not necessarily the entire boundary layer, which is of a greater depth and not necessarily adequately evaluated solely by looking the surface level. 37. Page 22, bottom-Page 23. This is really a stretch. The values for December 2000 and 2001 are only representative if the mean synoptic/mesoscale patterns (including cloud cover, etc) come somewhat close to the patterns for December 2002. You have provided no indication, via figure or discussion that this is the case. Ditto for the use of January 2006 observations. Unless you can demonstrate equivalence, this entire discussion is invalid.
© Copyright 2026 Paperzz