Weather or Not

Severe Outflow by R. Edwards

  • Home
  • About
  • Archives

Powered by Genesis

Human Weather Forecasting in an Automation Era, Part 2: Lessons of Air France 447

August 26, 2022 by tornado Leave a Comment

This short series (go to Part 1) arises from the recently published paper, “The Evolving Role of Humans in Weather Prediction and Communication“. Please read the paper first.

The authors briefly mention the need for forecasters to avoid the temptation to get lazy and regurgitate increasingly accurate and complex postprocessed output. I’m so glad they did, agree fully, and would have hammered the point even harder. That temptation only will grow stronger as guidance gets better (but never perfect). To use an example from my workplace, perhaps in 2022 we’re arriving at the point that an outlook forecaster can draw probabilities around ensemble-based (2005 essay), ML-informed, calibrated, probabilistic severe guidance most of the time and be “good enough for government work.”

Yet we strive higher: excellence. That necessarily means understanding both the algorithms behind such output, and the meteorology of situations enough to know when and why it can go wrong, and adapting both forecast and communication thereof accordingly. How much of the improvement curve of models and output vs. human forecasters is due to human complacency, even if unconscious? By that, I mean flattening or even ramping down of effort put into situational understanding, through inattention and detachment (see Part 1).

It’s not only an effect of model improvement, but of degradation of human forecast thinking by a combination of procedurally forced distraction, lack of focused training on meteorological attentiveness, and also, to be brutally honest, culturally deteriorating work ethic. I don’t know how we resolve the latter, except to set positive examples for how, and why, effort matters.

As with all guidance, from the early primitive-equation barotropic models to ML-based output of today and tomorrow: they are tools, not crutches. Overdependence on them by forecasters, being lulled into a false sense of security by their marginally superior performance much of the time, that complacency causing atrophy of deep situational understanding, invites both job automation and something perhaps worse: missing an extreme and deadly outlier event of the sort most poorly sampled by ML training data.

Tools, not crutches! Air France 447 offers a frightening, real-world, mass-casualty example of this lesson, in another field. Were I reviewing the Stuart et al. AMS paper, I would have insisted on that example being included, to drive a subtly made point much more forcefully.

The human-effort plateau is hidden in the objective verification because the models are improving, so net “forecast verification” appears to improve even if forecasters generally just regurgitate guidance and move on ASAP to the next social-media blast-up. Misses of rare events get averaged out or smoothed away in bulk, so we still look misleadingly good in metrics that matter to bureaucrats. That’s masking a very important problem.

Skill isn’t where it should or could be, still, if human forecasters were as fully plugged into physical reasoning as their brain capacity allows. The human/model skill gap has shrunk, and continues to, only in part because of model improvements, but also, because of human complacency. Again, this won’t manifest in publicly advertised verification metrics, which will smooth out the problem and appear fine, since the model-human combination appears to be getting better. Appearances deceive!

The problem of excess human comfort with, and overreliance on, automation will manifest as one or more specific, deadly, “outlier” event forecasts, botched by adherence to and ignorance of suddenly flawed automated guidance: the meteorological equivalent of Air France 447. This will blow up on us as professionals when forecasters draw around calibrated-guidance lines 875 times with no problem, then on the 876th, mis-forecast some notorious, deadly, economically disastrous, rare event because “the guidance didn’t show it.”

That disaster will be masked in bulk forecast verification statistics, which shall be of little consolation to the grieving survivors.

Consider yourself warned, and learn and prepare accordingly as a forecaster!

More in forthcoming Part 3…

Filed Under: Weather Tagged With: analysis, automation, communication, communication skills, education, ensemble forecasting, forecast uncertainty, forecaster, forecasting, meteorology, operational meteorology, science, severe storms, severe weather, understanding, weather

Dangerous Crutch of Automation in Weather Forecasting

October 27, 2016 by tornado Leave a Comment

In the last installment of “Scattershooting”, I offered the following short take regarding the evolving role of humans on the operational forecasting process. I’m reproducing it again here for additional elaboration based on input received since:

…

“IDSS” — a bureaucratic acronym standing for “Integrated Decision Support Services” — is the pop-fad buzzword in today’s public weather services. Behind all the frivolous sloganeering (“Weather Ready Nation”) and even more needless managerial lingo bingo, is a nugget of extreme truth: the science of operational forecasting must improve its messaging and communications capabilities. But it is possible to swing the pendulum too far, to turn professional forecasters into mere weather briefers (full-time IDSS) and let “forecasts” be straight model output. That would be a degradation of service, because models still cannot forecast multivariate, greatest-hazard events in the short term (tornado risks, winter weather, hurricane behavior, etc.) as well as human forecasters who are diligently analyzing and understanding the atmosphere. Today’s tornado and hail threat is not the same animal as single-variable temperature grids, and is far, far, far, far, far more important and impactful.

There is still a huge role in the forecast process for human understanding. Understanding has both physical and empirical bases — the science and art of forecasting. Models of course are great tools — when understood and used well. However, models themselves do not possess understanding. Moreover, models do not issue forecasts; people do. A truly excellent forecast is much more than regurgitation of numerical output in a tidy package. It is an assembly of prognostic and diagnostic understanding into a form that is communicated well to customers. A forecast poorly understood cannot be communicated well. And a bad forecast that is masterfully conveyed and understood is a disservice regardless. Eloquent communication of a crappy forecast is akin to spray-painting a turd gold. It is still a smelly turd.

Solution? Get the forecast right. All else will fall into place. Credibility is top priority! As for automation, it will proceed at the pace forecasters permit their own skills to atrophy, both individually and collectively. For those unfamiliar as to how, look up the prescient meteorological prophet Len Snellman and his term, “meteorological cancer”.

…

Now the elaboration:

Patrick Marsh supplied a remarkably insightful article on the role automation and resulting ignorance played in plunging Air France Flight 447 (Rio to Paris) into the Atlantic: Crash: How Computers Are Setting Us Up for Disaster. Please read that before proceeding; somewhat long but certainly riveting, the article should be well worth your time.

This has important ramifications to hazardous-weather forecasting too. It fits in nicely with a lot of concepts in meteorology, as a whole, about which I have been thinking and writing for many years (e.g., here, and here, and here, and here), and as have even more-accomplished scientists (e.g., here). Let’s hope the “powers that be” read these lessons and pay attention.

In the meantime, even though it has nothing directly to do with severe midlatitude storms or other hazardous-weather forecasting done out of Norman, the Air France crash article offers such a powerful and plainspoken lesson that it could and arguably should be included in all new-forecaster training materials. My unnamed office has remained ahead of the automation monster and in control of it, because the forecasters see and understand its limitations on a daily basis, as with (for example) SREF calibrated probabilities and so-called “convection-allowing models” (CAMs) at fine grid spacing, such as the SSEO members and the NCAR ensemble. My office is at the forefront of how to help to develop the tool and use it in its proper role, without dangerous over-reliance; the model is the slave, not the human. Let’s keep it that way.

One way we can do so as forecasters is through continuing to provide meaningful developmental and testing input, as we have. This way we still understand the monster we help to create, and know when and how to employ its growing powers. Another, even more important way is by maintaining the high level of both baseline meteorological understanding and diagnostic situational awareness of events for which we are properly well-known and -respected. As I have stated resolutely elsewhere: A forecaster who tries to predict the future without a thorough understanding of the present is negligent, and prone to spectacular failure.

Keeping up with the science, and maintaining and improving fundamental skills, are so important. How? Reading, training and practice…reading, training and practice! Writing (research papers) helps immensely too, in literature review and assembly. Forecasting experience matters too, as long as it involves experiential learning and progress, not one year experience repeated over and over! Again, they key word and concept here is understanding. When forecasters’ core abilities atrophy from disuse and weaknesses creep in, perhaps even unrealized (as was the case on that A380 disaster), is when the atmosphere analogously will make us pay.

Nearly four decades ago (Snellman 1977, Bulletin of the AMS), visionary scientist Len Snellman foresaw this when he coined the term, “meteorological cancer”, to describe the threat of over-reliance on automated output to our ability to predict the weather. This can include the extreme and/or exceptional events, the “money events” that disproportionately cause casualties and destroy property. What is my money event? The tornado outbreak, and to some extent the derecho.

Since such events are, by nature and definition, extreme and unusual, an “average” model forecast might not get there. Predicting the mean on a day with little precedent can be horribly wrong. Sometimes the ensemble outlier is the closest solution, not the ensemble mean or median. We will need to be able to recognize in advance when that could be reasonably possible. [Emphasis on “reasonable”–not “you never know, you can’t rule it out” grade of total CYA over-emphasis on the outliers.] Our probability of recognizing the correctness of the outlier increases with our understanding, both of the ensemble and of the meteorology behind the exceptions, then in turn, our ability to diagnose when the model is veering or could veer rapidly awry, thereby nailing a once-in-career success and saving some lives to boot. Either that, or go down like that plane and suffer spectacular, career-defining failure…

We can and must prevent the latter, on both individual and institutional levels. Meteorological understanding and diagnostic ability don’t guarantee success, but they make good forecasts much more likely and consistent. When we as a science will get in trouble is when we start treating automated output as an unquestioned crutch instead of as a tool in a holistic toolbox that still involves human diagnostics and scrutiny. Is that skill being lost in the field with less-experienced forecasters who have never had to fly anything but autopilot? Will “SuperBlend” become “SuperCrutch”? If not, when will we reach that tipping point, what sort of disaster will it take to reveal such a failure, and how can we see it coming in time to change course and lift that altitude? Vague as the proposal is for now regarding specifics on forecaster duties and staffing and office-hour reductions, does “Operations and Workforce Analysis” (OWA, NWS document and NWSEO union response), for all its superficial and advertised benefits for “IDSS”, carry a dark undercurrent of swinging the automation pendulum on the forecast side too far toward that fate?

These questions and more need to be discussed. Now that truly would be (to use a piece of bureaucratic jargon) a “vital conversation”.

Filed Under: Weather Tagged With: automation, ensemble forecasting, forecaster, meteorological cancer, meteorology, numerical guidance, operational meteorology, research, science, severe storms, severe weather, situational awareness, weather forecaster

Forecasting the Ensemble Outlier

February 9, 2011 by tornado 1 Comment

Several days back, I posted some thoughts about a pretty good short-range ensemble (SREF) spot forecast of a recent snow event in the Norman area. Now let’s look at even more recent ensemble forecasts that were absolutely wretched–so much so that even the most extremely low outlier in the latest of the two forecasts was overdone by a factor of nearly 2.5!

When is forecasting at or even beyond the outlier the best choice? When it’s the closest to correct! Granted, that’s easy to say in hindsight; but the cold truth is that the atmosphere does hold the answer key. Forecasts are judged in an objective sense by their closeness to the verifying quantity. In this case, it’s accumulated snowfall.

Below please witness two different SREF runs’ predictions for Norman snow, in the form of plume diagrams (explained in the previous link above). The answer key, in the form of the average snow depth measured at my house, is in magenta. [NOTE: The vertical scales are not exactly the same. Look at the numbers at left, in inches.]

Model initial hour 21Z Monday:

Model initial hour 15Z Tuesday:

The best spot forecast here clearly would have been for the greatest probabilities and/or snowfall ranges skewed strongly toward, or even beyond, the low end of the distribution. Following the cliff-leaping lemmings of ensemble consensus here would have been a miserable forecast failure. There probably are many more lessons here than my fading, sleep-deprived mind can muster, but I can offer two for now:

  1. The outlier sometimes is the best solution! Forecasters get paid to produce the best possible prediction; and sometimes that means forecasting way above or below the average or “consensus” model forecast. Clearly leaning heavy toward the low extreme would pay off here. I don’t pretend it’s easy; but since when did anyone promise that forecasting is supposed to be? For more discussion on this issue, see this 5-year old post on ensemble forecasting–specifically, scroll down to the “Consensus Forecasting vs. Extreme Event ‘Outliers'” section.
  2. Spot forecasting for something like forecasting a banded snow event, which can vary wildly based on subtleties that observations or models cannot resolve well, probably is a foolhardy endeavor. Parts of northeastern Oklahoma have had over 20 inches of snow from the same event! Norman stayed outside the southern fringes of a persistent, mostly W-E aligned snow belt that belted parts of northern Oklahoma, southern Kansas and northwestern Arkansas.

Filed Under: Weather Tagged With: ensemble forecasting, ensemble mean, ensemble outlier, plume diagram, snow, snow forecasting, spot forecast, SREF, winter weather, winter weather prediction

  • 1
  • 2
  • Next Page »

Search

Recent Posts

  • Scattershooting 230128
  • A Thanksgiving Message
  • Human Weather Forecasting in an Automation Era, Part 3: Garbage In, Garbage Out
  • Human Weather Forecasting in an Automation Era, Part 2: Lessons of Air France 447
  • Human Weather Forecasting in an Automation Era, Part 1: Situational Understanding

Categories

  • Not weather
  • Photographic Adventures
  • Scattershooting
  • Weather
  • Weather AND Not
@SkyPixWeather

- January 30, 2023, 3:38 am

@cschultzwx @TwisterKidMedia So many holds don’t get called. That looked quite familiar. I know this as a Cowboys/Micah Parsons fan.
h J R
@SkyPixWeather

- January 30, 2023, 3:27 am

@TwisterKidMedia @sdantwx Worst officiating I’ve ever seen was in a college game too, and Andrew should know this one. https://t.co/pC5fTkFFrF
h J R
@SkyPixWeather

- January 30, 2023, 3:20 am

@TwisterKidMedia @tempestchasing I still don’t understand fully WTH happened w/the “unheard whistle” clock debacle and play that wasn’t. That was bizarre.
h J R

Blogroll

  • CanadianTexan
  • Chuck's Chatter
  • Cliff Mass Weather & Climate
  • Digital Photography Review
  • DMN Dallas Cowboys BLOG
  • Dr. Cook's Blog
  • Dr. JimmyC
  • E-journal of Severe Storms Meteorology
  • Eloquent Science
  • Image of the Week
  • Jack's Cam Wall
  • Jim LaDue View
  • Laura Ingraham
  • MADWEATHER
  • Michelle Malkin
  • Photography Attorney
  • Severe Weather Notes
  • SkyPix by Roger Edwards
  • Tornatrix
  • With All My Mind

Meta

  • Log in
  • Entries feed
  • Comments feed
  • WordPress.org