This is a guest blog by Bill Fothergill, also known as billthefrog.
He sent it to me a couple of weeks ago, but it's still topical.
-----
Problematic Predictions
Many of the contributors to the Arctic
Sea Ice Blog have passed comment on the dangers of basing an
end-of-melt-season prediction on anything as simple as the current
area or extent values. The variables that determine the eventual
outcome of the annual melt season are truly legion, and it is sheer
folly to think that a single parameter provides anything other
than a
tenuous glimpse at what eventually may transpire.
However, whilst few might disagree with the above assertion, it would obviously be better to base this claim on something a trifle more rigorous than a vague gut feeling or some unshakeable belief that the universe should conform to one’s prejudices. There are obviously many ways to perform some form of mathematical analysis on the available data, but an example of such a method would be to look at the correlation between various mean monthly values over a predetermined time frame.
In the example outlined here, the NSIDC monthly area and extent values are used to provide the data, and the CORREL function in Excel is used to calculate a type of relationship known as the correlation coefficient. The dataset(s) were read into a rectangular array with “months” along one dimension and “years” along the other such that, say, the April extent for each year from 1979-2012 could be compared in turn with its respective equivalent figure for May, June, July etc.
Correlation coefficients can take on any value between +1 and -1. A value of +1 would indicate a perfect functional relationship between the sets of figures. For example, every value for, say, July might be exactly 25% larger than the corresponding value for August. Similarly, if every August value was, say, 1 million square kilometres above its corresponding September figure, then the correlation coefficient would again be +1. (This will of course be recognisable to many as our dear old (dys)functional friend “y = mx + c”.) As the correlation gets weaker, the value of the coefficient decreases, and eventually will become negative if a rise in one variable is matched by a drop in the corresponding partner value.
As those living in the real world know all too well, the anomalies over the summer and autumn (fall) seasons having been getting rapidly larger of late. This then begs the question “how does this impact upon the predictive skill of any early-season value?”
Initially, I simply repeated the correlation exercise on a subset (2007-20012) of the overall dataset. (The reason for this choice of subset should be pretty obvious.) However, as this contained just 6 years worth of data, there was a concern that any result could be interpreted as being only an artefact of the selection period. (i.e. inadvertent cherry picking) In order to allay any such fears, two further subsets were also included by way of comparison.
Table 1: Correlation
Coefficients for the periods indicated.
Data based on NSIDC monthly
Arctic averages.
Although the above table might seem somewhat confusing to those unversed in this form of analysis, some patterns should nevertheless be pretty obvious. Possibly the most striking is that, almost without exception, the coefficient based on area is significantly higher than that based on extent. In other words, when trying to predict what may transpire later in the year, one is more likely to meet with some degree of success if one concentrates on area, rather than extent.
One can also clearly see that the correlation coefficients pertaining to any of the subset periods are markedly lower than those for the entire 1979-2012 dataset. In other words, it has become more difficult to predict end-of-melt-season outcomes. (Assuming one’s prediction is predicated solely upon present extent/area data.) Although the coefficients for both area and extent drop significantly in each of the subsets, the deleterious effect on the extent coefficient is much more marked.
The third obvious thing that can be seen from the above is that the correlations grow weaker as the distance between the two sets of months grows. For example, the May: June correlation is stronger than that for May: August. This is of course exactly what one would expect. The Arctic sea ice can be regarded as a “system” that exhibits chaotic behaviour – as the time period increases, so does the scope for the unexpected.
By filtering and sorting, additional information can be gleaned as follows…
Adjacent Months
Table 2: Comparison of coefficients for adjacent months
Taken over the entire dataset, the adjacent month correlations are pretty strong. This behaviour is generally what one might expect and is an example of “auto correlation” – basically if the level of ice is high (or low) one month, then it is very likely to be somewhat similar the following month. However, if one looks at the subset figures, even this normally tight relationship appears to have broken down, particularly around the April – June period.
Two and three month separations
The above can obviously be repeated for non-adjacent months as shown below…
Table 3a and 3b: Comparison of coefficients for non-adjacent months
It can clearly be seen that, even when looking at the entire dataset, the April coefficients are distinctly lower than those starting in subsequent months. However, in more recent years (as demonstrated by the various subset values) it is becoming necessary to wait until June or even July before one can make a reasonably reliable prediction for even just two or three months ahead. The presence of negative correlations can also be seen in the April and May figures.
September minima
Inevitably, people want to get an early feel for what the situation is likely to be at the time of the September minima. One can easily extract the relevant coefficients for just this purpose…
Table 4: Coefficients for September and a variety of starting months
Even prior to recent years, the April and May figures appear to have been of limited value when trying to predict the September value. However, as can be seen from the amount of red in the above table(s), April and May seem to have become almost worse than useless as an indicator of what will happen later.
Summing up, attempts to predict the September extent based solely on the level of ice in April or May is basically a fool’s errand. Although there is a reasonable chance of being close based upon June figures (much more so for area, rather than extent) one really needs to wait until July before betting the family silver.
The physical structure of the ice, and the salinity of the surrounding sea water will have large effects, as well. The volume of the ice is probably the best measure; much better than area alone.
Neil
Posted by: Neil Blanchard | June 30, 2013 at 12:46
Very interesting that there are so many negative correlation values.
I think the approach of Kaleschke and Spreen in their "Sea Ice Outlook 2010" (see also
ftp://ftp-projects.zmaw.de/seaice/prediction/ ) is better, and reduces artifacts of ice in areas that will be gone by September no matter what (e.g. Hudson Bay, Bering Sea).
But they haven't updated their ongoing outlooks page to include 2013 yet :(
Posted by: Nightvid Cole | June 30, 2013 at 14:19
This is a nice analysis, and I hope readers can take a few minutes to look over it closely, including the tables of coefficients.
If “chaotic behavior” is hard to visualize, another way of picturing the change is that Arctic sea ice has become more like a pawn, at the mercy of larger forces at work. That is just a metaphor, of course, but in physical terms, it makes sense that as ice gets smaller and weaker, it is more easily pushed around.
Posted by: Rick Aster | June 30, 2013 at 15:06
"In other words, it has become more difficult to predict end-of-melt-season outcomes"
This is not a conclusion that can be drawn from the data shown here. To see why, consider the this data set: (10, 11), (11, 10), (10, 10) and (11,11). There is no correlation. And this dataset has no correlation either: (0, 1), (1, 0), (0, 0) and (1,1). Now, if you add these datasets together, suddenly there is a good correlation: a high number on the first datapoint always goes together with a high number on the second, and a low number with another low number.
This is exactly what happened in the arctic. In the 80s, there was high spring extent and high summer extent. In the '07-'12 period, there was much lower spring extent and summer extent. Apart, there is no correlation (natural variability overcomes the trend). Taken together, there is a correlation (the trend overcomes natural variability).
This phenomenon occurs in every dataset with a trend and random peturbations, so in itself this tells nothing about the predictability of the end-of-melt-season outcomes. A better way to test that hypothesis is to try to 'predict' each year by using the data of the other years, and see if there's a trend in the quality of the predictions. If people are interested, I can do that kind of analysis, but I suspect that any trend found is not significant. There's not enough data.
Posted by: Bvanderspoel | June 30, 2013 at 15:17
Nice post. The sea ice extent in the early melting season is indeed a poor predictor for the September mean extent.
The PIOMAS volume and the long term models (like Gompertz curves) have been more reliable. They are not predicting a new record for 2013 (yet). Based on the latest PIOMAS update.
The ice conditions this year are causing a high extent compared to volume. This pattern might change by September. But a complete pattern reversal seems unlikely. We'll have to wait and see.
Posted by: Mignonette | June 30, 2013 at 15:22
Fred,
Did you detrend before taking correlations?
I've previously posted a graph of correlations in 10 day increments with the ice area at 7/9/XX.
http://farm4.staticflickr.com/3744/8960722115_256a90cb1b_o.jpg
But this was done with the difference between each successive year. In other words I have a table of years in columns with rows being ten day increments of CT Area. From that I make a secondary table in which each value is the difference from the previous year. i.e. from 1980 to 2012.
It is on that table of differences that I calculate the correlations, because the correlation equation gives spurious results in the presence of trends.
When I calculate the correlations in a likewise manner to the above graph, but for just a five year period, I get a noisier plot (to be expected), but a plot that more or less tracks the evolution of correlations seen above.
I think your shorter periods show less correlation because the longer ones contain more of a trend.
Is the trend in CT Area greater than that for NSIDC Extent? This is implied by the ratio of the two indices. If so that would explain the difference between extent and area.
Posted by: Chris Reynolds | June 30, 2013 at 15:29
I miss Intrade. Last year, I cleaned up on the minimum extent markets.
While you guys are, of course, trying to make accurate predictions based on science, if we ever get any kind of prediction markets like Intrade back, the key to making money on betting on this (and I recommend against betting the family silver) is to keep going long on the low numbers for years. Over time, you'll win.
Just like with the GISS temp markets, or the last named storm of Atlantic hurricane season markets (much more risky, but over time, if you bet the signal, you'll make up for your loss on noise).
The real trick is convincing the deniers to put their money where their mouths are. It's amazing how many of them are too chicken to bet--proof that they don't really believe their own BS.
Posted by: Ssupak | June 30, 2013 at 18:12
I find this exceptionally interesting, but for different reasons.
It seems to me that every estimate is little more than a guess. OK pure statistical analysis does come closer to the actual end result than modelling, but pure statistical analysis did not predict 2005, 2007 or 2012. It can’t. Because it can only model on a trend. Events which are driven by factors we don’t understand and can’t predict, are not present in a statistical model. They can’t be.
In the don’t know camp we have
W/m^2 on average for the summer
Solar output
Weather events (including summer cyclones and blocking high’s)
El Nino
La Nina
Ice mechanics in dangerously weakened ice
Impact of Local methane
NAO future impact
In the “known for certain” camp we have
What happened last time
To me, to take the ONE variable we DO know and subject it to rigorous statistical analysis is simply creating a formula which states that
We are absolutely certain that our results equal “Don’t Know^3”
Every analysis, every prediction, every heavily worked out massaged and strictly analysed offering starts with. “If we see x y or z then we will get”.
It seems to me that the stricter the rules of analysis and the more time and effort put into it, the wider the margin of the guess. Because, on a year by year basis, the predictions are nothing more than a guess. Whether they are a better guess (more reasoned with more data), or a worse guess (gut call), or in the middle guess (taking all the changes into account I think….), they are nothing more than a guess. Because we do not clearly understand and cannot model, on a day by day basis, the environment that will unfold in each melting season.
Otherwise the climate scientists would beat us every time. They’ve had decades to work on all of this….
Posted by: NeilT | June 30, 2013 at 18:56
I see potential outcomes of melting seasons (extent/area minimum) as a battle between weather conditions and ice thickness/quality. Until a couple of years ago the minimum was determined almost entirely by weather conditions during the different phases of the melting season.
Last year we saw clearer than ever (more clearly than in 2010 and 2011) that ice thickness/quality has diminished to a point where it starts to challenge the dominance of weather conditions.
This year - with a start that was incredibly non-conducive to extent/area decrease - might turn out to be further evidence of this process, although it's too early to tell. I think we'll know more at the end of July.
Posted by: Neven | June 30, 2013 at 21:12
Neven;
I am pretty sure we know a lot more by the end of July ;)!
Posted by: Espen Olsen | June 30, 2013 at 21:50
I am reminded by my explorations a few years ago into algorithms for day trading.
The issue we are wrestling with is very similar, interestingly enough; it is the dynamic relationship between short term and long term trends in a system with multiple inputs, each of which *themselves* may be cyclical, and be modified by feedback from others.
In day trading, the result is almost indisinguishable from a casino; I found that with the inputs available, the output/success was essentially random, regardless of how involved the algorithm applied (No, I did not lose any money; I ran the models first....)
Chris Reynolds said I think your shorter periods show less correlation because the longer ones contain more of a trend.
I think this captures the essence of it. Short term trends are decoupled from the specific outcome of the longer term cycle. Further, interaction between aspects of the system could multiply or cancel out others contributing to the outcome. For example, as a practical example a long enough anticyclone over the Beaufort, or pan-arctic dipole, could cancel and roll back the negative feedback of cloud and low temperatures in late May and early June which preserved ice extent and concentration. Conversely, a mild arctic cyclone forming might further ablate the heat entering the arctic, and form a bastion against rising continental temperatures on the margins.
Those I think are the phenomena which in short term directly affect the end of season state. So to that end, the answer lies in understanding the dynamics of the atmosphere, and over the short term (solstice to solstice) determining what effects it will have on the conditions which either strengthen or weaken the ice. At any given juncture, the state of the ice is mostly independent of the weather conditions immediately to follow. To that end, I'd suggest we'd benefif by examining year over year weather data as a factor for correlations, as well as the state of the ice.
Posted by: jdallen_wa | June 30, 2013 at 23:07
I claim 0 expertise here.
But it strikes me that, besides weather, the _quality_ of the ice--how slushy, briny, thin, full of holes..--is becoming a much larger factor than the quantity measures of extent, area, and even volume that we're used to dealing with.
But, perhaps with the exception of thickness, we have few solid numbers to attach confidently to these more qualitative aspects of the sea ice.
We do, though, have a pretty good general idea that these qualities are generally deteriorating quite markedly.
That's why I agree with some other here and on the blog that we are just a few days or weeks of strong sunshine away from a major cliff, a new minimum and maybe even something more stunning. But I also agree that things are more uncertain and harder to read than ever, especially within any one year.
Posted by: wili | July 01, 2013 at 01:03
Bill, thank you for a great post.
I've been looking at these correlations of early area/extent as predictors for later area/extent as well, but from the perspective of physical effect of solar energy uptake during the melting season.
Specifically I focused on May and June as predictors of September ice extent, so my analysis does not have the 'width' of yours, although it has some 'depth' that I think may be valuable in this analysis.
Specifically, I worked with 4 variables (predictors) that should (on the basis of laws of physics of albedo effect) affect energy uptake and thus ice melt during the melting season :
1) Northern Hemisphere snow cover
2) Arctic sea ice area
3) Effect of the area of polynia and melting ponds throughout the ice extent, which I think is pretty well captured by the simple relation : (ice_extent - ice_area).
4) ice thickness of the ice that will melt before September.
I calculated (theoretically, and with a lot of guessing) how much of an effect each one of these variable should have on the September minimum, and then looked for evidence of these physical effects in the correlation data and linear regression results of observational data.
The results are quite interesting.
For starters, I can confirm that 'area' is a better predictor for the later state of the ice than 'extent', simply from theoretical physics point of view : area determines how much energy gets absorbed, and thus how much ice will melt later on, while 'extent' can vary simply how the wind blows, and does not (for the part of extent that does not include area) affect the amount of energy absorbed.
Secondly, I get a better correlation if I include early "snow cover" (using Rutgers data) than if I just include 'area' as a 'predictor'.
The effect is not large for Jun->Sept prediction, but it is significant, and for May, it seems that 'snow cover' is dominating the correlation.
Thirdly, I cannot find much evidence that point (4) (ice thickness) has changed much over the past 30 years, and this was surprising to me at first, until I realized that for any particular melting season, the ice that will melt out between May (or June) until September will be mostly First Year Ice. Whatever amount of MYI there is in any particular year will be mostly unaffected, and thus does not significantly affect the outcome of the prediction.
Finally, sorry that I don't have any data to present right now.
I don't have my programs and data at the computer I'm at right now, but tomorrow or Tuesday I'll present some result from my analysis.
For now, thank you for your post. It made me realize that, although correlation does not equal causation, it may be time to explain in more details that statistical analyses (of area, extent and snow cover) that we are often doing on this blog is actually grounded in the laws of physics, and that our observational data confirms theory. As a result, our predictions (if properly calculated) may not be as 'problematic' as you suggest.
More tomorrow.
Posted by: Rob Dekker | July 01, 2013 at 10:46
Thanks, Bill--nice post.
I agree, though, that "In other words, it has become more difficult to predict end-of-melt-season outcomes..." is an unwarranted conclusion. Since the reference period is much longer, variability is much lower; this by itself should improve the correlations.
Your conclusion could still be right, though; you could, for instance, analyze comparable periods from early in the record, say 1979-1984. How 'easy' was prediction then, as measured by the coefficients?
The thought about looking at detrended data is also a good one.
Posted by: Kevin McKinney | July 01, 2013 at 11:57
In a sense, the less dominant weather conditions become, the easier it should be to predict the minimum, right? Especially when the Arctic has become ice-free. ;-)
Posted by: Neven | July 01, 2013 at 13:00
Well done, just shows the sensitivity of this thin ice to changes in weather and climate. Compare that to thick glaciers that respond over long periods. Juneau Icefield a good example.
Posted by: Glacierchange.wordpress.com | July 01, 2013 at 13:56
Those numbers are great info. It seems SIA is more important to look at than extent through July.
Posted by: Henry1 | July 01, 2013 at 14:32
Easier, yes, but much less interesting!
Maybe you could look at this idea of the weather becoming less important. Rather than just doing one early comparison, follow the lead of a paper like Santer et al 2011 by computing and comparing coefficients for all possible bins of defined lengths.
For example, 5-year bins would look like this:
1979-1984;
1980-1985;
...and so on to:
2007-2012.
Would you find a clear pattern of change in the resulting coefficients? If so, then that would suggest that role of weather--the main ingredient in short-term variability, one would presume--might be changing.
(It wouldn't prove it, of course; you'd still need to look at whether the weather itself was changing in ways that bias the analysis. And we already know it is changing--it's getting warmer and warmer up there, on average!)
Posted by: Kevin McKinney | July 01, 2013 at 19:58
The SIA stats in here suggest that a 2012 min is almost impossible, but we shouldn't rule anything out. However, it makes me question why so many are confident of a new record. This isn't extent which is unreliable. This is area which by the numbers is strong. Why is this the year where we get area to go off the grid from June? Even last year didn't.
Posted by: Henry1 | July 02, 2013 at 09:10
As promised, a bit more on getting physics of the albedo effect involved in correlations.
In Bill's excellent overview of correlations here, he used 'area' earlier as a predictor for 'area' later (area->area), and similarly 'extent' earlier as a predictor of 'extent' later (extent->extent).
Here, Bill noted that correlation of area->area is better than extent->extent and also that when taken over the shorter timeframes (2003->2012, 2005->2012 and 2007->2012) that the correlation factor reduces significantly (and in the case of 'extent->extent' almost completely disappears.
Others have pointed out that the correlation over the longer time frames is better because it simply qualifies the long term down trend that we all know is happening. So overall, it looks like predictions based on statistics are indeed problematic.
I think the reason that the correlations fall apart is because no physical meaning has been attached to using simply 'area' or 'extent' as a predictor for later 'area' or 'extent' of sea ice.
What I tried is to come up few variables known in June, that reflect how much energy the Arctic absorbs in June, and see if a combination (formula) of these variables works better as a predictor for 'later' 'area' or 'extent'.
I choose my variables so that they reflect the albedo ('whiteness') of the Arctic, since that determines how much solar energy the Arctic will absorb for the remainder of the season. I keep it real simple, and thus choose 3 variable for albedo :
(1) Snow - Northern Hemisphere snow cover
(2) Extent - Arctic sea ice area
(3) (Extent - Area) : Effect of the area of polynia and melting ponds throughout the ice
The 'predictor' formula (how 'white' the Arctic is) can then be :
Snow + Extent - (Extent - Area)
Now, for each of these factors, we need to determine how much of the solar radiation will cause ice melt. As a start, which has physical measing, I choose the following weight factors :
For (3) : 1.0 (assuming that ALL solar radiation onto melting ice and into polynia will cause ice to melt later in the season.
For (2) : 0.5 (assuming that half of the heat absorbed in the ocean OUTSIDE of the main pack will cause ice melt (while the other half would cause the ocean to warm up.
For (1) : 0.25 (assuming that half the heat from lack of snow cover will be blown North, and half of that will go to ice melt.
With that rough theoretical 'guess', we then get to the formula that should serve as a 'predictor' for the amount of heat that is absorbed in June due to snow/ice cover :
0.25 * Snow + 0.5 * Extent - 1.0 * (Extent - Area)
Note that this formula, expressed in simple factors is :
0.25 * Snow - 0.50 * Extent + 1.0 * Area
When I use this formula as the 'predictor' I get improved correlation numbers (especially for the shorter terms if you use only 'Area' or 'Extent' as a predictor), which suggests we are on the right track !
But what I was really interested in, is if by tweeking these weight factors (which after all were just educated guesses), if we can improve the correlation numbers even more. If it turns out that the 'optimum' correlation is way off from the weight factors I suggested above, then we know that the physical effects of 'albedo' amplification are simply not significantly visible in the later ice cover numbers.
So I used the 1995-2012 series (long enough for statistical quantity and short enough to not be affected by completely different melting regions) and tweeked the numbers until I found optimal correlation. This was the resulting formula (normalized Area weight to 1.0) :
Formula1 = 0.26 * Snow - 0.59 * Extent + 1.0 * Area
This optimal weight factor choice based on observational data is almost scarily close to the very rough numbers that were based on pretty rough theoretical estimates, and we can talk more the meaning of that that later.
For now, let me show you the initial results I get for the June->September prediction.
The first two lines are the results I got by reproducing Bill's June->Sept correlations and the last two lines are the correlations I get by applying my formula to predict Sept extent/area from June extent/area/snow-cover data :
Now, what is interesting is that the correlation holds up, even for the shorter timeframes.
This means that Formula1 is a better predictor that simply only 'area' or 'extent' for determining the September sea ice extent from information available in June.
And the interesting thing is that this formula is based on physics of ice melt, rather than on statistical extrapolation of area and extent.
Another interesting thing is the prediction of the 2013 September ave min based on this formula. NSIDC has just published the June extent and area numbers, and in absense of Rutgers' snow cover numbers I plug in 200 k km^2 extra snow cover for June, then this formula predicts 4.5 M km^2 ice extent this September, with about 250 k km^2 standard deviation.
Sorry for the long post, but it may be that predictions are NOT as problematic as we thought, and are NOT simply guesses, as long as we use some physically relevant data.
Much more on this later, as there is a lot more interesting info to deduce now that we know that physical effects ARE recognizable in the observational data.
Posted by: Rob Dekker | July 02, 2013 at 09:22
Great post Rob. I think we can go pretty low if the weather sets up right. But my money is still higher than your predicted numbers, but it will be fun to watch. I am not on the new record train this year given the data to this point.
I like to see detailed reasoning like that. Thanks for the post.
Posted by: Henry1 | July 02, 2013 at 10:03
Thanks Henry. Of course weather for the remainder of the season still plays a role, but based on the simple model I presented, there is virtually no chance of reaching a record below 2012 this year.
The only thing I am specifically concerned about these holes (caused by the PAC 2013) in the Central Basin, which fall in category (3) of my formula. Heat absorbed there has the highest impact, and since they are in the Arctic Basin, they may cause quite a bit of havoc (significant ice loss in areas where we normally don't see it) later in the season.
Again, it is a very interesting melting season !
Posted by: Rob Dekker | July 02, 2013 at 10:39
Note that the resolution in area measurements gives possibility to a situation where numerous small leads amount to some albedo decrease. I can't come up with a proper measure for these. But great to see some advance in forecasting. My attempt failed badly in not taking the snow amount in consideration. Thanks Rob Dekker.
Posted by: Erimaassa.blogspot.com | July 02, 2013 at 10:55
Rob, your comment was so good that I elevated it to a follow-up guest blog: Problematic predictions 2.
Truly great stuff.
Posted by: Neven | July 02, 2013 at 11:08
the attempt (2011) (IJIS extent) http://erimaassa.blogspot.fi/2011/07/number.html (I've lost the file in the recent crash of the computer)
true 4663594 vs. predicted 4015554
Posted by: Erimaassa.blogspot.com | July 02, 2013 at 11:15