04-18-2018 10:16:00 AM -0700
04-16-2018 01:32:51 PM -0700
04-16-2018 09:59:36 AM -0700
04-12-2018 09:53:41 AM -0700
04-10-2018 11:19:03 AM -0700
It looks like you've previously blocked notifications. If you'd like to receive them, please update your browser permissions.
Desktop Notifications are  | 
Get instant alerts on your desktop.
Turn on desktop notifications?
Remind me later.

Climategate Computer Codes Are the Real Story

Poor Harry is in the first circle of programmer hell: the program runs fine; the output is wrong.

He presses on:

17. Inserted debug statements into anomdtb.f90, discovered that a sum-of-squared variable is becoming very, very negative! Key output from the debug statements:

some test output...

forrtl: error (75): floating point exception

IOT trap (core dumped)

..so the data value is unbfeasibly large, but why does the sum-of-squares parameter OpTotSq go negative?!!

This is not good -- the existing program produces a serious error when it's run on what is supposed to be the old, working data. Harry presses on, finding a solution to that bug, going through many more issues as he tried to recreate the results of these runs for the data from 1901 to 1995. Finally he gives up. He has spoken to someone about what should be done:

AGREED APPROACH for cloud (5 Oct 06).

For 1901 to 1995 - stay with published data. No clear way to replicate process as undocumented.

For 1996 to 2002:

1. convert sun database to pseudo-cloud using the f77 programs;

2. anomalise wrt 96-00 with anomdtb.f;

3. grid using quick_interp_tdm.pro (which will use 6190 norms);

4. calculate (mean9600 - mean6190) for monthly grids, using the published cru_ts_2.0 cloud data;

5. add to gridded data from step 3.

This should approximate the correction needed.

Catch that? They couldn't recreate the results, so they're going back to their published data for the first 95 years of the 20th century. Only ...

Next problem -- which database to use? The one with the normals included is not appropriate (the conversion progs do not look for that line so obviously are not intended to be used on +norm databases).

They still don't know what to use for the next several years. Harry gives up; it's easier to write new codes.

22. Right, time to stop pussyfooting around the niceties of Tim's labyrinthine software suites - let's have a go at producing CRU TS 3.0! since failing to do that will be the definitive failure of the entire project.

This kind of thing is as fascinating as a soap opera, but I want to know how it comes out. Near the bottom of the file, I find:

I am seriously close to giving up, again. The history of this is so complex that I can't get far enough into it before by head hurts and I have to stop. Each parameter has a tortuous history of manual and semi-automated interventions that I simply cannot just go back to early versions and run the update prog. I could be throwing away all kinds of corrections - to lat/lons, to WMOs (yes!), and more.

The file peters out, no conclusions. I hope they find this poor guy, and he didn't hang himself in his rooms or something, because this file is a summary of three years of trying to get this data working. Unsuccessfully.

I think there's a good reason the CRU didn't want to give their data to people trying to replicate their work.

It's in such a mess that they can't replicate their own results.

This is not, sadly, all that unusual. Simply put, scientists aren't software engineers. They don't keep their code in nice packages and they tend to use whatever language they're comfortable with. Even if they were taught to keep good research notes in the past, it's not unusual for things to get sloppy later. But put this in the context of what else we know from the CRU data dump:

1. They didn't want to release their data or code, and they particularly weren't interested in releasing any intermediate steps that would help someone else

2. They clearly have some history of massaging the data -- hell, practically water-boarding the data -- to get it to fit their other results. Results they can no longer even replicate on their own systems.

3. They had successfully managed to restrict peer review to what we might call the "RealClimate clique" -- the small group of true believers they knew could be trusted to say the right things.

As a result, it looks like they found themselves trapped. They had the big research organizations, the big grants -- and when they found themselves challenged, they discovered they'd built their conclusions on fine beach sand.

But the tide was coming in.