Buy

Books
Click images for more details

Twitter
Support

 

Recent comments
Recent posts
Currently discussing
Links

A few sites I've stumbled across recently....

Powered by Squarespace
« Mistaken identity | Main | The hearings - cheerleaders »
Monday
Mar012010

What Jones said about station data

Here's the bit from Jones statement that was bothering me:

Stringer: Well I will plug on because I've got one of the quotes from your emails which says "why should I make the data available to you when your aim is to find something wrong with it."...to Hughes. Now that's your email. Now that's the nature of science isn't it, that scientists make their reputations by proving or disproving what other scientists have done previously. Your statement there appears to be anti-scientific and the books that people have written around this issue have persuaded me that you have not provided all the information - the programs, the weather stations, - all the information available so that people can replicate your work, and to say the data is freely available in the United States doesn't enable anyone to go through your workings and agree with you or disagree with you.

Jones: Well the list of stations, we did make that available in 2008, so that has been on our website...

Stringer: How long had people been asking for it at that time?

Jones: Erm

Stringer: You're talking about some papers from 1990 aren't you, that have been kept secret?

Jones: No. There was a paper in1990 and we were asked for the data in that paper, which I was talking about in the previous question, that was made available straight away. The list of stations was made available after about six months, from the first FoI request in about early 2007.

Now I would be grateful if someone would check the transcript for me, but Jones' legendary email in which he rejected Hughes request for data was in early 2005, so I'm struggling to see how the list of stations was released "straight away" in early 2007. Am I missing something here?

 

 

 

PrintView Printer Friendly Version

Reader Comments (80)

OT, but Acton's statement to the press after the hearing is available at http://www.uea.ac.uk/mac/comm/media/press/CRUstatements/vcstatement2

Mar 1, 2010 at 8:49 PM | Unregistered CommenterPhillip Bratby

Regarding the 60ish FOI requests, I thought there was a leaked email explaining that Jones had persuaded the University to ignore the requests. Or am i just dreaming it.

Mar 1, 2010 at 9:05 PM | Unregistered Commentermartyn

"I would like to end by saying what enormous pride I have in UEA and particularly CRU and its staff who have already played an historic role in advancing humanity's ability to understand and monitor climate change."
- Vice Chancellor Acton

Yes, indeedy!

Mar 1, 2010 at 9:13 PM | Unregistered Commenterpettyfog

The BBC has a short video clip which seems to sum up the situation;;
http://www.meto.gov.uk/corporate/pressoffice/2010/pr20100301.html
'Most scientists don't want to deal with the raw station data, they would rather deal with a derived product'
When it was just emails, much of this could always be passed off as careless talk. Now it is public, and on video, it seems much harder to explain away. Awaiting the defensive posts from the PR team...

Mar 1, 2010 at 9:14 PM | Unregistered CommenterSean Houlihane

As to the delay from 2005, and the emails concerning the stonewalling, that is how I recall it, as well. But I'm just a common sense sceptic.
I'm sure that there's several journalists out there in the print media eager to refute those claims.

Wouldnt it be special if old Moonbat was among the first to totally discredit the Alarmists?!!!

Mar 1, 2010 at 9:19 PM | Unregistered Commenterpettyfog

Read carefully - "... from the first FOI request". Accurate. And misleading - because he got polite email requests, which he ignored or denied. When forced to by FOI requests that couldn't be denied, he supplied the requested information in a "timely" manner.
So now we know - don't bother asking politely, just hit them straight up with an FOI request. Given the history, who'd blame you?

Mar 1, 2010 at 9:19 PM | Unregistered CommenterNeil Fisher

My view is that things got mixed up at the point you quote. The dialogue did not really make sense.

Separately, there is a write up about the hearing at
http://www.guardian.co.uk/environment/cif-green/2010/mar/01/phil-jones-commons-emails-inquiry
It reports the truth. I think it says a lot about The Guardian's integrity that they do that even though they are strong global-warming advocates.

Mar 1, 2010 at 9:19 PM | Unregistered CommenterDouglas J. Keenan
Mar 1, 2010 at 9:26 PM | Unregistered CommenterSean Houlihane

Martyn,

Email 1228330629 from the escaped lump of files

Phil Jones writes:

Ben,
When the FOI requests began here, the FOI person said we had to abide by the requests. It took a couple of half hour sessions - one at a screen, to convince them otherwise showing them what CA was all about. Once they became aware of the types of people we were dealing with, everyone at UEA (in the registry and in the Environmental Sciences school - the head of school and a few others) became very supportive. I've got to know the FOI person quite well and the Chief Librarian - who deals with appeals. The VC is also aware of what is going on - at least for one of the requests, but probably doesn't know the number we're dealing with. We are in double figures.

There are other comments Phil writes which also show him in a less than stellar light.

Mar 1, 2010 at 9:51 PM | Unregistered CommenterGareth

I need somebody to cheer me up. This seemed very much like a sad day for all of us--skeptic, alarmist, lukewarmer. Oh, the questions that were not asked...

Mar 1, 2010 at 10:05 PM | Unregistered CommenterTom Fuller

Doug, agreed, the best write-up that I've seen so far. But that's because it's Fred Pearce and he's got it in for the AGW crowd right now. Why is that, remind me? Did someone mention he'd had a run-in with Gavin?

Mar 1, 2010 at 10:25 PM | Unregistered CommenterRichard Drake

The section quoted is pretty much right, except for some ums and ahs, do you know it is still watchable? the recording is still available, and if you go to about 1:16 into the debate you will see Stringers questioning start there

http://www.parliamentlive.tv/Main/Player.aspx?meetingId=5979

Mar 1, 2010 at 10:32 PM | Unregistered CommenterSteve2

Tom, it's much less than it should have been but I think the main point about reproducibility did hit home, thanks mostly to Mr Stringer. We've just gonna have to work from there.

Mar 1, 2010 at 10:32 PM | Unregistered CommenterRichard Drake

Just a note to thank His Grace for covering this event as thoroughly and as helpfully as he did . . . A credit to the diocese, so he is.

Mar 1, 2010 at 10:35 PM | Unregistered CommenterDave B

Ah yes thanks Gareth that’s the mail I was trying to recal, but I did totally forget the last sentence. “We are in double figures” Jones must have written it before he was inundated!

Mar 1, 2010 at 10:47 PM | Unregistered Commentermartyn

@Tom Fuller

I found it very illuminating. I have never bothered to follow one of these Gov media things before, and I wonder how many other UK people have discovered this resource based on the experisnce of engaging blogs like the Bishops. I actually had a lower expectation of what would happen based on the remit and the persons to be questioned. I agree with some of the critcs from both sides who criticise, for example, why have Lawson and Peiser representing scepticism? But they did well enough and I think there will be some understanding that this was a bit of a show, Acton/Stringers exchange was funny being sarky about the ICO, and I liked the line about the "lonesome pine" :) Ironically I think that the future of politics may turn out to based on this techno precedent of the peoples engagement with the "worlds biggest problem" TM

Mar 1, 2010 at 10:51 PM | Unregistered CommenterSteve2

Rejoice in Climategate! Could we have imagined these clowns being questioned by a parliamentary inquiry live on our computer screens a few months ago. I think not.

Mar 1, 2010 at 11:04 PM | Unregistered CommenterEddie O

I think the main point about reproducibility did hit home, thanks mostly to Mr Stringer.

And let's ignore that the results have been independently reproduced 4 times in the last two weeks - even Roy Spencer managed it for the NH - on top of at least two other other confirming results that existed before then.

Let's hope that nobody notices that the data was in fact available to do that all along.

Mar 1, 2010 at 11:39 PM | Unregistered CommenterFrank O'Dwyer

A bit early in the year for Cherry Picking Frank?

Mar 1, 2010 at 11:50 PM | Unregistered CommenterOrchard

Orchard,

Gesundheit. It takes more than parroting 'cherry picking' to constitute an argument.

Mar 1, 2010 at 11:55 PM | Unregistered CommenterFrank O'Dwyer

Getting the same result via different data, different adjustments and different computations does not verify what CRU has been doing.

Mar 1, 2010 at 11:56 PM | Unregistered CommenterGareth

Gareth,

Getting the same result via different data, different adjustments and different computations does not verify what CRU has been doing.

Yeah 6+ independent replications from 95% the same base data set using similar adjustments for the same purpose is such a red flag for nefarious goings-on, isn't it?

Obviously as 'auditors' and blog scientists this indicates that you must dig deeper in order to 'replicate'. You will require the source code for excel, windows, the fortran compiler, the microcode for the CPU, and detailed printed circuit board layouts so that you can construct an exact replica of the computer used for the calculations.

In fact as blog scientists you should be satisfied with nothing less than the actual computer used for the calculations.

Mar 2, 2010 at 12:17 AM | Unregistered CommenterFrank O'Dwyer

Frank

There's an interesting thread over at Lucia's on releasing the code. Ron Broberg gives what I think it a useful typology.

Some thought I had on scientific confirmation related somewhat to steven mosher (Comment#35171)
-
1) Given the exact data and exact code, an audit can be made on implementation errors.
-
2) Given the exact data but using independent methods, results of the first team can be bolstered/undermined by the second team.
-
3) Using independent data and independent methods, results of the first team can be bolstered/undermined by the second team.
-
Historically, 3) was the method used most often to advance or correct a previously published work. Satellite data -v- surface records is an example of 3).
-
McIntyre has gotten very good at 2) – that is in taking the original proxy data and using it to reconstruct the proxy results produced by other teams. NASA GISTEMP -v- Hadley CRU is another example of 2) where both take GHCN as input but apply their own methods to produce output. E.M. Smith is struggling with 1). Stokes reimplementation is essentially a version of 1). Whereas the work to reconstruct CRUTEMP from descriptions in published papers is an example of 2).
-
In my limited understanding, in some science fields (sociology, I’ve read), providing complete data and exact code is common. In most, it has not. Traditionally, rather than an exact copy of the code, a description of the methods used along with a few key equations have been provided. That way, a second team could re-implement the method independently, if they so chose. Times are changing.

Now I agree that people can perfectly well go away and do 2) but even if that's so, why should they not also have the option of doing 1)? Are you arguing in principle against people having the option of doing 1) and, thus, against releasing the code? If so, I'd be interested to hear your thoughts.

Mar 2, 2010 at 12:39 AM | Unregistered CommenterRichieRich

@FoD

Cherry Picking

reproducibility

I believe in some cultures, individuals independently interbreed with their siblings.

Cherries yup, (mostly windfall), but certainly not worth picking.

Mar 2, 2010 at 12:57 AM | Unregistered CommenterAnoneumouse

Frank, when I talked about reproducibility I was referring to the lamentable record of Climate Science in that area, including the Hockey Stick and all that fed into it, since 1990. I certainly enthusiastically agree that the situation's getting better - as I implied in my second message to Glyn Moody the week Climategate broke (before I knew that it had). And it's going to get better again. In fact, I'm sure that you and I will be dancing through the daisies together before we're done.

Mar 2, 2010 at 1:06 AM | Unregistered CommenterRichard Drake

RichieRich,

I don't think there's anything wrong in principle with releasing the code but the data is far more important. A description of the high level methods and algorithms should be sufficient for the code - it is normal practice to use a high level vocabulary when communicating concepts and algorithms and for example if somebody says to calculate an average or standard deviation there is no need to demand their code in order to understand what that means. I have implemented crypto algorithms from published papers before. In that field for example, a common piece of vocabulary is modular exponentation - you just go implement that, you don't demand the author provides you with the code. (In fact when I did that, their code would have been no good to me as my task was to write a more efficient implementation of it)

The code should only be of interest if the algorithms aren't clear or there is some dispute about them. That happens sometimes - there is some case like that over on realclimate. But before you can say that there should at least be a good faith attempt to replicate and some kind of wild discrepancy that requires explanation, and a failure to reach understanding based on simply describing the methods.

I'd also argue that it's not entirely obvious what 'releasing the code' means and there is no end to that kind of nitpicking. To take an example, let's say you have crunched some data using excel. If you give me the .xls file, have you 'released the code'? How do I know I'm using the same version of excel you are? Indeed I don't even have MS excel - I'd run that using openoffice or iwork for the mac. Independent implementations of the same thing. Should be good enough, right, until there is a problem? Do I need to demand the source code for excel? If not then why not - maybe there is a lesson there?

I would also say that if everyone used the same code all the time that would be a very bad idea. That would make the code a common point of failure as every 'replication' would have the same failure modes.

Examples of the sort represented by (2) and (3) above are far more powerful and convincing replications, because they are not mere repetition.

Mar 2, 2010 at 1:11 AM | Unregistered CommenterFrank O'Dwyer

Frank, those of us who have worked in software for over thirty years don't believe the crap about "description of the high level methods and algorithms". We want the source code first and the suite of unit tests that define its operational specification second. We don't believe anything else we read until we have both of those and the ability to run the second against the first.

In the case of climate science I haven't heard test suites mentioned at all but every single open source project has one as a matter of course and very soon, we believe, all climate science modeling, reconstructing, adjusting and what-have-you will be open source so we can expect major changes.

I'm speaking for a few other people there who've been in software a while but I think it's along the right lines. In fact you're touching on one the areas I'm most passionate about.

Mar 2, 2010 at 1:22 AM | Unregistered CommenterRichard Drake

Okay, then--stiff upper moustache and onwards and outdowngorightwards.

Mar 2, 2010 at 1:30 AM | Unregistered CommenterTom Fuller

I have had a recurring thought while watching Phil Jones' testimony. Is he committing seppuku in the hope that his act will somehow divert attention from the fiasco in which he played his part?

Mar 2, 2010 at 2:22 AM | Unregistered CommenterRayG

O'Dwyer writes:
"And let's ignore that the results have been independently reproduced 4 times in the last two weeks - even Roy Spencer managed it for the NH - on top of at least two other other confirming results that existed before then.Let's hope that nobody notices that the data was in fact available to do that all along."

The moral commitment in scientific method to repeatable results means exactly that. It does not mean doing parallel research about the same phenomena. So, same data and same results do not make repeatable research. The research should be published in such a way that one can literally repeat the work. That means "show all of your work."

Mar 2, 2010 at 2:30 AM | Unregistered CommenterTheo Goodwin

I have a slightly different take on it.

Let us assume Jones is a scientist first and foremost. We want such people to stay within their scientific fields, do a good job and not venture into political waters.

Well I cannot imagine too many real scientists wanting to be in Jones place yesterday, shaking hands and all. Irrespective of the outcome.

Those 'scientists' whose egos drive them onto the political/media stage will always do so (post-normal scientists?) Those scientists who are primarily scientists will think twice. So the event served a purpose, because I am sure that academics across the world and country took an interest..

Mar 2, 2010 at 3:24 AM | Unregistered CommenterJiminy Cricket

I think the most telling statement Jones made was when he said that he had never been asked by other scientist to provide raw data and code. This, if taken at face value, would mean that his studies have never been replicated by other scientists. If the studies have never been replicated, then can they truly been said to have been peer reviewed?

Mar 2, 2010 at 3:44 AM | Unregistered CommenterJmarz

Frank O'Dwyer And let's ignore that the results have been independently reproduced 4 times in the last two weeks - even Roy Spencer managed it for the NH - on top of at least two other other confirming results that existed before then.

What the hell are you talking about? What 4 times are you talking about? Roy Spencer did not manage to reproduce it.

The results (meaning the warming trends) have to be shown to be legitimate. The only way to do it is to get ALL the RAW DATA used. Find out which raw data WAS NOT USED and why not. Find out HOW the adjustments were carried out and WHY.

Correct me if I'm wrong, but I understand that much, if not all, of the RAW DATA HAS BEEN LOST AND GONE FOREVER (Dreadful sorry Phil Jones).

Reproducibility does not mean here is the adjusted data I used, this is the program I used to adjust it - do it and see you get my results.

Dr Spencer used the International Surface Hourly weather data of NOAA's, did not make ANY adjustments for the UHI effect, which Jones claims he has accounted for, and got a trend from 1973-2009, 20% less than Jones's. Which is odd because not adjusting for UHI SHOULD MAKE THE TRENDS GREATER NOT LESS.

Moreover Dr Spencer used only those stations operating over the entire period of record. In other words stations that did not require any more "adjustments" (read fudging) for the period.

Mar 2, 2010 at 3:44 AM | Unregistered CommenterRichard

Hi Frank

Thanks for your thoughts. Here's some more stuff from Lucia's blog. What follows is a rather long post, but I think (hope!) of interest.

Mosher writes

A. you need the code to verify that the figures provided are actually results of the current version of the code. You and I both know that there are times when you rerun your code and get
different results due to improper version control. basic QA for
any publication I ever did.
B: you need the code to prove that there are no compiler or platform issues.
C: you need the code if you want to rehost.
D: I read hansen’s papers. they did not describe all the steps.
in particular they did not describe the QA steps taken prior
to applying the algorithms. I requested descriptions of these
undocumented methods. Gavin responded “read the paper”
But the paper was entirely vague. When the code was released
the readme indicated that the QA check was not done via
and algorithm but was human judgement. So, the paper didnt describe the step. Gavin said the paper did. it didnt. code got
released and my question got answered.

A little later in the thread, carrot eater writes

As for not releasing code: in most any field of science, people in other lab groups don’t sit around waiting, just in case the one lab group releases some code and allows a shortcut. This just simply doesn’t happen. Each research group is moving forward on their own projects, in their way.

And it’s this perspective that Mosher just simply lacks. The two competing lab groups don’t have to produce anything that matches within 100.00000%. But when their results do diverge, then people take a closer look to see what’s going on. Maybe a third group jumps in the fray. People work until the reason for the divergence is understood. And then you move on. Maybe it gets to the point where comparing code has to happen, but I don’t think that’s very common.

Mosher replies

I don’t lack that perspective. I have a different perspective. I’m well aware that different “labs” have different approaches.
My approach is the engineering approach using validated models.

I’m designing an aircraft. The science behind radar cross section is pretty clear. Estimating radar cross section is not that easy. There are different approaches. Various R&D organizations each have their go at the problem. But the customer, the air force, needs to decide. One model gets validated. everybody gets that model. You want to improve it? submit your changes. everybody gets the changes.

From my perspective I am the customer of climate science. You want to use the results of science to inform a policy, then I am the customer of that science. I got customer requirements. they may differ from yours. they are my requirements. I will buy your science when you adopt practices of transparency. full stop.
I want a red car. you try to sell me a green car and tell me that color doesnt matter. go away. you try to argue that I dont need to be able to pick the color of my car. go away. it doesnt cost anything to share code. the sky doesnt fall. refusing to share it may have cost CRU an important scientist.

And on point. Jones didnt share code with mcintyre because he thought that Mcintyre should do the code himself.

He specifically refused to share the code because:

1. he didnt think he could find it
2. he thought it was an undocumented mess.
3. he knew the code would SHOW Mcintyre why mcintyre
could not replicate the results from the description in the PAPER.

you see #3 is really insideous. Mc could not replicate the results.
he tried. So he asked for code. Jones says NO. and in a mail to Mann, he said ” I wont give him code, and I know why he cant replicate it”
So, you live in this world where you think people will just share the code when I say I cannot replicate the results. you trust them to be as reasonable as you are. Witness also the problems that GAvin had replicationg scarfettis work. they read the paper, wrote their code. didnt match. they requested code. the answer was NO.

In response to Mosher's earlier point about the Hansen paper, Carrot Eater writes

Mosher: If a minor detail isn’t spelled out in the paper, then just make some decision for your own implementation, play with it to see how much it matters, and move on. This is how science works in the real world, where competing researchers don’t FOI each other. When you then write your own paper, you can put in as much detail as you want in the supplementary info, or the appendix to the thesis, or wherever.

Mosher replies

You don’t get it. I’m not doing science. I am trying to decide whether to BUY THE SCIENCE or not. So when you tell me that
I have to do my own science if I dont buy the science you are selling, it’s kinda funny.

I’ve said this before. Perhaps it bears repeating. If you want
me to accept a result or an argument, i want to see your work.
I may go through that work, I may not. But, I want to see your work. If you don’t want to show me your work, don’t demand
that I buy it. Dont expect me to buy it because of letters behind
your name. dont expect me to buy it because I havented proved the opposite. Don’t expect me to defend my requirements.
Now, if you dont want to make this customer happy, that is
your choice. Lucia doesnt share code. Do I buy her analysis?
sure. Why? because NOTHING happens to me if I buy or dont buy.
Lets say GISS didnt share code. Would I buy their analysis?
No. why? because it matters more to me.

Your perspective and that of Carrot Eater seems to be that one should (is morally obligated?) to try to reproduce the results in a paper from a high level description and only if one has problems doing so, should one ask to see code. And only at this point is the author obliged to release code.

Mosher's postion, and one I think I have sympathy with, is: if you want me to believe your work, show me your detailed workings (which includes code). I'm not obliged to try to reproduce your work but if you want me to take your work seriously, you are obliged to put your workings out there.

If not identical, this is very close to Theo's position above, when he writes

The moral commitment in scientific method to repeatable results means exactly that. It does not mean doing parallel research about the same phenomena. So, same data and same results do not make repeatable research. The research should be published in such a way that one can literally repeat the work. That means "show all of your work."

Mar 2, 2010 at 7:41 AM | Unregistered CommenterRichieRich

Frank O'Dwyer

The issue is why you do not want to see data and code released. What is the argument against doing it?

After all, the future of the planet is at stake, not to mention think of the children, trillions, perhaps 100s of trillions, are being requested to be spent on very controversial projects whose outcomes are forecast based on various studies. Before they stump up the money, and before they engage in projects which may have negative side effects, people are invoking the precautionary principle and saying they want to be sure. If even a little doubt is removed by releasing the data and the code, it should be done.

Your argument in reply to this is on a purely personal level. Its to do with whether the people making the requests are nice, reasonable, could do the work over....etc Why it will be a lot of work. Why the researchers may find it stressful and time consuming.

Who cares?

There is not and cannot be any argument why, if people really are advocating the spending of trillions and the modification of the climate of our planet, they do not have to supply EVERYTHING. They have another option if you like, do not engage in political action. Simply assert that the science is as they say, that their studies are legitimate, but accept that it is not a basis for investment. Most of us would have little problem with that.

Our problem is some guys saying 'trust me' to the tune of trillions and to the tune of dramatic interventions in the functioning of the climate of the planet. When you make arguments like this, you have to expect intense and detailed scrutiny.

Its like, I am building a 16 foot sailing dinghy for my own use. Go to it. I am going into business and will be selling the things by the dozen? Comply with health and safety. Show your stability calculations.

Why do you think there are going to be, or should be, exceptions for CRU and for proposals to forcibly modify the climate of the planet?

Mar 2, 2010 at 8:14 AM | Unregistered Commentermichel

@RichieRich, good points.

If you said to a Mathematician that you could confirm the Proof a theorem by just printing the final page of Proof but without the 'solution', they would laugh in your face.

The whole of science is based on proven building blocks. Even a simple building arithmetic block like 1+1 is always 2 has had to proved before it too can be used in further proofs.

In engineering a black box technique where you do not care what is in the box (usually supplied by a supplier) as long as you get the EXPECTED results from it can work, but ONLY if you know EXACTLY what is expected beforehand. That is, YOU CAN TEST THE RESULTS against a baseline.

What is it with Climate 'Science' that even these basic principles do not apply?

Mar 2, 2010 at 8:17 AM | Unregistered CommenterJiminy Cricket

@michel, also good points.

Science has historically shown itself to be amoral. If it was me in that position I would be going to nth degree to ensure that my results were sound. Because I believe it would be a moral responsibility to do. As an engineer it is my duty to find fault.

So I do not buy any of the 'personal' defence based arguments. This is not about personalities. Science does not care. Science is impersonal.

When I see a Scientist I do not think 'good guy' or 'bad guy'. it is irrelevant. History has always shown that scientists will follow paths not always dictated by morals.

This rigourous approach is missing from Climate 'Science', that is why the vacuum is being filled by the Sceptics.

Mar 2, 2010 at 8:34 AM | Unregistered CommenterJiminy Cricket

Correct me if I'm wrong, but didn't this all start with an attempt to reproduce the end result, and not getting anywhere because the 'obvious' guesses as to how each step could proceed not being correct? Where a method was mentioned in the literature, it was a trivial phrase which could be interpreted as being accurate, but was not sufficient to allow precise replication. The reason for the obfuscation - see Mann's response to MM03. The obfuscation provides a number of 'elementary errors' in the critical analysis which are used to dismiss the whole effort. At that stage, all M&M could say was 'there is something odd going on'. How are we all expected to make a judgment on the oddity if all the tricks are hidden?

Mar 2, 2010 at 9:03 AM | Unregistered CommenterSean Houlihane

Precisely, the argument that it is sufficient to provide the high level stuff and a statement of methods was blown out of the water by the study of the original MBH paper. It proved not to be possible to reconstruct the result from the information given in the paper, because they had used a strange, undocumented (and flawed) method of PC analysis.

I agree with Mosher, if the published data is to be used to drive policy or for major decision making it is essential that the code is made available. We all know that mistakes are endemic in computer programmes which is why those of us who do real science (I work in the pharmaceutical industry) have stringent Comval processes for ensuring appropriate documentation, version control and change validation procedures in place.

One of the most astonishing admissions at yesterdays hearing came from Dr Slingo of the Met Office who, with a straight face, said that their forecasting code was thousands of lines long but was validated twice a day by their forecast output. I wouldn't take any drug or fly in any aeroplane designed by her and her colleagues.

Mar 2, 2010 at 9:30 AM | Unregistered CommenterArthur Dent

Frank O'Dwyer said:

I would also say that if everyone used the same code all the time that would be a very bad idea. That would make the code a common point of failure as every 'replication' would have the same failure modes.

Strawman. People are not wanting to use the code they are wanting to check it.

Mar 2, 2010 at 9:58 AM | Unregistered CommenterGareth

Maybe the analogy isn't quite perfect but, consider the reaction of a athlete upon finding that (s)he's failed a drug test but then being told by the doctor that the original sample had been lost.

Would you think it fair that the only possible recourse for defence should be to argue against the test result and details of the test method? Should the statement "trust me, I'm a doctor" be sufficient to justify a conviction?

Mar 2, 2010 at 10:51 AM | Unregistered CommenterDave Salt

Frank's position would be, why should she see the original sample? Let her go do her own tests and replicate the results if she can. The same would be his argument about drug testing. Why should they reveal the details of their sample? Go collect your own sample and do your own statistics, if you think you can do better.

Right. And until you do release, you don't get to sell the drugs. You don't have to release data for audit, we don't have to approve your drugs.

Same with climate. You don't have to release, we don't have to spend. One or the other. Same by the way with financials. You want to be a listed company, produce accounts. You don't have to produce accounts. In that case, you can't sell stock.

Mar 2, 2010 at 10:57 AM | Unregistered Commentermichel

Prof Jones's response to the athlete would be:
You're guilty. We're not going to tell you how we worked that out, and we've lost your original sample. But you're still guilty. Honest.

Mar 2, 2010 at 12:57 PM | Unregistered CommenterStuck-Record

But it's far more important than this in the context of climate "science".

If I do my own stuff and come up with a different answer, the team, the media and the IPCC ignore it entirely. The sceptics, even those of us funded by the billions put into the sceptic side by big oil (natch), simply could not get any air time at all. There were and are sceptics who have been able to publish papers that go against the consensus but they are simply ignored. (except to be used now by the team, similarly disgracefully, to deny that the peer review process has been goosed - its only the people trying directly to undermine the team that have been pushed out)

The ONLY way to strike at the team and the consensus is to demonstrate that the consensus scientists have arrived at their results incorrectly. You have to demonstrate flaws in THEIR work.

The utterly disingenuous dismissal of the M&M work by the team shows that absolute and complete reproducability must be the start point. Sceptics must demonstrate complete mastery of the team results because it is only after that that they can show why it is flawed. Without the first step, the demonstration of flaws has no traction.

Frank is being completely disingenuous to sideline this.

Mar 2, 2010 at 1:00 PM | Unregistered CommenterThe Pedant-General

Dave Salt: very fair analogy.

Arthur Dent: What's a Comval process? That's new to me. I fully agree that 'mistakes are endemic in computer programmes' and to deal with that version control is one of the important tools in the kitbag. Others not so well known in a software team are pair programming (or co-coding) and continuous integration. Most vital of all, well ahead of documentation for me, is a comprehensive and readable set of regression tests. And all these processes depend on each other. Despite the immaturity of software development, there is accumulated wisdom to draw on by now. (The experts of course disagree to some extent - but the factors I mention have been coming to the fore in the last decade at least.)

Gareth:

Frank O'Dwyer said: I would also say that if everyone used the same code all the time that would be a very bad idea. That would make the code a common point of failure as every 'replication' would have the same failure modes.


Strawman. People are not wanting to use the code they are wanting to check it.


This is quite a deep point on both sides. You're right that the first and most basic issue is replicating and checking results. And Frank's right to point to the common point of failure. But it's quite wrong to use this to argue against open source. Because open source doesn't preclude anyone starting from scratch, to try and do a better job. In fact, it makes it a great deal easier to do so, which could accelerate advances in climate science no end.


An interesting example would be the introduction of the Chrome browser in Sep 2008 after Google had invested so much effort and funding in Mozilla Firefox. They got to the point where they felt that they could do better by starting again. Improved security was a key reason they gave. But of course when Chrome arrived people at once found new security holes, that Firefox didn't have. For over the years Firefox has built up a loyal base of testers and contributors who very quickly fix such things. Chrome didn't start with such a large 'ecosystem' around it - but it looks to have a lot of momentum by now.

The really cool thing is that both are open source and Google is actively encouraging Mozilla or anyone else to come and use the best bits of their efforts. And Mozilla is committed to moving to the WebKit rendering engine used by Chrome but largely developed by Apple for Safari. Thus there's a mixture of healthy competition and a desire not to duplicate resources too much. This all plays out very nicely in the open source world as we currently experience it in such core areas of the Web.

We desperately need this kind of culture in climate science. There's no reason not to. There are many trillions of dollars of reasons for every human being on the planet to want it to be so.

Mar 2, 2010 at 1:26 PM | Unregistered CommenterRichard Drake

To follow on from the Pedant-General, a further exchange from Lucia's. Carrot Eater again.

I strongly feel that somebody could have easily written a very good GISS emulator, well before they released the code. In fact, releasing the code probably slowed things down, as it was in such a messy condition. Writing a new GISS from scratch would have been faster than doing what the good ccc guys are doing, though it would not have been 100.000% perfectly matching. GISS really doesn’t do anything that sophisticated or complicated. I’m a bit annoyed actually that there wasn’t a high profile attempt to do so; it would have proven the point very nicely. But, better late than never.

Mosher again.

Yes, carrot we heard this argument a dozen times in 2007. People arguing that we should work their way. When the code was released people could actually look at it and decide which path THEY wanted to go down. The issue with an emulation is the lack of 100% matching. Especially when the problem we wanted to look at ( UHI) is likely at the edge of the noise floor. The other thing is that ANYTIME you show something with an emulation that contradicts the real code, you will get this answer:

The emulation is wrong. the emulation is not peer reviewed.
So from a practical standpoint emulations can’t show anything
wrong with a package they emulation.

What you would have is two pieces of code that didnt produce
similar results. Like so:

GISS shows a warming of 1C
Emulation shows a warming of .98C

Everyone agrees that the emulation is “valid”

GISS shows no effect from dropping stations.
Emulation shows effect from dropping stations.

Conclusion?

With no code, the GISS fans will say the emulation is wrong.

So, you need the code. more importantly we have a right to the code regardless of our identity, intentions, or ability to do anything with it. full stop right to the code.

Mar 2, 2010 at 1:54 PM | Unregistered CommenterRichieRich

Richard "ComVal" is the shorthand for "Computer Validation" a requirement under the GLP/GMP (Good Laboratory/Manufacturing Practice Regulations). Any climate scientist would probably have hysterics at the very notion of Good Laboratory Practice one of whose requirements is that any study being undertaken today should be capable of reconstruction (and replication) solely from the archived raw data at any point in the future.

This requires scientists to maintain and archive laboratory notebooks cross referenced and signed together with all the relevant statistical manipulations and computer codes and even the calibration and maintenance records of all the instrumentation used.

"The dog ate my homework" excuse is non viable and could lead to a requirement to run the complete study again or ultimately see your product withdrawn form the marketplace. The methodology of the climate scientists in academia and government institutes make me ashamed to be a scientist.

Mar 2, 2010 at 2:05 PM | Unregistered CommenterArthur Dent

RichieRich and others have hit it on the head. It vastly increases the work to check the results to not include the code (and data!), for no reason other than to make it harder or obfuscate the process. I admire McIntyre for even trying. The paper is the documentation, and any programmer will tell you the documentation neither perfectly matches the code nor is ever complete. Then, when there are differences, the reimplementation can be dismissed for its incorrectness, even though it is correct to spec.

The code is one of the places where mistakes are the most likely to occur. If there are mistakes in the code, and even the most trivial code has bugs, refusing to release the code means you are hiding mistakes and making verification difficult or impossible. Doing this for publicly funded work from which government policies are set world-wide? That is unacceptable.

Phil said no reviewer had ever asked for the code or data. What does this say for the state of peer review in climate science? Reviewers just glance at it and say "'looks OK to me"?

Mar 2, 2010 at 2:30 PM | Unregistered CommenterSpongebob Squarepants

Perhaps it's worth focussing on this point a little:

"The ONLY way to strike at the team and the consensus is to demonstrate that the consensus scientists have arrived at their results incorrectly. You have to demonstrate flaws in THEIR work."

This approach appears quite combative and perhaps not very constructive: it does not advance the science except in as much as it hauls the science back from what might be a wrong turn. It is therefore open to - and indeed has been the subject of - criticism from warmists.

If warmists did not want to have people crawling all over them and trying to destroy their work, they shouldn't have invested so much time and effort in it without getting it really properly validated. It is the warmists who have contrived to prevent such validation and now there is real and really pressing work to unpick the entire canon of literature before anyone can begin to make any progress at all - there is simply no confidence whatsoever in any of the foundations - there is nothing solid to build on.

In short, IT'S THEIR OWN BLOODY FAULT.

All that's happening now is that we are seeing this lamentable state of affairs.

Mar 2, 2010 at 3:06 PM | Unregistered CommenterThe Pedant-General

Did Jones/CRU commit to releasing the data, metadata and code? By when? If not, what is going to be done to compel it to be made public? This is the only issue that really matters.

Mar 2, 2010 at 3:25 PM | Unregistered CommenterPete Ballard

PostPost a New Comment

Enter your information below to add a new comment.

My response is on my own website »
Author Email (optional):
Author URL (optional):
Post:
 
Some HTML allowed: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <code> <em> <i> <strike> <strong>