Thursday, March 10, 2011

ISO erred on trueness

When ISO was set up in April 1947 at Paris, France, it was all about nuts and bolts. As a matter of fact, ISO/TC1 Screw heads came first and ISO/TC2 Fasteners was second. Ever since has ISO been setting up a broad range of standards while the world is putting its standards to the test. But I wonder why ISO did err on trueness. Here’s what ISO announced in its Technical Corrigendum 1 on 2005-08-15.

Accuracy (trueness and precision) of
measurement methods and results
Part 5: Alternative methods for the determination of
the precision of a standard measurement method


ISO/TC69, Applications of Statistical Methods, Subcommittee SC 6, Measurement methods and results published the above Technical Corrigendum 1. So what error was SC6 to correct? Of course, trueness and precision should never have been between brackets! What ought to be between brackets are precision and accuracy! A true test for bias would need first of all an unbiased variance estimate. Those who have kriged and smoothed cannot possibly test for bias or estimate precision. Neither can they test for spatial dependence by applying Fisher’s F-test to the variance of bogus data and the first variance term of the bogus data set. So much for kriging and smoothing when we study climate change on our planet!

What I would want between brackets is precision and bias. Derive the variance and then test for bias if enough degrees of freedom are available. Bias detection limits (BDLs) and Probable bias ranges (PBRs) for Type I risks and Type I&II risks are intuitive and powerful measures for the observed bias. Ignorance of precision and bias has irked me as long as have central values without variances. Surely, Matheron and his disciples have brought a big catch of bad science to our world.

I have juxtaposed precision and bias since 1974. That’s when I became a member of CAC/ISO/TC102-Iron ore. I am also a Member of ISO/TC27-Solid mineral fuels, of ISO/TC69-Applications of statistical methods, and of ISO/TC183-Copper, lead, zinc and nickel ores and concentrates. Much of what I have written on sampling and weighing of bulk solids became part of ISO/TC183. My son and I have written a software module on Precision and Bias for Mass Measurement Techniques. ISO has published it as an ISO Standard. I was told Canadian Copyright was not violated. Merks and Merks found it easy to work with precision and bias. What’s more, we are pleased to be encumbered with Fischerian (sic!) statistics.

The International Organization for Standardization was much on my mind when I posted false and true tests for bias. ISO comes from the Greek word isos which means “equal”. Scores of countries have set up national institutions to interface with ISO. The Standards Council of Canada Act received Royal Ascent in 1970. That’s when the CAC prefix was placed before ISO. I have nothing but praise for Standards Council of Canada. CAC/ISO/TC69 Applications of statistical methods has played a key role in my work. I have always juxtaposed Precision and Bias. But it’s a long a story. And it’s bound to get longer while I’m trying to keep it short. I do want to kill two nutty practices with the same stone. The first is to assume spatial dependence between measured values in an ordered set. The second is to not apply Fisher’s F-test to the variance of a set of measured values and the first variance term of the ordered set. Geostatistocrats assume, krige, smooth and select the least biased subset of any infinite set of kriged estimates. It may well have dazzled those who have never scored a passing grade on Statistics 101. I still find it funny how so few could write so much about so little.
Biased but high degree of precision

Here’s where ISO has created the error to be corrected. Are true and false antonyms or not? Wouldn’t the antonym of trueness be falseness, or perhaps falsehood? Of course, I would call this ISO document Trueness (precision and bias) of measurement methods and results. Surely, a significant degree of spatial dependence between measured values in an ordered set does impact precision. But I upped the odds of finding a false positive. I did so by inserting David’s “famous Central Limit Theorem” between each pair of measured values. Pop in more kriged estimates between measured values and bogus spatial dependence may make the odd mind spin. Is it a minor miracle or Matheronian madness?

Spatial dependence between measured values in an ordered set ought to be verified by applying Fisher’s F-test to the variance of the set and the first variance term of the ordered set. When applied to sets of test results for single boreholes I came to call it fingerprinting boreholes. SME’s reviewers liked it a lot. And so will members of ISO/TC69/SC6 once the upshot of spatial dependence on confidence limits for central values is clear. Assuming spatial dependence between measured values and interpolation by inserting functionally dependent values between measured values has made a mess of the study of climate change. Surely, CAC/ISO/TC69/SC6 has a role to play in selecting the most fitting statistical methods.

No comments: