MURDER, POLITICS, AND THE END OF THE JAZZ AGE
by Michael Wolraich
Order today at Barnes & Noble / Amazon / Books-A-Million / Bookshop
MURDER, POLITICS, AND THE END OF THE JAZZ AGE by Michael Wolraich Order today at Barnes & Noble / Amazon / Books-A-Million / Bookshop |
Statistics are the magic beans of pop science. Tell someone that heavy drinkers live longer than teetotallers, and she'll say, "How is that possible? What about cirrhosis of the liver?" Tell her there are statistics proving it, and she'll say, "Wow, I never would have guessed that!" That particular stat is making the rounds of the news talk shows right now, and I'd imagine so are a lot of beers.
The first problem with relying on statistics is that most people don’t understand the math very well. To someone trying to prove a point, stats are like a buffet of numbers. For true believers, confirmation bias takes over. Even dispassionate scientists don't always interpret the numbers correctly.
Odds Are, It’s Wrong. Science fails to face the shortcomings of statistics
It’s science’s dirtiest secret: The “scientific method” of testing hypotheses by statistical analysis stands on a flimsy foundation. Statistical tests are supposed to guide scientists in judging whether an experimental result reflects some real effect or is merely a random fluke, but the standard methods mix mutually inconsistent philosophies and offer no meaningful basis for making such decisions. Even when performed correctly, statistical tests are widely misunderstood and frequently misinterpreted. As a result, countless conclusions in the scientific literature are erroneous, and tests of medical dangers or treatments are often contradictory and confusing.
Replicating a result helps establish its validity more securely, but the common tactic of combining numerous studies into one analysis, while sound in principle, is seldom conducted properly in practice.
Experts in the math of probability and statistics are well aware of these problems and have for decades expressed concern about them in major journals. Over the years, hundreds of published papers have warned that science’s love affair with statistics has spawned countless illegitimate findings. In fact, if you believe what you read in the scientific literature, you shouldn’t believe what you read in the scientific literature.
“There is increasing concern,” declared epidemiologist John Ioannidis in a highly cited 2005 paper in PLoS Medicine, “that in modern research, false findings may be the majority or even the vast majority of published research claims.”
Ioannidis claimed to prove that more than half of published findings are false, but his analysis came under fire for statistical shortcomings of its own. “It may be true, but he didn’t prove it,” says biostatistician Steven Goodman of the Johns Hopkins University School of Public Health. On the other hand, says Goodman, the basic message stands. “There are more false claims made in the medical literature than anybody appreciates,” he says. “There’s no question about that.”
Nobody contends that all of science is wrong, or that it hasn’t compiled an impressive array of truths about the natural world. Still, any single scientific study alone is quite likely to be incorrect, thanks largely to the fact that the standard statistical system for drawing conclusions is, in essence, illogical. “A lot of scientists don’t understand statistics,” says Goodman. “And they don’t understand statistics because the statistics don’t make sense.”
Between poor understanding of the math, confirmation bias and the need to publish and perish, it is no wonder that the science establishment insists on peer review and replication of results. But even then, could something go wrong? Yes, the data being sampled may be narrow or skewed.
The Weirdest People in the World?
To be published in Behavioral and Brain Sciences
It’s about a truly unusual group: people from Western, Educated, Industrialized, Rich, and Democratic (WEIRD) societies. In particular, it’s about the Western, and more specifically American, undergraduates who form the bulk of the database in the experimental branches of psychology, cognitive science, and economics, as well as allied fields (hereafter collectively labeled the “behavioral sciences”). Given that scientific knowledge about human psychology is largely based on findings from this subpopulation, we ask just how representative are these typical subjects in light of the available comparative database. How justified are researchers in assuming a species-level generality for their findings? Here, we review the evidence regarding how WEIRD people compare to other populations.
The Behavioral Sciences Database is Narrow ... a randomly selected American undergraduate is more than 4000 times more likely to be a research participant than is a randomly selected person from outside of the West.
Researchers Often Assume their Findings are Universal ...despite their narrow samples, behavioral scientists often are interested in drawing inferences about the human mind and human behavior. ... Leading scientific journals and university textbooks routinely publish research findings claiming to generalize to “humans” or “people” based on research done entirely with WEIRD undergraduates.
[Update & thanks to DF]
Papers by Dr Andrew Wakefield led many doctors and parents to see a correlation between MMR vaccines and increasing frequency of autism in children. Research by Dr Mark Geier, and his son, seemed to implicate the vaccine preservative Thiomersal, which contains mercury, as well. Wakefield's study spurred the withdrawal of MMR vaccines, and vaccines containing Thiomersal were also withdrawn as a precaution.
But eventually, groups like the American Academy of Pediatrics, criticized the Geiers' research as having, "numerous conceptual and scientific flaws, omissions of fact, inaccuracies, and misstatements." Others cited "methodological problems in his research including not presenting methods and statistical analyses to others for verification, improperly analyzing data taken from Vaccine Adverse Event Reporting System, as well as either mislabelling or confusing fundamental statistical terms in his papers, leading to results that were "uninterpretable"."
Wakefield went beyond poor methodology and secrecy. As reported by Brian Deer, he was paid to attack MMR vaccines to support class action litigation that was planned by a UK solicitor. Even his subjects were paid by the solicitor. Geier has been somewhat discredited; Wakefield is no longer a physician.
As described by Tim Lambert on ScienceBlog, a scary combination of narrow samples, bad methodology and lack of ethics has been seen in John Lott's work to promote his theories that gun ownership leads to lower crime rates.
John Lott has claimed that he conducted a survey over three months in 1997 that found that in 98% of defensive gun uses the defender merely has to brandish the gun to break off an attack. It is almost certain that he never did a survey because:
* nine published surveys give numbers ranging from 21% to 67% as to how often defenders shoot, far more than the 2% Lott claims
* Lott claims that his survey found defenders firing in 2 out of 28 cases, which is 7%, not 2%. Nor is it possible that the weighting scheme Lott now claims to have used turned 7% into 2%.
* no evidence that Lott ever conducted such a survey can be found
* Lott has repeatedly changed his story about the source of the 98% figure, variously attributing it to "national surveys" and some particular polls, only publishing the story about the survey in 2000.
* Lott made the 98% claim on Feb 6, 1997, well before his survey was completed.
* Lott changed his story about the survey.
* Despite extensive coverage on the net, in many papers such as the New York Times, the Chicago Tribune and the Washington Post, an ad in the alumni magazine and in the best-selling book Freakonomics, none of the eight students Lott claims conducted the survey have been heard from.
* Lott claims to have "replicated" his survey with a new one that gives 95% brandishing, when in fact the new survey found only seven gun users (one of whom fired), far too small a sample to yield a meaningful brandishing number. (And if you correct his arithmetic you get 91% anyway.)
In 1998, John Lott published a book entitled More Guns, Less Crime. In that book he presented statistical evidence that concealed-carry laws were associated with lower crime rates.
In 2002, Ian Ayres and John Donohue analysed a more extensive data set and found that, if anything, concealed carry laws lead to more crime. Lott responded with a new analysis that he claimed confirmed the "more guns, less crime" hypothesis. Ayres and Donohue's response (April 2003) was devastating—Lott's data contained numerous coding errors that, when corrected, eliminated the results and, this was the second time these sort of errors had been found in Lott's data.
More at Reason.com:
The first round of dispersed investigation came when a Minnesota attorney named David Gross came forward to say he had been the subject of a survey that sounded like Lott's. The Washington Times ran a brief story implying that the question about Lott's survey was now closed.But bloggers were more skeptical: Gross turned out to be a gun rights activist himself, with the group Concealed Carry Reform, NOW! Historian Thomas Spencer unearthed a letter to the Minneapolis Star Tribune in which Gross wrote that gun control advocates "dance on the graves of the innocent victims and glory in their spilled blood." Another blogger, the pseudonymous Atrios, found news reports recounting how Gross had taken over the names of several gun control groups that had neglected to renew their corporate status with the state.
Of course, a gun activist would be the most likely to hear about the controversy and come forward. But skeptics continued to doubt Gross' account, and some of Lott's former defenders wrote that they wished for some further independent confirmation of his account. Lott, currently a resident scholar at the American Enterprise Institute, has conducted a new survey, which he includes in his new book, The Bias Against Guns (Regnery).
Meanwhile, several of the bloggers who had been writing about the controversy -- a group that included me -- drew the ire of someone called Mary Rosh. Rosh, who identified herself as a former student of Lott's who had long admired his fairness and rigor, said that it was irresponsible to post links to the survey debate without calling Lott first. This sounded odd, not only because bloggers very seldom do that kind of background research before posting a link, but because Lott had made precisely the same criticism several times in e-mails to bloggers covering the story.
A Google search revealed that Rosh had for several years been a prolific contributor to Usenet forums, where she regularly and vociferously defended the work of Lott. On a whim, I compared the I.P. address on Rosh's comment to the one on an e-mail Lott had sent me from his home. They were the same.
Lott admitted to using the sockpuppet on usenet, websites and Amazon reviews.
Despite all this, John Lott continues to be cited - at gunfacts.info.
Comments
The mere fact that these scientific studies are performed by, analyzed by and published by humans indicates that they are biased by default. There is no such thing as an unbiased scientific study.
by cmaukonen on Tue, 09/07/2010 - 11:16pm
Neil Young got it right, "Red means run, son... numbers add up to nuthin'."
Tough to argue with that.
by quinn esq on Tue, 09/07/2010 - 11:22pm
You forgot the part about how the whole anti-vax movement was spawned by one guy and one completely fraudulent piece of work that has been thoroughly discredited, but continues to be cited.
Seriously, it's easy to blame the numbers, but it's the people that are the problem. Statistics isn't that hard, but not even rigorous statistical analysis can prevent people from making false claims.
by DF on Thu, 09/09/2010 - 9:52am
I didn't mention the Geiers by name, but I think I covered it. My impression is that thiomersal was removed as a precaution because it contained mercury. The removal spurred parents' interest in Geier's research, which was eventually found to be severely flawed.
by Donal on Thu, 09/09/2010 - 10:10am
I was referring to the study and ensuing scandal involving this guy. The entire history of this episode informs us that in spite of a mountain of statistics proving that there was no evidence of the connection between ASD and the MMR vaccine, a bunch of people wanted to believe this guy. You can't blame the numbers for that.
by DF on Thu, 09/09/2010 - 10:29am
Now that you mention it, I do recalling hearing about Wakefield, but the articles I read highlighted the Geiers' errors.
by Donal on Thu, 09/09/2010 - 11:05am
One more thing that should be said, although it should also be obvious to anyone who thinks about it: assuming all of the statistics are performed flawlessly, you'd still expect about 5% of all scientific studies that show 95% confidence to be wrong (and 1% of all studies that show 99% confidence, etc.). The problem, of course, is in figuring out which 5%.
One of the founders of statistics liked to use 99.9% certainty (and even then, 1 out of every 1000 such results would be due to chance), but scientists quickly found out that such certainty typically required far too many experiments.
by Atheist (not verified) on Thu, 09/09/2010 - 10:28am
Quite right. Good statistical analysis doesn't tell you what's right or wrong. It tells you what the probability that you've arrived at the wrong conclusion is. That can't stop people from excitedly jumping to the wrong conclusion.
by DF on Thu, 09/09/2010 - 10:30am
I fought a losing battle over at TPM, trying to persuade them to change how they reported the reliability of polls. You can't just say, "The margin of error is X%." Typically, it needs to be, "The margin of error is X%, 19 times out of 20."
Big difference. It means you should expect every 20th poll to be off by more than the stated margin. In an election cycle, with hundreds of polls being taken, that makes for lots and lots of really wrong poll results. Yet individual, small-sample polls showing minuscule movement within the margin of error consistently get hyped as showing "trends." They simply can't be used that way, but they are.
by acanuck on Thu, 09/09/2010 - 3:39pm
Sorry I missed this Donal.
I always laugh when some cop in civil dress flashes a badge for 1/10 of a second and threatens someone with 'obstruction of justice' on TV. How am I supposed to know he did not win the 'badge' at a carnival side show?
And 'surveys', well I have been yelling about them for two years but not from a strictly scientific perspective. I just know that Rassmussen has shown the same damn results for approval ratings of President Obama for two frickin years. 45% approve and 55% think he should be shot.
And, at the same time, Gallop's figures change all the time. And they never add up to 100%. Because you and I know that at least 3% of our population does not have the faintest idea who is president.
This is a nice follow up on this Lott guy. Is he related to the ex gay and fanatically dishonest senator?
Real fine work here Donal.
by Richard Day on Mon, 09/13/2010 - 3:19am