Kindly comments on Googling is the first step. Consider adding scholarly searches to your arsenal. - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (45)
As a test case, I tried applying this technique to the Dangers of Red Meat, which is apparently a risk factor for colorectal cancer. The abstracts of the first few papers claimed that it is a risk factor with the following qualifications:
if you have the wrong genotype (224 citations)
if the meat is well-done (178 citations)
if you have the wrong genotype, the meat is well done, and you smoke (161 citations)
only for one subtype of colorectal cancer (128 citations)
only for a different subtype not overlapping with the previous one (96 citations)
for all subtypes uniformly (100 citations)
no correlation at all (78 citations)
Correct me if I'm wrong, but most of those look like the result of fishing around for positive results, e.g. "We can't find a significant result... unless we split people into a bunch of genotype buckets, in which case one of them gives a small enough p-value for this journal." I haven't read the studies in question so maybe I'm being unfair here, but still, it feels fishy.
You may be right. It's not quite M&M colors, though; there was apparently some reason to believe this allele would have an effect on the relationship between red meat and cancer. If anything, you might claim that the fishing around is occurring at the meta level: the buckets are "genetics has an effect", "the cancer's location has an effect", "how the meat is cooked has an effect", and so on.
I believe at least part of the reason for this is that "the correlation between red meat and cancer is 0.56" or whatever is not an interesting paper anymore, so we add other variables like smoking to see what happens. (Much like "red meat causes cancer" is a more interesting paper than "1% of people have cancer".) I'm not sure whether this is good or bad.
I punched in "red meat" to google scholar.
http://care.diabetesjournals.org/content/27/9/2108.short 197 citations - concluding that eating red meat "may" increase your risk of type II diabetes.
http://ajcn.nutrition.org/content/82/6/1169.short 173 citations - Shows more "correlations" and "associations" for the "beneficial effect of plant food intake and an adverse effect of meat intake on blood pressure."
Seems accurate.
People who eat red meat tend to:
Do you understand why it's not... entirely honest... to blame red meat? It shows up as a statistical correlate. It can be used to identify people at risk for these conditions, but then researchers make a leap and infer a causal relationship.
It's an ideological punchline they can use to get published. And that's all.
I followed the first link http://care.diabetesjournals.org/content/27/9/2108.short and the abstract there had "After adjusting for age, BMI, total energy intake, exercise, alcohol intake, cigarette smoking, and family history of diabetes, we found positive associations between intakes of red meat and processed meat and risk of type 2 diabetes."
And then later, "These results remained significant after further adjustment for intakes of dietary fiber, magnesium, glycemic load, and total fat." though I'm not sure if the latter was separate because it was specifically about /processed/ meat.
So long as they keep the claim as modest as 'eating red meat "may" increase your risk of type II diabetes.' it seems reasonable. They could still be wrong of course, but the statement allows for that. I should note here that the study was on women over 45, not a general population of an area.
If there's better evidence that the search is not finding, that is a problem.
You do understand that scientist don't just look for correlations but form a bit more complex models than that. Do you seriously think that things like that are not taken into account!? Hell, I am willing to bet that a bunch of the studies test those correlations by comparing for example smokers who eat more red meat versus smokers who eat less/none read meat.
I mean, come on.
Taking everything into account is difficult, especially when you have no idea exactly what you aught to be taking into account. Even if you manage to do that exactly right, there is still publication bias to deal with. And if you are using science for practical purposes, it's even harder to make sure that the research is answering the right question in the first place, Sieben's comments sound anti-science...but really they are frustration directed towards a real problem. There really is a lot of bad science out there, sometimes it is even published in top journals - and even good science is usually extremely limited insofar as you can use it in practice.
I think it's just important to remember that while scientific papers should be given more weight than almost every other source of evidence, that's not actually very much weight. You can't instrumentally rely on a scientific finding unless it's been replicated multiple times and/or has a well understood mechanism behind it.
Red meat adds a literal sizzle to research papers.