Friday, November 14, 2014

A Groundhog Day blog redux

Someone wondered the other day why we keep saying the same thing over and over on our blog.  "Ok, ok, we know things are complex, get on with your life." We, of course, wonder why we have to keep repeating ourselves.  But his query reminded me that we've dealt with this issue before, so we're rerunning a post from 2012.

The Groundhog Day blog?

Sometimes it seems that we're posting the same story over and over again.  Here are some new study results, here's what the authors say they mean, and here's what we think they really mean.  Usually a lot less than the authors report.  Just this week, does aspirin prevent cancer?  Should we eat eggs?  And a post asking simply how we can tell if results are credible.  If you read us regularly you know we don't just pick on epidemiology.  We give genetics the same treatment -- why should we believe any GWAS results, e.g.?  And should we expect to find genes 'for' most diseases?  Or behaviors?  The same for all those adaptive stories that 'explain' the reason some trait evolved.  And Holly is equally circumspect about claims in paleoanthropology, which of course is why we love her posts!

Is it just being curmudgeonly to ask these questions?  Or is it that where some see irreducible complexity others see a simple explanation that actually works?

An isomorphic problem
The important thing about these various issues in modern science is that from the point of view of gaining knowledge about the causal world, they are isomorphic problems.  They have similar characteristics and are (currently) addressed by approaches with similar logic--in terms of study design, and similar assumptions on which both study design, data collection, and methods of analysis are based.  The similarities in underlying causal structure include the following:
  1. Many different factors contribute causally to the outcome
  2. Most of the individual factors contribute only a small amount
  3. The effect of a given factor depends in various ways on the other factors in the individual
  4. The frequency of exposure to the factors varies greatly among individuals
  5. Sampling conditions (how we get the data we use to identify causal elements) vary or can't really be standardized
  6. The conditions change all the time
  7. The evidence for causation is often indirect (esp. in reconstructing evolution)
  8. We have no underlying theory that is adequate to the task, and so we use 'internal' criteria
These days, we use the word 'complexity' to describe such situations.  That word is often used in a way that seems to imply wisdom or even understanding on the part of those who use it, so it has become a professionalized flash-word often with little content.

Often, people use the word, but persist in applying enumerative, reductionist approaches that we inherited over the past 400 years largely from the physical sciences (we've posted on this subject before).  This is based essentially on the repeatability of experiments or situations.  We try to identify individual causal elements and study them on their own.  But if the nature of causation is the integrated effects of uniquely varying individuals, then only the individual strong (often rare) factors will be easily identified and characterized in this way.

Item #8 above is important.  In physics we have strongly formal theory which yields precise predictions under given conditions.   There is measurement error, and the predictions are sometimes probabilistic, but the probabilities involved and the statistics of analyzing error, were designed for such situations.  We compare actual data to predictions from that externally derived theory.  That is, we have a theory not derived from the data itself.  It is critical to science that the theory is largely derived not just in our heads but from prior data.  But it's external to new data that we use to test the theory's accuracy.

In the situations we are facing in genetics, evolution, biomedicine, and health, we have little similar theory, and the predictions of what we have are not precise or our assumptions too general.  Even the statistical aspects of measurement error or probabilistic causation are not based on rigorously specified expectations from theory.  Our theory is simply too vague at this stage.  So what do we do?

We use internal test criteria.  That is, we test the data against itself.  We compare cases and controls, or different species of apes' skeletons, or different diets.  We don't use some serious-level theory to predict that so many eggs per day, or some specific genotype at many sites in the genome will have some specific effect based on primary biological theory, but only that there is a per-egg outcome. We don't know why, so we can't really test the idea that eggs really are causal, because we know there are many variables we just aren't adequately measuring or understanding.  When we do find strong causal effects, however, which does happen and is our goal of this kind of research, then subsequently we can perhaps develop a real theoretical base for our ideas.  But the track record of this approach is mixed.

This is also often called a hypothesis-free approach.  For most of the glory period in science, the scientific method was specifically designed to force you to declare your idea in a controlled way, and test it (the 'scientific method').  But when this didn't work very well, as in the above areas, we adopted a hypothesis-free approach that allowed internal controls and tests: our 'hypothesis' is just that eggs do something: we don't have to specify how or why.  In that sense, we are simply ignoring the rules of historically real science, and even boasting that we are doing science anyway, by just collecting as much data as we can, as comprehensively as we can, in the hopes that some truth will fall out.

The central tenet of science for the last 400 years has been the idea that a given cause will always produce the same effect.  Even if the world is not deterministic, and the result will not be the same exact one, it will at least have some probability distribution specifying the relative frequency with which we'll observe a given outcome (like Heads vs Tails in coin-flipping).  But we really don't even have such criteria in the problems we're writing about.  Even when we try to replicate, we often don't get the same answer, and do not have good explanations for that.
When we're in this situation, of course we can expect to get the morass of internally inconsistent results that we see in these areas, and it's for the same basic epistemologicalreason!  That is, the same reason relative to the logic of our study designs and testing in these very different circumstances (genetics, epidemiology, etc.).  Yet that doesn't seem to slow down the machine that cranks out cranky results: our system is not designed to let us slow down to do that.  We have to keep the funds coming in and the papers coming out.

And then of course there's cause #9. Most of us have some underlying ideology that shapes our interpretation of results.

This is all a fault of us and the system.  We can't be faulted for Nature's complexity.  The issues are much more--yes--complex than we've described here, but we think this captures the gist of the problem.  Scientific methods are very good when we have a good theory, or when we are dealing with collections of identical objects (like oxygen or water molecules, etc.), but not when the objects and their behavior are not identical and we can't specify how they aren't.  We all clearly see the problem.  But we haven't yet developed an adequate way to deal with it.

No comments:

Post a Comment