Saturday, July 30, 2011

ents

Last post I mentioned that this post would be a technical one on statistics. However, after studying further on the subject of statistics I quickly realized that I am not qualified to author a technical post on the matter. Instead I will here offer a post that works to hold together statistics and scientific epistemology. Think of scientific epistemology being somewhat like Tolkien's tree Ents that he depicted as gathering around in a counsel to decide on a course of action. They refuse to be rushed in their decision making by the immediacy of danger: rather they thoroughly work the problem over to ensure that their decision places them on the side of good. Likewise with scientific epistemology, establishment of true knowledge occurs slowly and without any deference to the urgent need of human beings. So here ensues an elaboration of Ent-like knowledge building interspersed with pictures of past heroes of statistics (everybody should choose at least one statistical hero...such a thing ought to be as ubiquitous as having a NASCAR driver to root for).

I believe that natural human knowledge is capable of more than a mere approximation of reality. I believe that natural human knowledge can achieve a one-to-one relationship with reality; that is to say that theories of reality (i.e.: how things are) can represent specified slices of reality with perfect accuracy (e.g.: Copernicus and heliocentrism; Newton and gravitational forces, etc.). However, I also believe that the vast majority of natural knowledge produced in a given calendar year is specious at worst (i.e.: false knowledge) and an approximation of reality at best (i.e.: analogous knowledge). Here I refer not to the sort of assertions that pass off as knowledge among the popular pundits, but rather I am referring to critically tested, peer-reviewed scientific knowledge. It is this high level of knowledge that ranges from false to analogous. To characterize this range of knowledge quality I have generated a list of four (4) descriptors beginning with the most basic and contingent sort of knowledge progressing up to true knowledge or "fact" (as science would have us call it).

Ronald Fisher: what's not to like about this statistician? The very incarnation of an Ent.

Possible. Definition: that can be; capable of existing. For example an oval circle is quite possible; an oblong circle is quite possible. However, a square circle could never be a circle (i.e.: essentialistically impossible). An MRI machine that produces only rare cheese could never be an MRI machine (i.e.: nominalistically impossible). Certain things are impossible. Importantly, though, many possible things are not true. For instance, although it is possible for a live cow to go over the moon, if I claim to know that a live cow actually went over the moon last night this would be a false knowledge claim. Although it is possible for the sun to rise in the West and set in the East, this is not the truth of what happens on planet Earth. That which is conceivably possible is not necessarily existentially so. This being held true, there are three possible descriptions of possibilities and existence: existent possibility, nonexistent possibility, nonexistent impossibility. There is no such thing as an existent impossibility.

Karl Pearson: note with what alacrity he wields his pen in noble service of statistics.

Probable. Definition: likely to occur or to be so; that can reasonably be expected or believed on the basis of the available evidence, though not proved or certain. For example, a 2002 study published in Spine journal deduced a clinical prediction rule to identify which patients with low back pain are most likely to be successfully treated with spinal manipulation. The study deduced these 5 parts for a prediction rule: 1) pain onset less than 16 days prior to treatment; 2) no symptoms distal to the knee; 3) Fear Avoidance Behavior Questionnaire score less than or equal to 19; 4) one or more hypomobile segments in the lumbar spine; 5) at least one hip with more than 35 degrees of internal rotation motion. A patient must exhibit at least 4 of the above 5 traits to be considered positive on the rule. A subsequent randomized controlled trial published in 2004 compared patients who satisfy this rule with those who do not. They found that patients positive on the rule and receiving spinal manipulation have an adjusted odds ratio for successful treatment of 60.8 when compared to those negative on the rule and receiving exercise. That is to say, a person who satisfies the prediction rule and is treated with spinal manipulation is 60.8 times more likely to have a successful outcome than someone negative on the rule and receiving exercise. Although adjusted odds ratios are not exactly probabilities, they approximate mathematical probabilities closely when baseline risks are low. Consequently we can say that a person with low back pain who satisfies the prediction rule is very likely to gain significant function after only 2 treatments of spinal manipulation and this functional gain is likely to last for 6 months. But, this is a probability. This means that a small number of people who are positive on the rule and receive spinal manipulation will feel no improvement or worsening symptoms after treatment. So this is a practical example of probabilistic knowledge. The above studies did not elucidate fact, they elucidated pragmatic probabilities to help clinicians, insurance companies and patients make a decision about treatment options.

Valid. Definition: sound; well grounded on principles or evidence; able to withstand criticism or objection, as an argument. The move from probable knowledge up to valid knowledge is like passing over the Natural Light beer for Kentucky Bourbon Barrel Ale. Like declining the Hostess Snack Cake to save room for Graeter's ice cream. Like operating on the spine with a high precision Medtronic drill instead of a high speed DeWalt house drill. Valid knowledge within medical diagnostics is produced when a new test for diagnosis is measured against a gold standard. Assessing the new test (lets say for detecting prostate cancer) against the gold standard can result in four possible outcomes: true positive, false positive, true negative, and false negative. The box below illustrates this well:

Gold Standard Negative (no prostate cancer present)

Gold Standard Positive (prostate cancer present)

New Test Positive

False Positive (invalid test result)

True Positive (valid test result)

New Test Negative

True Negative (valid test result)

False Negative (invalid test result)

A variety of statistics are available to describe these results, however I will spare the reader details of these at this time. Let the reader note, though, that validity does not measure probability of departure from truth but rather it measures reality of departure from truth. In this way validity is not a guess at the truth but a true measure of the truth. The greatest mistake in popular conceptions of research occurs when studies analyzed by probabilistic statistics (i.e.: p-values and confidence intervals) are interpreted as if they utilize validity statistics (i.e.: likelihood ratios). In this manner, many studies are presented as valid when they ought truthfully be presented as some grade (e.g.: low, moderate, high grade) of probabilistic evidence for or against a hypothesis.

Thomas Bayes: the right Reverend had a penchant for more than divine truth.

Veritable. Definition: true; real; actual. I am taking the name of this descriptor from the Latin word veritas: truth. How does the descriptor "veritable" differ from the descriptor "valid"? Knowledge that is veritable is perfectly true (1-for-1 correspondence with reality or logic) whereas knowledge that is described by validity is expressed as a degree of departure from truth. Given the above diagnostic example: the new prostate cancer test would be characterized as being valid if it deviates from the gold standard to only a small degree and invalid if it deviates from the gold standard to a large degree. However, the gold standard test is veritable: a container of 100% true knowledge, finding prostate cancer whenever it exists and ruling out prostate cancer whenever it does not exist. Veritable knowledge is not simply produced by a mathematical formula, but is rather arrived at through years of practice, research (statistically analyzed), critical review (e.g.: statistical assessment, construct assessment, etc.), technical reformulation, and even large scale disciplinary enactment (i.e.: use within the guild at large).

Jerzy Neyman: a fierce opponent to Fisher; his calculations were as killer as his Hitler-ish look.

One final cautionary note is due here. Having moved through this post, and engaged with the proposed hierarchical categories of knowledge (low quality "possibility" knowledge up to highest quality "veritable" knowledge) the reader might be tempted to scorn knowledge that is beneath "veritable" in my hierarchical scheme. The purpose of these hierarchies is not to cast aspersion upon "inferior" knowledge, but rather to appropriately characterize the levels of knowledge and their relation to truth. This has very pragmatic implications. Possible knowledge should be utilized with extreme caution and proposed lightly with a ready willingness to believe alternative possibilities; probable knowledge should be utilized discriminantly and defended publicly with openness to evidence to the contrary; valid knowledge should be ubiquitously employed and publicly defended with great vigor and energy; veritable knowledge should be published and re-published with regular frequency as well as being proclaimed with certainty to winsomely disabuse falsehoods consistently where falsehoods appear.

Next post will be devoted to a description of basic statistics and some of their mathematical formulations.

Saturday, July 23, 2011

monochrome

Two and one half years have ticked off the mantle top clock since the last post on this blog. Even more embarrassing is a swift look at the goals mentioned in that last post: publishing a case study in the national physical therapy journal (that didn't happen), become a researcher (not much traction on that over 2.5 years), becoming a voice that shapes physical therapy knowledge (I haven't even started talking yet). Alas and alack, my life was actually interesting over the last couple years: job changes, selling/buying a house, first born daughter arrived on the scene, theological growth, starting graduate classes again. No time for the monochrome world of research. I had some life to live. Life is settling down a little now. Hopefully a little more tranquility will invade my work life over the coming months and then I can really begin in earnest tackling scientific epistemology viz a vi statistics and clinical research.

Interestingly enough, my introduction to biostatistics course this summer is a refreshing reminder that learning is fun. My instructors Drs. Shukla and Dwivedi are doing a superb job of instructing the class at a philosophical level. They are engaging questions of statistical certainty within the constraints of probabilistic mathematics (a world they refer to as fuzzy or gray) counterbalancing the clinical drive for a definite answer for ailing patients. The most unique concept I have learned to date is this: hypothesis testing (i.e.: level of significance and p-value) does not result in a mathematical acceptance or rejection of the hypothesis. Rather, hypothesis testing adds or removes weight to an a priori belief that the hypothesis is true. In this manner, statistics are legal procedures. The test statistic is the prosecuting trial attorney arguing the state's case in front of the judge and jury. The data is considered insignificant unless otherwise proven beyond reasonable doubt by the test statistic.

Next post will be a technical one exploring some aspect of statistics and how they relate to contemporary medical knowledge.