Talk:Positive predictive value

From Wikipedia, the free encyclopedia

WikiProject Medicine This article is within the scope of WikiProject Medicine. Please visit the project page for details or ask questions at the doctor's mess.
B This page has been rated as B-Class on the quality assessment scale
Mid This article has been rated as Mid-importance on the importance assessment scale


Contents

[edit] Table and edits

See Talk:Sensitivity (tests) re past wish list for simpler description, setting what it is before launching in mathematical jargon. I have also added a table and in Sensitivity (tests) added a worked example. The table is now consistant in Sensitivity, Specificity, PPV & NPV with relevant row or column for calculation highlighted. David Ruben Talk 02:45, 11 October 2006 (UTC)

[edit] "Physician's Gold Standard" Remove?

"Physician's gold standard" seems to be an unhelpful phrase as it is used in this article.

My experience has been that when "gold standard" is used in this context it refers to the reference test against which the accuracy of a test is measured. As we all know, sensitivity, specificity, PPV, etc., require a "gold standard" test for reference -- otherwise we don't have a basis for claims about % true positives and % true negatives.

Here it seems that "physician's gold standard" means something like "it is the statistical property of a test that is most useful to physicians".

It seems that either the author was confused about the use of "gold standard" in biostatistics or there's another (unfortunate) use of the phrase that I'm not familiar with. Since I don't know which, I'm not editing the page. If others agree, perhaps this phrase should be replaced.

--will 02:19, 24 July 2007 (UTC)

I agree with 'will' above (24 july 2007) that PPV cannot be seen as a gold standard. I suggest to replace 'It is the physician's .....' by 'It is the most important measure for ruling in disease' Any objections? Michel soete 16:52, 19 October 2007 (UTC)
Agree "physician's gold standard" is awful and needs rehrasing. Like your suggestion Michel soete... go ahead :-) David Ruben Talk 20:50, 19 October 2007 (UTC)

[edit] the need for an unequivocal definition of positive predictive value

Let's consider following tabel (Grant Innes, 2006, CJEM. Clinical utility of novel cardiac markers: let the byer beware.)

Table 3. Diagnostic performance of ischemia modified albumin (IMA) in a low (5%) prevalence population.

ACS   Yes No Total Sensitivity (true-positive rate) = 35/50 = 70% 
IMA + 35 722  757 Specificity (true-negative rate) = 228/950 = 24% 
IMA – 15 228  243 Positive predictive value = 35/757 = 4.6% 
      50 950 1000 Negative predictive value = 228/243 = 94% 

The positive predictive value is smaller than the prevalence. We must conclude that a positive test result decreases the probability of disease or in other words that the post-test probability of disease, given a positive result, is smaller than the pre-test probability (prevalence): very strange and unusual conclusion.

From a statistical point of view this very strange conclusion can be avoided by interchanging the rows of thet table: IMA- becomes a positive test result. This operation results in a predictive value of 6.17%. The conclusion is that a positive test result, if the test is of any value at all, increases the post-test probability as it is expected to do and in no case decreases this value.

This example illustrates the need for an unequivocal definition of a positive test result. If a positive test result is unequivocally defined, the positive predictive value is mathematically unequivocally defined. A text providing such an unequivocal definition was removed by someone who called it 'garble'. I intend to put the text back, any objections? —Preceding unsigned comment added by Michel soete (talkcontribs) 18:57, 22 September 2007 (UTC)


Yes - makes no sense, 'garble' indeed. I've removed it and placed here in talk page where we can work on this.

And, alternatively, too: PPV = PR * LR+ / (PR * (LR+ - 1) + 1)
wherein PR = the prevalence (pre-test probability) of the disease, * = the multiplication sign and LR+ = the positive likelihood ratio. LR+ = sensitivity / (1 - specificity). The prevalence, the sensitivity and the specificity must be expressend in per one, not in percentage or in pro mille a.s.o.. The frequency of the True Positives must be this frequency that exceeds or equals the expected value, mathematically expressed: True Positives >= (True positives + False Positives) (True Positives + False Negatives) / N wherein N = True Positives + False Positives + True Negatives + False Negatives. If this condition is not met and if the sensitivity differs from .50 (50%) then two different results after the calculation of sensitivity are possible since the rows of two by two tables can be interchanged and then a former positive result can be called a negative, a former negative result can be called a positive (Michel Soete, Wikipedia, dutch version, Sensitiviteit en Specificiteit, 2006, december 16th).

As a start, lets use same terminology as rest of article, ie call PR just Prevalence, no need explain maths symbols. If LR+ is "sensitivity / (1 - specificity)", then I get:

PPV = Prevalence * sensitivity / (1 - specificity)
     --------------------------------------------
     Prevalence * ((sensitivity / (1 - specificity)) - 1) + 1

Lets multiply through by (1 - specificity):

PPV = Prevalence * sensitivity 
     --------------------------------------------
    (Prevalence * (sensitivity - (1 - specificity)) + (1 - specificity)

Which is:

PPV = Prevalence * sensitivity 
     --------------------------------------------
     Prevalence * sensitivity - Prevalence + specificityPrevalence + 1 - specificity

and so to:

PPV = Prevalence * sensitivity 
     --------------------------------------------
     Prevalence * sensitivity + (1-specificity)(1- prevalence)

ie exactly the same as the last formula already given in the article ! This fails to add therefore a new insight into its derivation or meaning.

As for "The frequency of the True Positives must be this frequency that exceeds or equals the expected value, mathematically expressed: True Positives >= (True positives + False Positives) (True Positives + False Negatives) / N wherein N = True Positives + False Positives + True Negatives + False Negatives. If this condition is not met and if the sensitivity differs from .50 (50%) then two different results after the calculation of sensitivity are possible since the rows of two by two tables can be interchanged and then a former positive result can be called a negative, a former negative result can be called a positive" - sorry can't even begin to get my head around this.

  • Why must TP be larger than the expected values?
  • The conditional formula your seek is the same as TP => Positive predictive value * Sensitivity, but what is this expressing in everyday words ?
  • How can there be two different results possible ?
  • Surely just needless convolution to start supposing what happens if switching rows about ? Might as well say switching a "test result that excluded a disease" to a "test result that confirmed a disease" - one can't start switching values. One defines at the start what a positive or negative result means (ie what the null hypothesis is) and then should stick to it thoughout the analysis. David Ruben Talk 11:46, 27 September 2007 (UTC)

[edit] allowing ambiguity

My mother tongue is dutch. Initially I did not understand quite well what garble is but now I think it is the same of nonsense.

Not quite the meaning I meant, more that it was so convoluted/mixed up/unclear as to loose the intended meaning.David Ruben Talk 15:00, 29 September 2007 (UTC)

Let us assume that allowing ambiguity is a good option. Following tables can then be constructed:

                 D+      D-               D+        D-
blue (P)         99 (a)  1  (b)    red (P) 1        99
red  (N)          1 (c) 99  (d)    blue(N) 99        1

Constructing these tables I respected some conventions: The frequencies of diseased people are in the first column, the frequencies of the positives in the first row, the frequency of the true positives in cell a.... a.s.o..

Now we can write that sensitivity is a / (a + c). For those for whom blue is positive the sensitivity is 99%, for those for whom red is positive the sensitivity is 1%. The positive predictive value ( a / (a + b)) is 99% (blue is positive) or 1% (red is positive).

I now understand where you see the alternative way of looking at the data (indeed one could go switching sensitivity for specificity), but this is precisely my point about needing to be very clear from the outset about the meaning of the test (the null hypothesis) and what a positive or negative result means. To start talking about how well a test result confirms a disease and then start considering how the same test might be viewed as a marker of no disease (ie is a positive result that for picking up disease or is a positive result that of identifying the normal) is to dither between positive & negative results, PPV & NPV, specificity & sensitivity. One should define what the test indicates and then staying with that, interpret the results - there can only be a single PPV, a single NPV, a single specificity, a single sensitivity for any given set of data.David Ruben Talk 15:00, 29 September 2007 (UTC)

Such a possibility for ambiguity is not in line with traditional medical thinking and therefore it leads to (at least seemingly) contradicory statements and therefore confusion.

Megan Davdson writes (2002, The interpretation of diagnostic tests: A primer for physiotherapists): 'Where sensitivity or specificity is extremely high (98-100%, interpretation of test results is simple. If the sensitivity is extremely high, we can be sure that a negative test result will rule the disease out.' If ambiguity is allowed we have to add 'or extremely low (0-2%)' and 'If the sensitivity is extremely low, we can be sure that a positive test result will rule disease out'. Moreover, the relatively new concepts SpPIn and SnNOut are described in the article. It are acronyms. A SpPIn is a test with such an extreme high Specificity that if a test result is Positive disease can be ruled In. A SnNOut is a test with such an extremely high Sensitivity that if the test result is Negative the disease can be ruled Out.

Thus our demand that a > the expected value in cell a is a solid basis for these concepts and their names and for the classical ideas that they incorporate. Also the strong living idea that a positive test result always points to disease find in this demand a firm basis.

I hope that the argumentation above were convincing enough and that the removed text will be put back by the person that removed it.

81.244.101.52 12:07, 29 September 2007 (UTC)


Have to disagree with "If the sensitivity is extremely high, we can be sure that a negative test result will rule the disease out'" - where sensitivity is high, this means only that with a positive result we can be reasonably sure that the disease is identified. Sensitivity has no direct bearing on the truely healthy, only on those with disease. Consider:
       Disease   Healthy
   +ve    980       10 
   -ve     20       10
This has sensitivity of 98% (980/(20+980)) yet it can hardly be said that "a negative test result will rule the disease out" - quite the opposite, of those with a negative result, two thirds will have the disease (20/(20+10)).
Now for the second claim of "'If the sensitivity is extremely low, we can be sure that a positive test result will rule disease out'", equally untrue:
       Disease   Healthy
   +ve    20        1 
   -ve   980       99
This test has low sensitivity of 2% (20/(20+980)), but high specificity of 99% (99/(1+99)), yet a positive result is far from reassuring but suggests over a 95% chance for really being ill (20/20+1)). Of course this test is so poor that it fails to meaningfully help identify disease vs healthy, given that in this example 91% of subjects had the disease !
I think your textbook for physiotherapists is being simplistic in its outlook and attempts to guide the reader - would have been better if its author had stuck to describing the standard terms, rather than trying to create new "rules-of-thumb".David Ruben Talk 15:00, 29 September 2007 (UTC)

[edit] unequivocal definition of statistical measures

Hi Davidruben,

I did not claimed that a test with a high sensitivity, given a negative test result, ruled disease out, it was Megan Davidson. It was not Megan Davidson who claimed that a low negative test, given a a positive test result ruled disease out, I was it who stated that this should be added in her article if ambiguity was allowed. Davidson did not wrote a textbook but an excellent article of six pages on the subject.

You disagree with those claims, I disagree too but Davidson and Grant Innes are examples of classical thinking about the subject. Moreover, they have strong argumentation for their point of view. Davidson writes: 'Unfortunately the predictive values only apply when the clinical prevalence is identical to that reported in the study. Prevalence changes dramatically depending on where the test is being performed.'

Grant Innes writes: 'In reality, predictive value is less a measure of test performance than it is a reflection of disease prevalence in the population being tested.'(op.cit.)

His illustrating examples are good. But I disagree with both of them. Your examples are good and following example is it too:

             D+         D-
       +ve   99        99 
       -ve    1         1

No further comment on this tabel needed, I suppose.

I consider their point of view as an expression of what was generally believed in the former century and, I suppose, by many if not most of the physicians today.

I stress the point that allowing ambiguity in defining a positive result does not result in ambiguity of the conclusion for the testee. Blue remains in any case the color that ends up in the conclusion D+. For the patient it is of no importance if the sensitivity is called 90% or 10% and if this conclusion is the result of what is called a positive or negative test result. For him, blue is disease.

The null hypothesis on its own does not says wich test result is positive. For two by two tables the null hypothesis says that experimental data will not deviate (significantly) from the table of expected values. My demand is decisive for what must be considered as a positive result (a must be higher than the expected value in cell a). It results in a situation wherein only one sensitivity a.s.o. is possible. A positive result is then not a result of a decision but of a calculation.

By the way, in my opinion, a cheap, innocent poor test may have very good utility. A potential good test is a test where the test result shows association with disease. The utility of a test is depending on decisions and is not only a characteristic of a test if there is association between test results and disease. Let us assume that the physician or the patient is satisfied with a probability of 97% to decide to a dangerous treatment then a very poor, cheap test increasing the post-test probability from 93% to 97% is potentially a very usefull test.

So, I hope that you will convinced that it is to be preferred to put my 'confusing' text back.

Michel soete 20:05, 29 September 2007 (UTC)

I hope my phrasing was clear that I was critised the textbook and certainly not yourself. I think their points probably try to refer to situations where the disease is quite rare and hence my previous very skewed examples would not apply. Hence their observations have a prerequest (and unstated) assumption that the disease occurs in a small fraction of the total population). The test examples I gave though would still be reasonable if the "disease" is not living past 80 (most people do not), and the "test" was say a mad-scientists claim that enquiring if people ate more than 2 apples a day could predict those who would live healthy long lives (one can conceive that this would be a useless test). I accept your comments immediately above demonstrate understanding of the situation, but the points are I'm sure too esoteric (?pedantic or convoluted better adjectives?) to help explain these parameters in what is just a general encycloipaedia (i.e. they might be appropriate in a full textbook on statistics, but that is not Wikipedia's role). So, whilst I've appreciated discussing this with you (it certainly made me review my own understanding of several issues), I still do not personally feel the paragraph should be included - sorry :-) Would be interesting to hear if other editors have any thoughts on this... David Ruben Talk 23:14, 29 September 2007 (UTC)

Hi David Ruben

I think I can understand quite well your hesitation. I suppose that nobody should hesitate to prefer a sensitivity of 99% in the last example I gave and therefore my new example was not convincing (but perhaps somewhat shocking). It is logical for a measure as sensitivity that everyone desires that it is high and for these tables there is no good reason to prefer a very low sensitivity. Applying my requirement the conclusion is too that the sensitivity is 99% and not 1%. But the problem of the table of Grant Innes is therewith not solved and this is not a esoteric, pedantic problem. It is a real life problem.

I looked on the website of wynneconsult.com and there I found the following (in dutch): 'The probability of a positive test result, as the patient has the disease, is called sensitivity. The sensitivity has to be as high as possible.' This is quite reasonable, I believe. They write too: 'The probability of a negative test result in absence of disease the disease is called specificity and it must also be as high as possible.' This too is quite reasonable, I think, but it is a pity that, reconsidering the table of Grant Innes, both requirements cannot be met at the same time. Indeed both should be at least as high as minimum 50%. We must make a choice and on what basis? So the requirements are not of general value and it is for that reason that I proposed a new requirement, it is an objective basis to make this choice.

Moreover if the unindependant variable is a numerical variable sensitivity and specificity can be manipulated by changing the cut-off points. If the positivity decreases the sensitivity will decrease too and there will be a cut-off point that is low enough to cause a sensitivity that is lower than 50%. What then? Switch positive results into negative results to meet the requirement of as high sensitivity as possible again? I do'nt like the idea.

For all those reasons and yet a few others I proposed my requirement that solves those problems. There is even no loss: if sensitivity in some cases will be lower than 50% it will be at the profit of specificity and it will be justified.

I thank you for your efforts to answer.

81.244.101.52 20:11, 30 September 2007 (UTC)

(this is reaching the limit of my understanding/recollection on the topic). I agree that adjusting the cut-off point for considering a test positive or negative will effect the resulting sensitivity & specificty of the overall test. Two queries:
  1. Is this not the difference between Discrete and Continuous variables (eg have/have not cancer as against a continuum of say haemoglobin values being used to describe someone as having anaemia ?
  2. The sliding of cut-off point to define a continuous parameter to affect sensitivity&specificity that you mentioned - has this not got a formal name ? The term "test utility" came to mind, but I could only find the following articles which confused me (and I'm not sure therefore what is the correct concept): Posterior probability and Likelihood-ratio test David Ruben Talk 00:39, 4 October 2007 (UTC)

Ah, remembered - the changing of a cut-off point and assessing best specificity/sensitivity is called Receiver operating characteristic
Simple binary parameters (has or has not cancer, lesion seen or not seen on Xray) gives a single set of data, and so a single value for specificity & sensitivity - and we either like or dislike the test being considered. However where one can collect continuous parameter data (eg some measure of irregular patterning on say an ultrasound scan of the liver against whether a liver biopsy does/does not confirm cirrhosis), then one has an option of deciding what arbitary cut-off value to use in deciding whether the test is to be regarded as being positive or negative (ie whether the ultrasound scan is suggesting cirrhosis or not). Clearly if the cut-off irregularity is decided to be none (0) then every scan is considered positive (sensitivity will be 100%, but specificity is 0%). Conversely a cut-off of maximal irregularity (an impossible concept as any diseased liver could always be that bit worse) means that no scan is ever considered positive (specificity would be 100%, but test sensitivity is 0%). More reaslistically, some mid-point degree of abnormality on the scan may give us best combination of sensitivity & specificity (ie a "good test").
The Receiver operating characteristic article mentions the very concept of switching data (inverting all positives/negitives) to invert sensitivity & specificity that you previously mentioned (and I now understand why you raised this as being of importance). But this does not belong in this article - these issues are to do with selecting the test parameter and an overview should perhaps be added to Design of experiments, Dependent and independent variables, Cost-benefit analysis or Statistics (none seem quite the right place), in addition to the current limited unclear explanation at the Receiver operating characteristic article. But it does not belong in the lower heirachy articles of Positive predictive value, Negative predictive value, Sensitivity or Specificity.
Finally just to conclude the description as to why selecting the cut-off is important, this comes down to trying to decide what counts as a "good test". This depends on a subjective assessment, in part based on Cost-benefit analysis of the various outcomes (giving rise to an assessment of utility). Two examples to illustrate the extremes of assessment:
  1. Consider a cheep test that is set to pick up a potentially catastrophic (but easily fixable) fault on a space shuttle. Clearly ensuring that all possible faults are detected even if quite a large number of false positives occur, is going to be desirable (given the outcome for missing a fault). Whereas a different cut-off for the test which has less false positives but might allow a fault go undetected (ie some false negatives), is not going to be acceptable to NASA.
  2. Going back to our ultrasound scans for possible cirrhosis, if a positive scan must be followed by a risky biopsy investigation to confirm the diagnosis, then it is imperative to minimise the false positives (ie risk of harming those who in reality are healthy), and a lesser sensitivity is accepted to give a better specificity (the "imperative" being the ethical Risk-benefit analysis which in turns derives from the instruction to "first, do no harm" being a higher priority than the hippocratic oath's "for the good of my patients") David Ruben Talk 00:54, 5 October 2007 (UTC)


Researching for another article, I coincidentally just found PMID 17350133 which is an example of using [[Receiver operating characteristic to "determine the independent clinical risk factors and their optimal cut-points associated with impaired glucose tolerance (IGT) and dysglycemia (IGT or diabetes)."David Ruben Talk 01:01, 5 October 2007 (UTC)

[edit] PPV and Prevalance

Hi David Ruben

If noboby will make obejections in the first days I will add to the text after 'is the proportion of patients with positive test who are correctly diagnosed.' the following: 'The positive predictive value must exceed the prevalence.' This results in the same as the text that was remved.

Justification:

Let us consider following table filled with expected frequencies.

               D+        D-
     red (pos) 12 (a)    28 (b)    40
     blue (neg)18 (c)    42 (d)    60  
               30        70       100

Prevalence = 30%, positive predictive value = 30%. If there is no association between colour and D+, D- prevalence and positive value are equal. Let prevalence and positivity (a + b)/(a + b + c + d) remain equal but a = 13 (then b = 27, c =17, d = 43). The positive predictive value will then increase with 2,5% and will exceed the prevalence. The more a increases the more the positive predictive value will exceed the prevalence. So far no problems. But what if a decreases? Let a = 11 (then b = 29, c = 19 and d = 41). The predictive value decreases to 27,5% and is lower than the prevalence. The more a decreases the more the positive predictive value will decrease. Conclusions: In some cases it is better to predict the presence of the disease with the prevalence (preferring indicating at random 30 persons in hundred patients as possibly having the disease than on a positive test result)and a positive test result can decrease the possibility of the presence of the disease in comparison with the prevalence. I think many will dislike such conclusions. The problem can easily be solved by interchanging the rows in the table and call blue positive and red negative. I suppose that Grant Innes in the table above thought that a high level of IMA would make it more probable that ACS was present (or become present)and therefore called IMA+ a positive test result. The data does not confirm this theory and it should have been better to adapt to the data (or reject the study for being without quality) and call a low level of IMA positive. Accepting the demand that a positive predictive value must exceed the prevalence makes conclusions possible that are easy to accept and perhaps in earlier times and yet nowadays believed by most people: that a positive test result always increases the possibility of the presence of the disease, that LR+ is always greater than 1 and yet others. This demand brings more order in this area of medical statistics. Therefore this demand (in this form or in onother) is, in my opinion an essential element of the definition of the positive predictive value and cannot be omitted without risking seemingly contradictory statemants in regard with some tables.

Michel soete 18:46, 6 October 2007 (UTC)

Other than using an alternative for the word "must" in the suggested additional phrase "The positive predictive value must exceed the prevalence.", I agree as far as yoiur point goes, but just as sensitivity is paired with specificity in considering a good test, so a high PPV ought to be paired with NPV. The word "must" assumes always, and as you point out a test that has no ability at all to discriminate disease from healthy will have PPV=Prevalance. Rather what one hopes for with a test is that indeed PPV>Prevalance, and ideally PPV>>Prevalance, so better phrasing might be "One measure of a good test is that its ability to correctly identify those with a condition (PPV) is higher than the underlying prevalance rate".
But PPV>Prevlance is not in itself a guarantee that the test is "good", consider a test good at identifying severe cases of a disease but hopeless with milder forms:
               D+        D-
     Test pos   9        1    10
     Test neg  31       59    90  
               40       60   100
So a positive test result is a very strong pointer to having the disease: PPV=9/(9+1)=90% vs Prevalance=(9+31)/100=40%, so yes PPV>>Prevalence. However the test is lousy as it misses 31 of the 40 with the disease (77.5%), ie its sensitivity is just 22.5% (9/(9+31)). Keeping to the consideration of the worth of the study by test outcomes, also we should seek NPV>1-Prevalance. In this example NPV=59/(31+59)=65.6% and (1-Prevalance)=60% - hardly any difference and hence a poor test. Lets improve on out test and find we get:
               D+        D-
     Test pos  29        1    30
     Test neg  11       59    80  
               40       60   100
Here the same Prevalence 40% and improved PPV=96.7%, so PPV>>P. But now NPV=73.75% which has pulled further ahead of (1-Prevalance)
So perhaps to both PPV & BPV article we should add "Ideally a test is better able to both identify and exclude those with a condition (PPV & NPV) than the underlying prevalance rate. Hence one seeks a test where PPV>>Prevalance and NPV>>(1-Prevalance)" David Ruben Talk 00:05, 7 October 2007 (UTC)

Hi David Ruben,

If PPV = prevalence than a predictive value was calculated but not a positive predictive value. We could perhaps say we calculated a neutral predictive value. If we thought we calculated a PPV and find that 'PPV' < prevalence then we calculated the post-test probability of disease given a negative test result but not PPV. Our initial hypothesis was proven to be wrong if 'PPV' < prevalence. A test result is positive if it makes we can assess that there is a higher probability of disease than we can assess on the basis of prevalence alone. Therefore I remain convinced that PPV must exceed prevalence and that this is essential for the definition of PPV. PPV is not a measure of the quality of a test. For every PPV it is possible to construct tables that show that the test is of no, low, high value. The suggestion of taking in account both PPV and NPV in assessing the quality of a test is very interesting but it seems to me that it is not relevant for the definition of PPV. By the way, accuracy ((a + d)/ (a + b + c + d) is an overal measure of the quality of a dichotomous test. Perhaps your suggestion leads to a yet more meaningfull measure for the overall quality of a test. I fear that such overall measures could hide the fact that a test can have very moderate overall quality but can be execellent in ruling in or ruling out disease. For instance the ANA test is, on its own, a very moderate test in ruling in SLE but excellent in ruling out SLE.

Michel soete 15:43, 8 October 2007 (UTC)

Michel, you are needlessly complicating the terminology. "Positive" Predictive Value is fine irrespective of the relative values of PPV and prevalence. Without the word "positive" it is unclear what is being measured. It also doesn't make sense for the name of a statistic to depend on its value in comparison to another statistic. PPV tells me the number of actual diseased individuals who had a "positive" value on the test. Prevalence is not part of that definition.

--Loonatickle —Preceding unsigned comment added by Loonatickle (talk • contribs) 21:07, 14 May 2008 (UTC)

Hi David Ruben,

I have the feeling that thus far I could not fully convince. In every day words my demand equals stating that the post-test probability of disease given a positive result must always be greater than the post-test probability of disease given a negative test result except in the case of no association between variable and disease.

Let's prove it: Let the expected frequencies be a', b', c' and d'. Fur such a table we agreed that PPV = prevalence. My demand was initially that a > a', thus a = a' + x, x being a positive number. Since the marginal totals do not change we can write a' + b' = a + b and b = b' - x. It is obvious that (a' + x)/(a' + b') > a'/(a' + b') since a' + x > a'. Thus we can write a / (a + b) > prevalence since a'/(a' + b) = prevalence. Remark that this is the same as saying that the PPV > prevalence. If a = a' + x then c = c' - x. It can be proven in a similar manner that c/(c + d) < prevalence. Now we can write a / (a + b) > prevalence > c/ (c + d). Thus a/(a + b) > c/ (c + d) what was to be proven.

Now we can state that the post-test probability of disease given a positive test result is always greater than the post-test probability of disease given a negative test result. Without my demand this cannot be stated. This demand makes also others statements possible as for instance that LR+ > 1 and that LR- <1 a.s.o. For those who think that this conclusion is possible without a demand for a I recommend the calculation of LR+ and LR- on the table of Grant Innes above.

Michel soete 11:47, 17 October 2007 (UTC)