Dear Paul,
since I was in the consensus for my last response, I give you again my
response to this new problem.
The principle of my solution is always the same: to try to build a
probabilistic model.
a) I first reformulate the problem in more familiar notations for me,
with a diagnosis D and 2 signs S1 et S2.
The first report is S1, the second one is S2, and the diagnosis is
location Y for X.
Your data are: P(D/S1) = p1, and P(D/S2) = p2
Your question is: P(D/S1 and S2) = p12 = ?
b) the problem is clearly underparametrized, so I would make 2 assumptions:
A1: S1 and S2 are independant conditionnally to D and not D (it is
possible not to make this assumption, but you have then to give a
value for the dependence).
A2: P(D)=P(not D) = 0.5 (this is for simplyfing the computations, but
the solution is easy to compute also if you give another value for
P(D)).
c) now the solution is straihtforward:
let's denote OR(12) = P(D/S1,S2)/P(not D/S1,S2),
and OR(i)=P(D/Si)/P(not D/Si), i=1,2
we have by simple Bayes's formula and with assumptions A1 and A2:
OR(1,2)= OR(1) OR(2)
and p12 = OR(1) OR(2) / (1 + OR(1) OR(2) )
I don't know if it is your answer, since it is a very simple one...
I think it is the solution based on the simplest probabilistic computations.
sincerely yours
Jean-louis
Quoting "Lehner, Paul E." <pleh...@mitre.org>:
Austin, Jean-Lous, Konrad, Peter
Thank you for your responses. They are very helpful.
Your consensus view seems to be that when receiving evidence in the
form of a single calibrated judgment, one should not update personal
judgments by using Bayes rule. This seems incoherent (from a
strict Bayesian perspective) unless perhaps one explicitly
represents the overlap of knowledge with the source of the
calibrated judgment (which may not be practical.)
Unfortunately this is the conclusion I was afraid we would reach,
because it leads me to be concerned that I have been giving some bad
advice about applying Bayesian reasoning to some very practical
problems.
Here is a simple example.
Analyst A is trying to determine whether X is at location Y. She
has two principal evidence items. The first is a report from a
spectral analyst that concludes "based on the match to the expected
spectral signature I conclude with high confidence that X is at
location Y". The second evidence is a report from a chemical
analyst who asserts, "based on the expected chemical composition
that is typically associated with X, I conclude with moderate
confidence that X is at location Y." How should analyst A approach
her analysis?
Previously I would have suggested something like this. "Consider
each evidence item in turn. Assume that X is at location Y. What
are the chances that you would receive a 'high confidence' report
from the spectral analyst, ... a report of 'moderate confidence'
from the chemical analyst. Now assume X is not a location Y, ...."
In other words I would have lead the analyst toward some simple
instantiation of Bayes inference.
But clearly the spectral and chemical analyst are using more than
just the sensor data to make their confidence assessments. In part
they are using the same background knowledge that Analyst A has.
Furthermore both the spectral and chemical analysts are good at
their job, their confidence judgments are reasonably calibrated.
This is just like the TWC problem only more complex.
So if Bayesian inference is inappropriate for the TWC problem, is it
also inappropriate here? Is my advice bad?
Paul
From: uai-boun...@engr.orst.edu [mailto:uai-boun...@engr.orst.edu]
On Behalf Of Lehner, Paul E.
Sent: Monday, February 16, 2009 11:40 AM
To: uai@ENGR.ORST.EDU
Subject: Re: [UAI] A perplexing problem - Version 2
UAI members
Thank you for your many responses. You've provided at least 5
distinct answers which I summarize below.
(Answer 5 below is clearly correct, but leads me to a new quandary.)
Answer 1: "70% chance of snow" is just a label and conceptually
should be treated as "XYZ". In other words don't be fooled by the
semantics inside the quotes.
My response: Technically correct, but intuitively unappealing.
Although I often council people on how often intuition is
misleading, I just couldn't ignore my intuition on this one.
Answer 2: The forecast "70% chance of snow is ill-defined"
My response: I agree, but in this case I was more concerned about
the conflict between math and intuition. I would be willing to
accept any well-defined forecasting statement.
Answer 3: The reference set "winter days" is the wrong reference set.
My response: I was just trying to give some justification to my
subjective prior. But this answer does point out a distinction
between base rates and subjective priors. This distinction relates
to my new quandary below so please read on.
Answer 4: The problem inherently requires more variables and cannot
be treated as a simple single evidence with two hypotheses problem.
My response: Actually I was concerned that this was the answer. As
it may have implied that using Bayes to evaluate a single evidence
item was impractical for the community of analysts I'm working with.
Fortunately ...
Answer 5: The problem statement was inherently incoherent. Many of
you pointed out that if TWC predicts "70% snow" on 10% of the days
that it snows and on 1% of days that it does not snow, and a 5% base
rate for snow, then the P("70% snow" & snow) is .005 and P("70%
snow" & ~snow) = .0095. So for the days that TWC says "70% snow" it
actually snows on a little over 34% of the days. Clearly my
assertion that TWC is calibrated is incoherent relative to the rest
of the problem statement. The problem was not underspecified, it
was over specified. (I hope I did the math correctly.)
My response: Thanks for pointing this out. I'm embarrassed that I
didn't notice this myself. Though this clearly solves my initial
concern it leads me to an entirely new quandary.
Consider the following revised version.
The TWC problem
1. Question: What is the chance that it will snow next Monday?
2. My subjective prior: 5%
3. Evidence: The Weather Channel (TWC) says there is a "70%
chance of snow" on Monday.
4. TWC forecasts of snow are calibrated.
Notice that I did not justify by subjective prior with a base rate.
From P(S)=.05 and P(S|"70%") = .7 I can deduce that
P("70%"|S)/P("70%"|~S) = 44.33. So now I can "deduce" from my prior
and evidence odds that P(S|"70%") = .7. But this seems silly.
Suppose my subjective prior was 20%. Then P("70%"|S)/P("70%"|~S) =
9.33333 and again I can "deduce" P(S|"70%")=.7.
My latest quandary is that it seems odd that my subjective
conditional probability of the evidence should depend on my
subjective prior. This may be coherent, but is too counter
intuitive for me to easily accept. It would also suggest that when
receiving a single evidence item in the form of a judgment from a
calibrated source, my posterior belief does not depend on my prior
belief. In effect, when forecasting snow, one should ignore priors
and listen to The Weather Channel.
Is this correct? If so, does this bother anyone else?
paull
From: uai-boun...@engr.orst.edu [mailto:uai-boun...@engr.orst.edu]
On Behalf Of Lehner, Paul E.
Sent: Friday, February 13, 2009 4:29 PM
To: uai@ENGR.ORST.EDU
Subject: [UAI] A perplexing problem
I was working on a set of instructions to teach simple
two-hypothesis/one-evidence Bayesian updating. I came across a
problem that perplexed me. This can't be a new problem so I'm
hoping someone will clear things up for me.
The problem
5. Question: What is the chance that it will snow next Monday?
6. My prior: 5% (because it typically snows about 5% of the
days during the winter)
7. Evidence: The Weather Channel (TWC) says there is a "70%
chance of snow" on Monday.
8. TWC forecasts of snow are calibrated.
My initial answer is to claim that this problem is underspecified. So I add
9. On winter days that it snows, TWC forecasts "70% chance of
snow" about 10% of the time
10. On winter days that it does not snow, TWC forecasts "70%
chance of snow" about 1% of the time.
So now from P(S)=.05; P("70%"|S)=.10; and P("70%"|S)=.01 I apply
Bayes rule and deduce my posterior probability to be P(S|"70%") =
.3448.
Now it seems particularly odd that I would conclude there is only a
34% chance of snow when TWC says there is a 70% chance. TWC knows
so much more about weather forecasting than I do.
What am I doing wrong?
Paul E. Lehner, Ph.D.
Consulting Scientist
The MITRE Corporation
(703) 983-7968
pleh...@mitre.org<mailto:pleh...@mitre.org>
_______________________________________________
uai mailing list
uai@ENGR.ORST.EDU
https://secure.engr.oregonstate.edu/mailman/listinfo/uai