Probabilistic Reasoning & Uncertainty for Artificial Intelligence (CIS 479)
Probabilistic Reasoning
Why use it?
Ad Hoc Uncertainty (MYCIN)
Questions?
Certainty Factors (ad-hoc)
how certain are you that the rule will fire?
Probability
P(E) = Number of desired outcomes / Total number of outcomes = | event | / |sample space|
P(not E) = 1 – P(E)
P(A or B) = P(A) + P(B) – P(A Ç B) P(A Ç B) = f ( a and b are mutually exclusive)
P(A Ç B) = P(A) * P(B|A) = P(B) * P(A|B)
For independent events
P(B|A) = P(B)
P(A|B) = P(A)
Bayes Theorem (Non-Independent)
P(Hi |E) º Probability Hi is true given evidence E
P(E | Hi) º Probability evidence E is observed given Hi is true.
P(Hi) = Hi true regardless of evidence
P(Hi |E) = P(E | Hi) * P(Hi) / S P(E | Hk) * P(Hk)
Tanimoto Example on Rain in Seattle
P(H) = 0.8
P(E|H) = 0.2
P(E | ~H) = 0.025
tilde (~) = not
Probability of rain tomorrow with geese on the lake
P(H | E) = (P( E | H) * P(H)) / P(E) = (0.016 / 0.021) = 0.7619
P(E) = P(E | H) * P(H) + P(E | ~H) +P(~H) = (0.02)(0.8) + (0.025)*(0.2) = (0.016) + (0.005) = 0.021
P(~H | E) = ?
P( ~H | E) = (P(E | ~H)P(~H)) / P(E) = (0.005 / 0.021) = 0.2381
Weakness of Bayes Theorem
Probabilistic Inference Nets
Criteria for Problem Characteristics
Fuzzy Set Theory
X Î S This is not longer true
X Ï S Either in or out for most set theory
Question?
What constitutes a big pile of chalk?
If we have three pieces of chalk in the room is that considered a big pile of chalk? Some people might say, yes that is a big pile and some would not. Someplace between those three pieces of chalk and a whole room full of chalk the pile of chalk turns from a small pile into a big pile. This could be a different spot for different people.
F:[0,1]n ® [0,1]
{0,1}
Tanimoto has examples of independent events and dependent events
Fuzzy Inference Rules
Possiblistic Dependent |
Probabilistic Independent |
|
A |
a |
a |
B |
b |
b |
~A |
1-a |
1-a |
A Ù B |
min(a,b) |
a*b |
A n B |
max(a,b) |
a+b – a*b |
A ® B |
max(1-a,b) |
(1-a)+a*b |
A Å B |
max(min(a,1-b)),min(1-a,b)) |
a+b-2ab+a2b+ab2-a2b2 |
P (X ® (X n Z))
P(X) = 0.5 Possibilistic
P(X) = 0.1 P(Y n Z) = max (P(Y),P(Z))
P(Z) = 0.2 P(Y n Z) = max(0.1, 0.2) = 0.2
P(X ® P(Y n Z)) = max(1-P(X), P(Y n Z))
= max(1-(5), 0.2)
= 5
Probabilistic
P(Y n Z) = P(Y) + P(Z) – P(Y)P(2) = (0.1 + 0.2) – ((0.1)(0.2)) = (0.3 – 0.2) = 0.28
P(X ® X n Z) = 1 – P(X) + P(X)P(Y n Z) = (1 – 0.5) + ((0.5)(0.28)) = 0.5 + 0.14 = 64
Probabilistic Reasoning
S1: Clanking Sound
S2: Low pickup
S3: Starting problem
S4: Parts are hard to find
C1: Repair Estimate > $250
H1: Thrown connecting rod
H2: Wrist Pin Loose
H3: Car Out of Tune
P(H: | S)
Second Level
H4: Replace or Rebuild Engine
H5: Tune Engine
Steps for Building Inferences
Dempster / Schaffer Theorem of Evidence
X,Y,Z
{ X } { Y } { Z }
{ X,Y } { X,Z } { Y,Z } ® [0,1]
{ X,Y,Z } { }
This all needs to add up to 1
as more evidence is acquired views might change
Belief
B(A) + B(~A) ¹ 1
A = total exposed spots is 7
6 of 36 elements from m
m
Î P(m ) ® [0,1]m(f ) = 0
S
m (x) = 1m
³ Am
³ FEvidence
m(F) = 1 ® Evidence is certain
F Ç A ¹ 0
Belief (A) = S m(B), A ³ B
Doubt in (A) = Belief (> A)
Plausibility (A) = 1- Doubt(A)
Belief (f ) = 0 Plausibility(f ) = 0
Belief(m ) = 1 Plausibility(m ) = 1
Plausibility (A) ³ Belief (A)
1 ³ Belief(A) + Belief(~A)
Plausibility(A) + Plausibility (~A) ³ 1
If B ³ A then
Belief (B) ³ Belief A
Plausibility (B) ³ Plausibility (A)
Rule of combination for uncertainty evidence – Orthogonal
A ¹ f
if true
[m1 + m2] A = S xny = A M1(x)M2(y)
1 - S xny = f M1(x)M2(y)
if A ¹ f then [M1 + M2] f = 0
Numeric Example
S = Snow
R = Rain
D = Dry
m
= {S, R, D}|P(m )| = 8
p. 275 Tanimoto
Two Pieces of evidence
f |
{S} |
{R} |
{D} |
{S,R} |
{S,D} |
{R,D} |
{S,R,D} |
|
Mfreeze |
0 |
0.2 |
0.1 |
0.1 |
0.2 |
0.1 |
0.1 |
0.2 |
Mstorm |
0 |
0.1 |
0.2 |
0.1 |
0.3 |
0.1 |
0.1 |
0.1 |
Mboth |
0 |
0.282 |
0.282 |
0.128 |
0.18 |
0.51 |
0.51 |
0.026 |
All values in the chart should add up to one
The belief ({S,R}) = Mboth({S,R}) + Mboth({S}) + Mboth({R}) = 0.744