1 slide slide probability is conditional. theorems of increase of probabilities. theorems of...
Post on 13-Jan-2016
225 Views
Preview:
TRANSCRIPT
11 11 SlideSlide
SlideSlide
Probability is conditional. Theorems of increase of
probabilities. Theorems of addition of probabilities.
22 22 SlideSlide
SlideSlide
STATISTICSSTATISTICS in in PRACTICEPRACTICE
• Carstab Corporation, a subsidiary of Morton International, produces specialty chemicals and
offers a variety of chemicals designed to meet the unique specifications of its customers.• Carstab’s customer agreed to test each lot after receiving it and determine whether the catalyst
would perform the desired function.• Each Carstab shipment to the customer had a .60
probability of being accepted and a .40 probability of
being returned.
33 33 SlideSlide
SlideSlide
Chapter 4Chapter 4 Introduction to Probability Introduction to Probability
4.1 Experiments, Counting Rules,
and Assigning Probabilities4.2 Events and Their Probability4.3 Some Basic Relationships
of Probability4.4 Conditional Probability4.5 Bayes’ Theorem
44 44 SlideSlide
SlideSlide
ProbabilityProbability• Managers often base their decisions on an
analysis of uncertainties such as the following:
1. What are the chances that sales will decrease if we increase prices?
2. What is the likelihood a new assembly method will increase productivity?
3. How likely is it that the project will be finished on time?
4. What is the chance that a new investment will be profitable?
55 55 SlideSlide
SlideSlide
Probability as a Numerical Probability as a Numerical MeasureMeasure
of the Likelihood of of the Likelihood of OccurrenceOccurrence
0 11.5
Increasing Likelihood of Occurrence
Probability:
The eventis very
unlikelyto occur.
The occurrenceof the event is
just as likely asit is unlikely.
The eventis almostcertain
to occur.
66 66 SlideSlide
SlideSlide
4.1 Experiments, Counting 4.1 Experiments, Counting Rules, Rules,
and Assigning Probabilities and Assigning Probabilities• Example: Experiment and Experimental
Outcomes
77 77 SlideSlide
SlideSlide
4.1 Experiments, Counting 4.1 Experiments, Counting Rules, Rules,
and Assigning Probabilities and Assigning Probabilities An experiment is any process that generates well-defined outcomes. An experiment is any process that generates well-defined outcomes.
The sample space for an experiment is the set of all experimental outcomes. The sample space for an experiment is the set of all experimental outcomes.
An experimental outcome is also called a sample point. An experimental outcome is also called a sample point.
88 88 SlideSlide
SlideSlide
An Experiment and Its Sample An Experiment and Its Sample SpaceSpace
• Example:
1. tossing a coin — the sample space
S = {Head, Tail}
2. selecting a part for inspection — the sample
space S = {Defective, Nondefective}
3. rolling a die — the sample space
S = {1, 2, 3, 4, 5, 6}
99 99 SlideSlide
SlideSlide
Example: Bradley InvestmentsExample: Bradley Investments
Bradley has invested in two stocks, Markley Oil and Collins Mining. Bradley has determined that thepossible outcomes of these investments three monthsfrom now are as follows. Investment Gain or LossInvestment Gain or Loss
in 3 Months (in $000)in 3 Months (in $000)
Markley OilMarkley Oil Collins MiningCollins Mining
1010 55 00-20-20
88-2-2
1010 1010 SlideSlide
SlideSlide
Counting Rules, Counting Rules, Combinations, Combinations,
and Permutationsand Permutations• If an experiment consists of a sequence of k steps in which there are n1 possible results for the first
step, n2 possible results for the second step,
and so on, then the total number of experimental outcomes is given by (n1)(n2) . . . (nk).
• A helpful graphical representation of a
multiple-step experiment is a tree diagram.
A Counting Rule for Multiple-Step Experiments
1111 1111 SlideSlide
SlideSlide
Counting Rules, Counting Rules, Combinations, Combinations,
and Permutationsand Permutations• Example: Tree Diagram for the Experiment of
Tossing Two Coins
1212 1212 SlideSlide
SlideSlide
Bradley Investments can be viewed as atwo-step experiment. It involves two stocks, each with a set of experimental outcomes.
Markley Oil: n1 = 4
Collins Mining: n2 = 2Total Number of
Experimental Outcomes: n1n2 = (4)(2) = 8
A Counting Rule for A Counting Rule for Multiple-Step ExperimentsMultiple-Step Experiments
1313 1313 SlideSlide
SlideSlide
Tree DiagramTree Diagram
Gain 5Gain 5
Gain 8Gain 8
Gain 8Gain 8
Gain 10Gain 10
Gain 8Gain 8
Gain 8Gain 8
Lose 20Lose 20
Lose 2Lose 2
Lose 2Lose 2
Lose 2Lose 2
Lose 2Lose 2
EvenEven
Markley Oil(Stage 1)
Collins Mining(Stage 2)
ExperimentalOutcomes
(10, 8) Gain $18,000
(10, -2) (10, -2) Gain $8,000 Gain $8,000
(5, 8) Gain $13,000
(5, -2) Gain $3,000
(0, 8) (0, 8) Gain $8,000 Gain $8,000
(0, -2) Lose $2,000
(-20, 8) Lose $12,000
(-20, -2) Lose $22,000
1414 1414 SlideSlide
SlideSlide
A second useful counting rule enables us to count the number of experimental outcomes when n objects are to be selected from a set of N objects.
Counting Rule for CombinationsCounting Rule for Combinations
CN
nN
n N nnN
!
!( )!C
N
nN
n N nnN
!
!( )!
Number of Combinations of N Objects Taken n at a Time
where: N! = N(N - 1)(N - 2) . . . (2)(1) n! = n(n - 1)(n - 2) . . . (2)(1) 0! = 1
1515 1515 SlideSlide
SlideSlide
Counting Rule for Counting Rule for CombinationsCombinations• Example:
1.
2. Lottery system uses the random selection of six integers from a group of 47 to determine the weekly lottery winner. The number of ways six different integers can be selected from a group of 47.
1012
120
)1)(2)(3)(1)(2(
)1)(2)(3)(4)(5(
)!25(!2
!5
2
552
C
573,737,10)1)(2)(3)(4)(5)(6(
)42)(43)(44)(45)(46)(47(
!41!6
!47
)!647(!6
!47
6
47
1616 1616 SlideSlide
SlideSlide
Number of Permutations of N Objects Taken n at a Time
where: N! = N(N - 1)(N - 2) . . . (2)(1) n! = n(n - 1)(n - 2) . . . (2)(1) 0! = 1
P nN
nN
N nnN
!!
( )!P n
N
nN
N nnN
!!
( )!
Counting Rule for PermutationsCounting Rule for Permutations
A third useful counting rule enables us to
count the number of experimental outcomes
when n objects are to be selected from a set of
N objects, where the order of selection is
important.
1717 1717 SlideSlide
SlideSlide
Assigning ProbabilitiesAssigning Probabilities• Two basic requirements for assigning probabilities
1. The probability assigned to each experimental outcome must be between 0 and 1, inclusively. If we let Ei denote the ith experimental outcome and P(Ei) its probability, then this requirement can be written as
0 P(Ei) 1 for all I
2. The sum of the probabilities for all the experimental outcomes must equal 1.0. For n experimental outcomes, this requirement can be written as
P(E1)+ P(E2)+… + P(En) =1
1818 1818 SlideSlide
SlideSlide
Assigning ProbabilitiesAssigning Probabilities
Classical Method
Relative Frequency Method
Subjective Method
Assigning probabilities based on the assumption of equally likely outcomes
Assigning probabilities based on experimentation or historical data
Assigning probabilities based on judgment
1919 1919 SlideSlide
SlideSlide
Classical MethodClassical Method If an experiment has n possible outcomes, this method would assign a probability of 1/n to each outcome.
Experiment: Rolling a die
Sample Space: S = {1, 2, 3, 4, 5, 6}
Probabilities: Each sample point has a 1/6 chance of occurring
Example
2020 2020 SlideSlide
SlideSlide
Relative Frequency MethodRelative Frequency Method
Number ofPolishers Rented
Numberof Days
01234
4 61810 2
Lucas Tool Rental would like to assign
probabilities to the number of car polishers
it rents each day. Office records show the following
frequencies of daily rentals for the last 40 days.
• Example: Lucas Tool Rental
2121 2121 SlideSlide
SlideSlide
Each probability assignment is given by dividing
the frequency (number of days) by the total frequency
(total number of days).
Relative Frequency MethodRelative Frequency Method
4/404/40
ProbabilityNumber of
Polishers RentedNumberof Days
01234
4 61810 240
.10 .15 .45 .25 .051.00
2222 2222 SlideSlide
SlideSlide
Subjective MethodSubjective Method• When economic conditions and a company’s circumstances change rapidly it might be inappropriate to assign probabilities based solely on historical data.
• We can use any data available as well as our experience and intuition, but ultimately a probability value should express our degree of belief that the experimental outcome will occur.
• The best probability estimates often are obtained by combining the estimates from the classical or relative frequency approach with the subjective estimate.
2323 2323 SlideSlide
SlideSlide
Subjective MethodSubjective MethodApplying the subjective method, an analyst
made the following probability assignments.
Exper. Outcome Net Gain Net Gain oror Loss Loss Probability(10, 8)(10, -2)(5, 8)(5, -2)(0, 8)(0, -2)(-20, 8)(-20, -2)
$18,000 Gain $8,000 Gain $13,000 Gain $3,000 Gain $8,000 Gain $2,000 Loss $12,000 Loss $22,000 Loss
.20
.08
.16
.26
.10
.12
.02
.06
2424 2424 SlideSlide
SlideSlide
Exercises Exercises
• Consider the experiment of tossing a coin three times– Develop a tree diagram– List the experimental outcomes– What is the probability for each experimental
outcomes?• A decision maker subjectively assigned the following
probabilities to the four outcomes of an experiment: Pr(E1)=0.10, Pr(E2)=0.15, Pr(E3)=0.40 and Pr(E4)=0.20. Are these probability assignments valid?
2525 2525 SlideSlide
SlideSlide
An event is a collection of sample points. An event is a collection of sample points.
The probability of any event is equal to the sum of the probabilities of the sample points in the event. The probability of any event is equal to the sum of the probabilities of the sample points in the event.
If we can identify all the sample points of an experiment and assign a probability to each, we can compute the probability of an event.
If we can identify all the sample points of an experiment and assign a probability to each, we can compute the probability of an event.
4.2 Events and Their Probabilities4.2 Events and Their Probabilities
2626 2626 SlideSlide
SlideSlide
Events and Their ProbabilitiesEvents and Their ProbabilitiesEvent M = Markley Oil Profitable
M = {(10, 8), (10, -2), (5, 8), (5, -2)}
P(M) = P(10, 8) + P(10, -2) + P(5, 8) + P(5, -2)
= .20 + .08 + .16 + .26
= .70
2727 2727 SlideSlide
SlideSlide
Events and Their ProbabilitiesEvents and Their Probabilities
Event C = Collins Mining Profitable
C = {(10, 8), (5, 8), (0, 8), (-20, 8)}
P(C) = P(10, 8) + P(5, 8) + P(0, 8) + P(-20, 8)
= .20 + .16 + .10 + .02
= .48
2828 2828 SlideSlide
SlideSlide
CommentsComments
• The sample space, S, is an event. Pr(S)=1.
• When the classical method is used to assign probabilities, the assumption is that the experimental outcomes are equally likely.
2929 2929 SlideSlide
SlideSlide
ExercisesExercises
• Consider the experiment of selecting a playing card from a deck of 52 playing cards. Each card corresponds to a sample point with a 1/52 probability.– List the sample points in the event an ace is selected– List the sample points in the event a club is selected– List the sample points in the event a face card is
selected– Find the probabilities associated with each of the
events above
3030 3030 SlideSlide
SlideSlide
ExercisesExercises
• Suppose that a manager of a large apartment complex provides the following subjective probability estimates about the number of vacancies that will exist next month.
Provide the probability of each of the following events a) no vacancies b) at least four vacancies and c) two or fewer vacancies
Vacancies
0 1 2 3 4 5
Probability
.05 .015 .35 .25 .10 .10
3131 3131 SlideSlide
SlideSlide
3 3 Some Basic Relationships of ProbabilitySome Basic Relationships of Probability
There are some basic probability relationships that can be used to compute the probability of an event without knowledge of all the sample point probabilities.
Complement of an Event Complement of an Event
Intersection of Two Events Intersection of Two Events
Mutually Exclusive Events Mutually Exclusive Events
Union of Two EventsUnion of Two Events
3232 3232 SlideSlide
SlideSlide
The complement of A is denoted by Ac. The complement of A is denoted by Ac.
The complement of event A is defined to be the event consisting of all sample points that are not in A. The complement of event A is defined to be the event consisting of all sample points that are not in A.
Complement of an EventComplement of an Event
Event Event AA AAcc
SampleSpace SSampleSpace S
VennDiagram
3333 3333 SlideSlide
SlideSlide
The union of events The union of events AA and and BB is denoted by is denoted by AA BB The union of events The union of events AA and and BB is denoted by is denoted by AA BB
The The unionunion of events of events AA and and BB is the event containing is the event containing all sample points that are in all sample points that are in A A oror B B or both.or both. The The unionunion of events of events AA and and BB is the event containing is the event containing all sample points that are in all sample points that are in A A oror B B or both.or both.
Union of Two EventsUnion of Two Events
SampleSpace SSampleSpace SEvent Event AA Event B
3434 3434 SlideSlide
SlideSlide
Union of Two EventsUnion of Two Events
Event M = Markley Oil Profitable
Event C = Collins Mining Profitable
M C = Markley Oil Profitable or Collins Mining Profitable
M C = {(10, 8), (10, -2), (5, 8), (5, -2), (0, 8), (-20, 8)}
P(M C) = P(10, 8) + P(10, -2) + P(5, 8) + P(5, -2)
+ P(0, 8) + P(-20, 8)
= .20 + .08 + .16 + .26 + .10 + .02
= .82
3535 3535 SlideSlide
SlideSlide
The intersection of events A and B is denoted by A The intersection of events A and B is denoted by A
The intersection of events A and B is the set of all sample points that are in both A and B. The intersection of events A and B is the set of all sample points that are in both A and B.
SampleSpace SSampleSpace SEvent Event AA Event Event BB
Intersection of Two EventsIntersection of Two Events
Intersection of A and BIntersection of A and B
3636 3636 SlideSlide
SlideSlide
Intersection of Two EventsIntersection of Two Events
Event M = Markley Oil Profitable
Event C = Collins Mining Profitable
M C = Markley Oil Profitable and Collins Mining Profitable
M C = {(10, 8), (5, 8)}
P(M C) = P(10, 8) + P(5, 8)
= .20 + .16
= .36
3737 3737 SlideSlide
SlideSlide
The addition law provides a way to compute the probability of event A, or B, or both A and B occurring. The addition law provides a way to compute the probability of event A, or B, or both A and B occurring.
Addition LawAddition Law
The law is written as: The law is written as:
P(A B) = P(A) + P(B) - P(A B
3838 3838 SlideSlide
SlideSlide
Event M = Markley Oil ProfitableEvent C = Collins Mining Profitable
M C = Markley Oil Profitable or Collins Mining Profitable
We know: P(M) = .70, P(C) = .48, P(M C) = .36
Thus: P(M C) = P(M) + P(C) - P(M C)
= .70 + .48 - .36
= .82
Addition LawAddition Law
(This result is the same as that obtained earlierusing the definition of the probability of an event.)
3939 3939 SlideSlide
SlideSlide
Mutually Exclusive EventsMutually Exclusive Events
Two events are said to be mutually exclusive if the events have no sample points in common.
Two events are said to be mutually exclusive if the events have no sample points in common.
Two events are mutually exclusive if, when one event occurs, the other cannot occur. Two events are mutually exclusive if, when one event occurs, the other cannot occur.
SampleSpace SSampleSpace SEvent Event AA Event B
4040 4040 SlideSlide
SlideSlide
Mutually Exclusive EventsMutually Exclusive Events
If events A and B are mutually exclusive, P(A B = 0. If events A and B are mutually exclusive, P(A B = 0.
The addition law for mutually exclusive events is: The addition law for mutually exclusive events is:
P(A B) = P(A) + P(B)
there’s no need toinclude “- P(A B”
4141 4141 SlideSlide
SlideSlide
The probability of an event given that another event has occurred is called a conditional probability. The probability of an event given that another event has occurred is called a conditional probability.
A conditional probability is computed as follows : A conditional probability is computed as follows :
The conditional probability of A given B is denoted by P(A|B). The conditional probability of A given B is denoted by P(A|B).
4.4 Conditional Probability4.4 Conditional Probability
( )( | )
( )P A B
P A BP B
( )( | )
( )P A B
P A BP B
4242 4242 SlideSlide
SlideSlide
Joint Probabilities, Marginal ProbabilitiesJoint Probabilities, Marginal Probabilities • Promotion Status of Police Officers Over the Past Two
Years
• Joint Probability Tables for Promotions
4343 4343 SlideSlide
SlideSlide
Event M = Markley Oil Profitable
Event C = Collins Mining Profitable
We know: P(M C) = .36, P(M) = .70
Thus:
Conditional ProbabilityConditional Probability
( ) .36( | ) .5143
( ) .70P C M
P C MP M
( ) .36( | ) .5143
( ) .70P C M
P C MP M
= Collins Mining Profitable given Markley Oil Profitable
( | )P C M( | )P C M
4444 4444 SlideSlide
SlideSlide
Conditional ProbabilityConditional Probability• Conditional Probability
( )
( | )( )
P A BP A B
P B
( )
( | )( )
P A BP A B
P B
4545 4545 SlideSlide
SlideSlide
Independent EventsIndependent Events
If the probability of event A is not changed by the existence of event B, we would say that events A and B are independent.
If the probability of event A is not changed by the existence of event B, we would say that events A and B are independent.
Two events A and B are independent if: Two events A and B are independent if:
P(A|B) = P(A) P(B|A) = P(B)or
4646 4646 SlideSlide
SlideSlide
Multiplication LawMultiplication Law
The multiplication law provides a way to compute the probability of the intersection of two events. The multiplication law provides a way to compute the probability of the intersection of two events.
The law is written as: The law is written as:
P(A B) = P(B)P(A|B)
4747 4747 SlideSlide
SlideSlide
Event M = Markley Oil ProfitableEvent C = Collins Mining Profitable
We know: P(M) = .70, P(C|M) = .5143
Multiplication LawMultiplication Law
M C = Markley Oil Profitable and Collins Mining Profitable
Thus: P(M C) = P(M)P(M|C)= (.70)(.5143)
= .36
(This result is the same as that obtained earlierusing the definition of the probability of an event.)
4848 4848 SlideSlide
SlideSlide
The multiplication law also can be used as a test to see if two events are independent. The multiplication law also can be used as a test to see if two events are independent.
The law is written as: The law is written as:
P(A B) = P(A)P(B)
Multiplication LawMultiplication Lawfor Independent Eventsfor Independent Events
4949 4949 SlideSlide
SlideSlide
Multiplication LawMultiplication Lawfor Independent Eventsfor Independent Events
Event M = Markley Oil ProfitableEvent C = Collins Mining Profitable
We know: P(M C) = .36, P(M) = .70, P(C) = .48 But: P(M)P(C) = (.70)(.48) = .34, not .36
Are events M and C independent?DoesP(M C) = P(M)P(C) ?
Hence: M and C are not independent.
5050 5050 SlideSlide
SlideSlide
5 Bayes5 Bayes’’ Theorem Theorem
NewNewInformationInformation
NewNewInformationInformation
ApplicationApplicationof Bayes’of Bayes’TheoremTheorem
ApplicationApplicationof Bayes’of Bayes’TheoremTheorem
PosteriorPosteriorProbabilitiesProbabilities
PosteriorPosteriorProbabilitiesProbabilities
PriorPriorProbabilitiesProbabilities
PriorPriorProbabilitiesProbabilities
• Often we begin probability analysis with initial or prior probabilities.• Then, from a sample, special report, or a product test we obtain some additional information.
• Given this information, we calculate revised or posterior probabilities.
• Bayes’ theorem provides the means for revising the prior probabilities.
5151 5151 SlideSlide
SlideSlide
A proposed shopping center will provide strong competition for downtown businesses like
L. S. Clothiers. If the shopping
center is built, the owner of
L. S. Clothiers feels it would be best to
relocate to the center.
Bayes’ TheoremBayes’ Theorem
• Example: L. S. Clothiers
The shopping center cannot be built unless a
zoning change is approved by the town council. Theplanning board must first make a recommendation, foror against the zoning change, to the council.
5252 5252 SlideSlide
SlideSlide
• Prior Probabilities
Let:
Bayes’ TheoremBayes’ Theorem
A1 = town council approves the zoning change
A2 = town council disapproves the change
P(A1) = .7, P(A2) = .3
Using subjective judgment:
5353 5353 SlideSlide
SlideSlide
• New Information
The planning board has recommended against the zoning change. Let B denote the event of a negative recommendation by the planning board.
Given that B has occurred, should L. S. Clothiers revise the probabilities that the town council will approve or disapprove the zoning change?
BayesBayes’’ Theorem Theorem
5454 5454 SlideSlide
SlideSlide
• Conditional Probabilities
Past history with the planning board and the
town council indicates the following:
Bayes’ TheoremBayes’ Theorem
P(B|A1) = .2 P(B|A2) = .9
P(BC|A1) = .8 P(BC|A2) = .1
Hence:
5555 5555 SlideSlide
SlideSlide
P(Bc|A1) = .8P(Bc|A1) = .8P(A1) = .7P(A1) = .7
P(A2) = .3P(A2) = .3
P(B|A2) = .9P(B|A2) = .9
P(Bc|A2) = .1P(Bc|A2) = .1
P(B|A1) = .2P(B|A1) = .2 P(A1 B) = .14P(A1 B) = .14
P(A2 B) = .27P(A2 B) = .27
P(A2 Bc) = .03P(A2 Bc) = .03
P(A1 Bc) = .56P(A1 Bc) = .56
Bayes’ TheoremBayes’ TheoremTree Diagram
Town Council Planning Board ExperimentalOutcomes
5656 5656 SlideSlide
SlideSlide
Bayes’ TheoremBayes’ Theorem
1 1 2 2
( ) ( | )( | )
( ) ( | ) ( ) ( | ) ... ( ) ( | )i i
in n
P A P B AP A B
P A P B A P A P B A P A P B A
1 1 2 2
( ) ( | )( | )
( ) ( | ) ( ) ( | ) ... ( ) ( | )i i
in n
P A P B AP A B
P A P B A P A P B A P A P B A
• To find the posterior probability that event Ai will occur given that event B has occurred, we apply Bayes’ theorem.
• Bayes’ theorem is applicable when the events for which we want to compute posterior probabilities are mutually exclusive and their union is the entire sample space.
5757 5757 SlideSlide
SlideSlide
• Posterior Probabilities
Given the planning board’s recommendation not to approve the zoning change, we revise the prior probabilities as follows:
1 11
1 1 2 2
( ) ( | )( | )
( ) ( | ) ( ) ( | )P A P B A
P A BP A P B A P A P B A
1 11
1 1 2 2
( ) ( | )( | )
( ) ( | ) ( ) ( | )P A P B A
P A BP A P B A P A P B A
(. )(. )(. )(. ) (. )(. )
7 27 2 3 9
(. )(. )(. )(. ) (. )(. )
7 27 2 3 9
Bayes’ TheoremBayes’ Theorem
= .34= .34
5858 5858 SlideSlide
SlideSlide
Conclusion
The planning board’s recommendation is good news for L. S. Clothiers. The posterior probability of the town council approving the zoning change is .34 compared to a prior probability of .70.
Bayes’ TheoremBayes’ Theorem
5959 5959 SlideSlide
SlideSlide
Tabular ApproachTabular Approach
• Step 1 Prepare the following three columns:
Column 1 - The mutually exclusive events for which posterior probabilities are desired.
Column 2 - The prior probabilities for the events.
Column 3 - The conditional probabilities of the new information given each event.
6060 6060 SlideSlide
SlideSlide
Tabular ApproachTabular Approach
(1) (2) (3) (4) (5)
Events
Ai
PriorProbabilities
P(Ai)
ConditionalProbabilities
P(B|Ai)
A1
A2
.7
.3
1.0
.2
.9
6161 6161 SlideSlide
SlideSlide
Tabular ApproachTabular Approach
• Step 2
Column 4
Compute the joint probabilities for each event and the new information B by using the multiplication law.
Multiply the prior probabilities in column 2 by the corresponding conditional probabilities in column 3. That is, P(Ai B) = P(Ai) P(B|Ai).
6262 6262 SlideSlide
SlideSlide
Tabular ApproachTabular Approach
(1) (2) (3) (4) (5)
Events
Ai
PriorProbabilities
P(Ai)
ConditionalProbabilities
P(B|Ai)
A1
A2
.7
.3
1.0
.2
.9
.14
.27
JointProbabilities
P(Ai ∩ B)
.7 x .2.7 x .2
6363 6363 SlideSlide
SlideSlide
Tabular ApproachTabular Approach
• Step 2 (continued)
We see that there is a .14 probability of the town council approving the zoning change and a negative recommendation by the planning board. There is a .27 probability of the town Council disapproving the zoning change and a negative recommendation by the planning board.
6464 6464 SlideSlide
SlideSlide
Tabular ApproachTabular Approach
• Step 3
Column 4 Sum the joint probabilities. The sum is The probability of the new information, P(B). The sum .14 + .27 shows an overall probability of .41 of a negative recommendation by the planning board.
6565 6565 SlideSlide
SlideSlide
Tabular ApproachTabular Approach
(1) (2) (3) (4) (5)
Events
Ai
PriorProbabilities
P(Ai)
ConditionalProbabilities
P(B|Ai)
A1
A2
.7
.3
1.0
.2
.9
.14
.27
JointProbabilities
P(Ai B)
P(B) = .41
6666 6666 SlideSlide
SlideSlide
• Step 4 Column 5
Compute the posterior probabilities using the basic relationship of conditional probability.
The joint probabilities P(Ai B) are in column 4 and the probability P(B) is the sum of column 4.
Tabular ApproachTabular Approach
)(
)()|(
BP
BAPBAP i
i
)(
)()|(
BP
BAPBAP i
i
6767 6767 SlideSlide
SlideSlide
(1) (2) (3) (4) (5)
Events
Ai
PriorProbabilities
P(Ai)
ConditionalProbabilities
P(B|Ai)
A1
A2
.7
.3
1.0
.2
.9
.14
.27
JointProbabilities
P(Ai B)
P(B) = .41
Tabular ApproachTabular Approach
.14/.41.14/.41
PosteriorProbabilities
P(Ai |B)
.3415
.6585
1.0000
6868 6868 SlideSlide
SlideSlide
EndEnd
top related