To install click the Add extension button. That's it.

The source code for the WIKI 2 extension is being checked by specialists of the Mozilla Foundation, Google, and Apple. You could also do it yourself at any point in time.

4,5
Kelly Slayton
Congratulations on this excellent venture… what a great idea!
Alexander Grigorievskiy
I use WIKI 2 every day and almost forgot how the original Wikipedia looks like.
Live Statistics
English Articles
Improved in 24 Hours
Added in 24 Hours
What we do. Every page goes through several hundred of perfecting techniques; in live mode. Quite the same Wikipedia. Just better.
.
Leo
Newton
Brights
Milds

Chain rule (probability)

From Wikipedia, the free encyclopedia

In probability theory, the chain rule[1] (also called the general product rule[2][3]) describes how to calculate the probability of the intersection of, not necessarily independent, events or the joint distribution of random variables respectively, using conditional probabilities. The rule is notably used in the context of discrete stochastic processes and in applications, e.g. the study of Bayesian networks, which describe a probability distribution in terms of conditional probabilities.

YouTube Encyclopedic

  • 1/5
    Views:
    2 806
    5 832
    146 153
    44 288
    802
  • Chain Rule in Probability
  • The Chain Rule for Probability
  • The Chain Rule
  • (PP 2.4) Bayes' rule and the Chain rule
  • Chain Rule for Events in Probability

Transcription

Chain rule for events

Two events

For two events and , the chain rule states that

,

where denotes the conditional probability of given .

Example

An Urn A has 1 black ball and 2 white balls and another Urn B has 1 black ball and 3 white balls. Suppose we pick an urn at random and then select a ball from that urn. Let event be choosing the first urn, i.e. , where is the complementary event of . Let event be the chance we choose a white ball. The chance of choosing a white ball, given that we have chosen the first urn, is The intersection then describes choosing the first urn and a white ball from it. The probability can be calculated by the chain rule as follows:

Finitely many events

For events whose intersection has not probability zero, the chain rule states

Example 1

For , i.e. four events, the chain rule reads

.

Example 2

We randomly draw 4 cards without replacement from deck with 52 cards. What is the probability that we have picked 4 aces?

First, we set . Obviously, we get the following probabilities

.

Applying the chain rule,

.

Statement of the theorem and proof

Let be a probability space. Recall that the conditional probability of an given is defined as

Then we have the following theorem.

Chain rule —  Let be a probability space. Let . Then

Proof

The formula follows immediately by recursion

where we used the definition of the conditional probability in the first step.

Chain rule for discrete random variables

Two random variables

For two discrete random variables , we use the eventsand in the definition above, and find the joint distribution as

or

where is the probability distribution of and conditional probability distribution of given .

Finitely many random variables

Let be random variables and . By the definition of the conditional probability,

and using the chain rule, where we set , we can find the joint distribution as

Example

For , i.e. considering three random variables. Then, the chain rule reads

Bibliography

  • René L. Schilling (2021), Measure, Integral, Probability & Processes - Probab(ilistical)ly the Theoretical Minimum (1 ed.), Technische Universität Dresden, Germany, ISBN 979-8-5991-0488-9{{citation}}: CS1 maint: location missing publisher (link)
  • William Feller (1968), An Introduction to Probability Theory and Its Applications, vol. I (3 ed.), New York / London / Sydney: Wiley, ISBN 978-0-471-25708-0
  • Russell, Stuart J.; Norvig, Peter (2003), Artificial Intelligence: A Modern Approach (2nd ed.), Upper Saddle River, New Jersey: Prentice Hall, ISBN 0-13-790395-2, p. 496.

References

  1. ^ Schilling, René L. (2021). Measure, Integral, Probability & Processes - Probab(ilistical)ly the Theoretical Minimum. Technische Universität Dresden, Germany. p. 136ff. ISBN 979-8-5991-0488-9.{{cite book}}: CS1 maint: location missing publisher (link)
  2. ^ Schum, David A. (1994). The Evidential Foundations of Probabilistic Reasoning. Northwestern University Press. p. 49. ISBN 978-0-8101-1821-8.
  3. ^ Klugh, Henry E. (2013). Statistics: The Essentials for Research (3rd ed.). Psychology Press. p. 149. ISBN 978-1-134-92862-0.
This page was last edited on 10 October 2023, at 00:00
Basis of this page is in Wikipedia. Text is available under the CC BY-SA 3.0 Unported License. Non-text media are available under their specified licenses. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc. WIKI 2 is an independent company and has no affiliation with Wikimedia Foundation.