is the branch of mathematics concerned with analysis of random phenomena
is a way of expressing what the chances are that an event will occur
is a variable \(v\) whose values depend on
outcomes of a random phenomenon
deal with events that occur in a countable sample space
deal with events that occur in a continuous sample space
Thus, the probability \(p\) of one discrete random variable (event) \(v\) which is defined for \(n\) states comprises \(n\) chances for every state $$p(v) \equiv \left\{p(v=s_1), p(v=s_2),\dots,p(v=s_n)\right\}$$
In the DGM library such probabilities are stored as vectors of floating-point numbers:
std::vector<float> probability;
Here \(p(v_1,v_2)\) is a joint probability and is verbalised as
the probability of \(v_1\) and \(v_2\)
In DGM library this is a two-dimensional matrix whose size is equal to the number of states of each of the random variables and the value of each element is equal to the probability of events \(v_1\) and \(v_2\) occurring simultaneously
cv::Mat probability(n, m, CV_32FC1);
Joint probability is commutative: $$p(v_1,v_2)=p(v_2,v_1)$$
Here the quantity \(p(v1~|~v_2)\) is a conditional probability and is verbalised as
the probability of \(v_1\) given \(v_2\)
In DGM library this is a two-dimensional matrix whose size is equal to the number of states of each of the random variables and the value of each element is equal to the probability of event \(v_1\) occurring, provided event \(v_2\) has occurred
cv::Mat probability(n, m, CV_32FC1);
Conditional probability is not commutative: $$p(v_1~|~v_2)\not=p(v_2~|~v_1)$$
Two random variables \(v_1\) and \(v_2\) are statistically independent (also unconditional independent) if and only if
$$p(v_1~|~v_2)=p(v_1)$$
As consequence we have:
Two random variables \(v_1\) and \(v_2\) are statistically independent (also unconditional independent) if and only if
$$p(v_1~|~v_2)=p(v_1)$$
As consequence we have:
Two random variables \(v_1\) and \(v_2\) are conditional independent given a third random variable \(v_3\) if and only if
$$p(v_1~|~v_2, v_3)=p(v_1~|~v_3)^1$$
As consequence we have:
\(^1\)This notation is equivalent to \(v_1\perp\!\!\!\perp v_2~|~v_3\)
Two random variables \(v_1\) and \(v_2\) are conditional independent given a third random variable \(v_3\) if and only if
$$p(v_1~|~v_2, v_3)=p(v_1~|~v_3)^1$$
As consequence we have:
\(^1\)This notation is equivalent to \(v_1\perp\!\!\!\perp v_2~|~v_3\)
which must be hold for every possible value of \(v_3\), and not just for some values
By application of the product rule of probability
\(p(v_i,v_j)=p(v_i~|~v_j)\cdot p(v_j)\), we can write the joint distribution for an arbitrary number \(n\) of random variables \(\vec{v} = (v_1,\dots,v_n)^\top\)
Note, that this decomposition holds for any choice of the joint distribution
or
c. 1701 - 7 April 1761
British mathematician and priest
Portrait purportedly of Bayes used in a 1936 book, but it is doubtful whether the portrait is actually of him. No earlier portrait or claimed portrait survives
Bayes law may be derived directly from the product rule and the commutative property of joint probability:
$$p(v_1,v_2)=p(v_1~|~v_2)\cdot p(v_2)$$
$$p(v_2,v_1)=p(v_2~|~v_1)\cdot p(v_1)$$
$$p(v_1,v_2)=p(v_2,v_1)\Longrightarrow p(v_1~|~v_2)\cdotp(v_2) = p(v_2~|~v_1)\cdot p(v_1)$$
c. 1701 - 7 April 1761
British mathematician and priest
Portrait purportedly of Bayes used in a 1936 book, but it is doubtful whether the portrait is actually of him. No earlier portrait or claimed portrait survives
Bayes law may be derived directly from the product rule and the commutative property of joint probability:
Posterior probability
Likelyhood
Prior probability
"Bayes law for probability theory, the same as Pythagorean theorem for geometry"
-Harold Jeffreys, British mathematician
Total probability
c. 1701 - 7 April 1761
British mathematician and priest
Portrait purportedly of Bayes used in a 1936 book, but it is doubtful whether the portrait is actually of him. No earlier portrait or claimed portrait survives
Bayes law may be derived directly from the product rule and the commutative property of joint probability:
Posterior probability
Likelyhood
Prior probability
Total probability
Total probability: $$p(B)=\displaystyle\sum^{n}_{i=1}{p(B|A_i)\cdot p(A_i)}$$
Allows to calculate the probability of an event of interest through conditional probabilities of this event in respect to the probabilities of some sypothesises
Bayes law may be derived directly from the product rule and the commutative property of joint probability:
Posterior probability
Likelyhood
Prior probability
Total probability: $$p(B)=\displaystyle\sum^{n}_{i=1}{p(B|A_i)\cdot p(A_i)}$$
Allows to calculate the probability of an event of interest through conditional probabilities of this event in respect to the probabilities of some sypothesises
Total probability
Bayes law may be derived directly from the product rule and the commutative property of joint probability:
$$p(B)=\displaystyle\sum^{n}_{i=1}{p(B|A_i)\cdot p(A_i)}$$
Bayes law may be derived directly from the product rule and the commutative property of joint probability:
$$p(B)=\displaystyle\sum^{n}_{i=1}{p(B|A_i)\cdot p(A_i)}$$