PHD

PHDProbability and Statistics


Probability Theory


Probability theory is a fundamental branch of mathematics that deals with the analysis of random events. The central aspect of probability theory is the study of uncertainty and the quantification of randomness through the use of mathematical probabilities. This field of study is essential for understanding concepts in a variety of fields such as statistics, finance, gambling, science, and engineering, where outcomes cannot be determined with certainty.

Basic concepts

Probability theory begins by defining some fundamental concepts that are important to understanding the subject. These include:

Randomized experiment

A random experiment is an action or process that results in one or more possible outcomes. For example, tossing a coin is a random experiment where the possible outcomes are "heads" or "tails."

Sample space

The sample space, often denoted as S, is the set of all possible outcomes of a random experiment. For the example of tossing a coin, the sample space is S = { text{"Heads"}, text{"Tails"} }.

Events

An event is a subset of the sample space. It represents the occurrence of one or more outcomes. For example, when a six-sided dice is thrown, an event could be throwing an even number, which is the set E = { 2, 4, 6 }.

Possibility

Probability is a measure of the likelihood of an event occurring. It is measured as a number between 0 and 1, where 0 represents impossibility and 1 represents certainty. The probability of an event E is represented as P(E).

Understanding probability through examples

To better understand the concept of probability, let us examine some examples of calculating probability:

Example 1: Tossing a fair coin

Suppose we have a fair coin, which means that the probability of getting heads is equal to the probability of getting tails. We can calculate the probability of each outcome as follows.

Sample Space, S = { "Heads", "Tails" }

Probability of getting heads, P(text{"Heads"}) = frac{1}{2}

Probability of getting tails, P(text{"Tails"}) = frac{1}{2}

Example 2: Throwing a six-sided dice

Consider a normal six-sided die. Each face shows a number from 1 to 6. The probability of any specific number coming up can be calculated.

Sample Space, S = { 1, 2, 3, 4, 5, 6 }

Probability of getting 3, P(3) = frac{1}{6}

Probability of getting an even number, P({ 2, 4, 6 }) = frac{3}{6} = frac{1}{2}

Laws of probability

Probability theory is governed by a set of fundamental rules or axioms formulated by Russian mathematician Andrey Kolmogorov. Here are the primary rules:

Axiom 1: Non-negativity

The probability of any event is non-negative.

P(E) ≥ 0

Axiom 2: Generalization

The probability that at least one outcome of a random experiment will occur is 1. That is, the probability of the entire sample space is 1.

P(S) = 1

Axiom 3: Additivity

For any two mutually exclusive events, the probability of occurrence of either event is equal to the sum of their individual probabilities.

If E1 ∩ E2 = ∅, then P(E1 ∪ E2) = P(E1) + P(E2)

Illustrating probability with a Venn diagram

Venn diagrams are useful for representing events and their probabilities, especially when trying to understand complex probability relationships, such as conjunction and intersection of events.

E1 E2

In this Venn diagram, event E1 is represented by the blue circle and event E2 by the red circle. The overlapping region represents the intersection of events E1 and E2. This intersection is important in understanding the probability of two events occurring simultaneously.

Conditional probability and independence

Conditional probability and independence are important concepts in probability theory. They help assess the relationship between different events.

Conditional probability

Conditional probability is the probability of an event occurring, given that another event has occurred. If A and B are two events from the sample space, then the conditional probability of A given B is written as P(A|B)

P(A|B) = frac{P(A ∩ B)}{P(B)}, quad text{if} P(B) > 0

For example, if a card is drawn from a standard deck of 52 cards, find the probability that it is an ace, given that it is a heart card.

P(text{Ace} | text{Heart}) = frac{text{Probability of Ace of Hearts}}{text{Probability of Heart}} = frac{1/52}{13/52} = frac{1}{13}

Independence of events

Two events A and B are independent if the occurrence of one does not affect the probability of the other occurring. Mathematically, A and B are independent if:

P(A ∩ B) = P(A) * P(B)

For example, when two different dice are thrown, the outcome of one dice does not affect the outcome of the other dice. Therefore, the events are independent.

Bayes' theorem

Bayes' theorem is one of the most important results in probability theory. It describes the probability of an event based on prior knowledge of conditions that may be related to the event.

P(A|B) = frac{P(B|A) cdot P(A)}{P(B)}

Bayes' theorem allows probabilities to be updated based on new evidence. For example, if a medical test is positive for a disease, Bayes' theorem can be used to revise the probability that a person has the disease, taking into account the accuracy of the test.

Random variables

In probability theory, a random variable is a variable whose possible values are the outcomes of a random event. There are two types of random variables: discrete and continuous.

Discrete random variable

A discrete random variable is one that can take on a finite or countably infinite set of values. Examples include the outcome of throwing a dice or the sum of the outcomes when throwing two dice.

Continuous random variable

A continuous random variable, on the other hand, can take on an infinite set of values, often forming a continuum or involving measurements that form part of the real number line. An example involves the measurement of height in a group of people.

Probability distributions

Probability distributions describe how probabilities are distributed over the values of a random variable. Depending on the type of random variable they are classified into two types: discrete probability distributions and continuous probability distributions.

Discrete probability distributions

Discrete probability distributions are characterized by the probability mass function (PMF), which assigns a probability to each possible value of the random variable.

A common example of this is the binomial distribution, which describes the number of successes in a given number of independent Bernoulli trials.

P(X = k) = binom{n}{k} p^k (1-p)^{nk}

where n is the number of trials, p is the probability of success in each trial, and k is the number of successes.

Continuous probability distributions

Continuous probability distributions are characterized by a probability density function (PDF). The most famous example is the normal distribution, or bell curve.

f(x) = frac{1}{sigma sqrt{2pi}} e^{-frac{1}{2}(frac{x-mu}{sigma})^2}

where mu is the mean and sigma is the standard deviation.

Expected value and variance

The expected value is the long-term average value of the repetitions of the experiment it represents. It is often referred to as the mean.

E(X) = sum_{i} x_i P(x_i), quad text{for discrete} E(X) = int_{-infty}^{infty} xf(x) dx, quad text{for continuous}

Variance measures the spread of the values of a random variable. It is the expected value of the squared deviation of a random variable from its mean.

Var(X) = E[(X - E[X])^2] = sum_{i} (x_i - E[X])^2 P(x_i)

Variance measures how much variability exists in a distribution around the mean.

Law of large numbers and central limit theorem

The law of large numbers and the central limit theorem are two important concepts that deal with the behavior of probabilities in the context of large datasets.

Law of large numbers

The law of large numbers states that as the number of trials of a random process increases, the experimental probability will tend to approach the theoretical (true) probability.

Central limit theorem

The Central Limit Theorem states that the distribution of the sample means of a large number of independent, identically distributed variables will be approximately normally distributed, regardless of the shape of the original distribution.

Conclusion

Probability theory provides a comprehensive framework for reasoning about randomness and uncertainty. It underlies many real-world applications where outcomes are inherently unpredictable. The field aids decision-making processes under uncertain circumstances and forms the essential foundation for the development of statistical methods used to interpret data.


PHD → 8.1


U
username
0%
completed in PHD


Comments