Distributions

A distribution is a mathematical description of a random variable's probability space.

Random Variable


A random variable is a mapping of a probability space to a set of real values. Define your own discrete random variable for the uniform probability space on the right and sample to find the distribution.

  1. Click and drag to highlight sections of the probability space, add a real number value, then press submit.

  2. Sample from probability space to generate the distribution for your random variable.

Sample Distribution
Color Value
0

Discrete and Continuous


A discrete random variable is a random variable that has countable values, such as a list of integers. Choose one of the following major discrete distributions to explore.


The Bernoulli distribution is the probability distribution of a random variable which takes the value 1 with success probability of \(p\) and the value 0 with failure probability of \(1-p\). It is frequently used to represent binary experiments, such as a coin toss.

The binomial distribution describes the result of \(n\) independent Bernoulli trials with probability \(p\). It is frequently used to model the number of successes in a specific number of identical binary experiments, such as the number of heads in five coin tosses.

The negative binomial distribution is the probability distribution of the number of successes in a sequence of independent Bernoulli trials with probability \(p\) before \(r\) failures occur. For example this distribution could be used to model the number of heads before three tails in a sequence of coin tosses.

The geometric distribution is the probability distribution that the first success is after exactly \(k\) failures in a sequence of independent Bernoulli trials with probability \(p\). For example this distribution can be used to model the number of rolls of a dice before the first six is rolled.

The Poisson distribution describes the probability of a given number of events occurring in a fixed interval of time and/or space if these events occur with a known average rate \(\lambda\), and independently of the time since the last event. This distribution has been used to model events such as meteor showers and goals in a soccer match.

The uniform distribution is a continuous distribution such that all intervals of the same length on the distribution's support are equally probable. For example this distribution might be used to model people's birth dates, where it is assumed all times are equally likely.

The normal (or Gaussian) distribution is used in the natural and social sciences to represent real-valued random variables whose distributions are not known. For example the normal distribution might be used to model people's height, where it is assumed most people are around the same height.

The Student's t-distribution, or simply the t-distribution, arises when estimating the mean of a normally distributed population in situations where the sample size is small and population standard deviation is unknown.

The chi-squared distribution arises from the sum of \(k\) independent and identically distributed squared standard normal distributions. It is used frequently in hypothesis testing and in the construction of confidence intervals.

The exponential distribution is the continuous analogue of the geometric distribution. It is frequently used when describing the time between independent events such as modelling atomic decay.

The F-distribution, also known as the Fisher–Snedecor distribution arises frequently as the null distribution of a test statistic, most notably in the analysis of variance.

The gamma distribution is a general family of continuous probability distributions used frequently to model data that is skewed normal. The exponential distribution and chi-squared distribution are special cases of the gamma distribution.

The beta distribution is a general family of continuous probability distributions bound between 0 and 1. The beta distribution is frequently used as a conjugate prior distribution in Bayesian Statistics.








Central Limit Theorem


The central limit theorem (CLT) states that the means from sufficiently large samples of independent random variables (each with a finite expected value and variance) will follow a normal distribution.

  1. Change the parameters alpha and beta to change the sampling beta distribution.


  2. Choose a sample size and a draw number, then press submit. Toggle the theoretical distribution.