The density and distribution functions always return values between zero and

Perhaps the following thought-experiment helps you to understand better why the probability $Pr(X=a)$ is zero in a continuous distribution: Imagine that you have a wheel of fortune. Normally, the wheel is partitioned in several discrete sectors, perhaps 20 or so. If all sectors have the same area, you would have a probability of $1/20$ to hit one specific sector (e.g. the main price). The sum of all probabilities is 1, because $20\cdot 1/20 = 1$. More general: If there are $m$ sectors evenly distributed on the wheel, every sectors has a probability of $1/m$ of being hit (uniform probabilities). But what happens if we decided to partition the wheel into a million sectors. Now the probability of hitting one specific sectors (the main prize), is extremely small: $1/10^{6}$. Further, note that the pointer can theoretically stop at an infinite number of positions of the wheel. If we wanted to make a separate prize for each possible stopping point, we would have to partition the wheel in an infinite number of "sectors" of equal area (but each of those would have an area of 0). But what probability should we assign to each of these "sectors"? It must be zero because if the probabilities for each "sectors" would be positive and equal, the sum of infinitely many equal positive numbers diverges, which creates a contradiction (the total probability must be 1). That's why we only can assign a probability to an interval, to a real area on the wheel.

More technical: In a continuous distribution (e.g. continuous uniform, normal, and others), the probability is calculated by integration, as an area under the probability density function $f(x)$ (with $a\leq b$): $$ P(a\leq X \leq b) = \int_{a}^{b} f(x) dx $$ But the area of an interval of length 0 is 0.

See this document for the analogy of the wheel of fortune.

The Poisson distribution on the other hand is a discrete probability distribution. A random Poisson variable can only take discrete values (i.e. the number of children for one family cannot be 1.25). The probability that a family has exactly 1 child is certainly not zero but is positive. The sum of all probabilities for all values must be 1. Other famous discrete distributions are: Binomial, negative binomial, geometric, hypergeometric and many others.

A probability density function describes a probability distribution for a random, continuous variable. Use a probability density function to find the chances that the value of a variable will occur within a range of values that you specify. More specifically, a PDF is a function where its integral for an interval provides the probability of a value occurring in that interval. For example, what are the chances that the next IQ score you measure will fall between 120 and 140? In statistics, PDF stands for probability density function.

Don’t worry! I’ll break all that down to simpler terms!

Unlike distributions for discrete variables where specific values can have non-zero probabilities, the likelihood for a single value is always zero in a continuous probability distribution function. Instead, the probability density function provides the chances of a value falling within a specified range for continuous variables. For a PDF in statistics, probability density refers to the likelihood of a value occurring within an interval length of one unit.

In short, probability density functions can find non-zero likelihoods for continuous variable X falling within the interval [a, b]. Or, in statistical notation: P (A < X < B).

Related posts: Discrete vs Continuous Variables and Probability Distributions

Obtaining a Probability from a PDF

Graphing a probability density function gives you a probability density plot. These graphs are great for understanding how a PDF in statistics calculates probabilities. The chart below displays the PDF for IQ scores, which is a probability density function of a normal distribution.

The density and distribution functions always return values between zero and

On a probability density plot, the total area under the curve equals one, representing the total probability of 1 for the full range of possible values in the distribution. In other words, the likelihood of a value falling anywhere in the complete distribution curve is 1. Finding the chances for a smaller range of interest involves finding the portion of the area under the curve corresponding to that range. In our example, we need to find the percentage of the area that falls between 120 and 140.

The density and distribution functions always return values between zero and

In the example above, my statistical software finds that P(120 < X < 140) = 0.08738. The probability of a value falling within the range of 120 to 140 is 0.08738.

These graphs also illustrate why probability density functions find a zero likelihood for an individual value. Consider that the probability for a PDF in statistics equals an area. For a non-zero area, you must have both a non-zero height and a non-zero width because Height X Width = Area. In this context, the height is the curve’s height on the graph, while the width relates to the range of values. When you have a single value, you have zero width, which produces zero area. Hence, a zero chance!

Mathematics Behind a Probability Density Function

In mathematical terms, you or your software will use integrals with a probability density function to find probabilities. Among other uses, the integration process finds areas for mathematical functions that describe shapes. That’s perfect for finding likelihoods for intervals in a continuous distribution curve based on a PDF. You’ll usually use statistical software to find PDF probabilities, so I won’t discuss integrals here. For more information, read the Wikipedia article for more details: Integrals.

Examples of Probability Density Functions

While the normal distribution is the most common, it can fit only symmetrical distributions. Fitting skewed distributions requires using other probability density functions. Learn more about the Normal Distribution and Skewed Distributions.

There are a variety of other probability density functions that correspond with distributions of different shapes and properties. Each PDF has between 1-3 parameters that define its shape. Before using a PDF to find a probability, you must identify the correct function and parameter values for the population you are studying.

To accomplish that, you’ll typically gather a random sample from that population and use a combination of parameter estimation techniques and hypothesis tests to identify the distribution and parameters that most likely produced your sample.

For more information about that process, read my post about Identifying the Distribution of Your Data.

Let’s look at examples of the lognormal and Weibull distributions, which can fit skewed data.

Lognormal Probability Density Function

A lognormal PDF models right-skewed distributions. Below is an example of the probability density function for a lognormal distribution that displays the distribution of body fat percentages for teenage girls. The data are from a study I performed.

The density and distribution functions always return values between zero and

Notice how it shows a right-skewed distribution. Additionally, the probability distribution plot indicates that 18.64% of the population will fall between 20-24%. The highest probability densities occur near the peak at 26%.

Weibull PDF

The Weibull PDF is a highly flexible distribution that can model a wide variety of shapes. The example below displays a left-skewed distribution.

The density and distribution functions always return values between zero and

Examples of Other Probability Density Functions

There are a variety of PDFs for other distributions. They include the following:

  • Weibull distribution: A remarkably versatile distribution that analysts use in many settings. Can model skewed distributions in both directions and approximate the normal distribution.
  • Lognormal distribution: Models right-skewed distributions, especially in cases where growth rates are unconnected to size.
  • Exponential distribution: Models distributions where small values occur more frequently than large values. Use to model the amount of time between independent events.
  • Gamma distribution: Models right-skewed distributions. Use to model the time until the kth event.
  • Uniform distribution: Models symmetric distributions where all equal length ranges have the same probability.
  • Beta distribution: Models distributions with all values falling within a finite range.