· Casper van Elteren · Statistics  Â· 10 min read

From Self-Limiting Growth to the Logit Function

Why concerns about population growth led to the logistic equation, and how the same curve later became central to probability models.

Mathematics is often taught as if formulas arrive fully formed. We are shown the final expression, perhaps a short derivation, and then asked to remember the result. What usually gets lost in that process is the original motivation: what concrete problem made someone invent the model in the first place?

I find mathematical ideas much easier to hold onto when they are tied back to the reasoning that produced them. The logistic and logit functions are a good example. Today they show up in statistics, machine learning, and forecasting, but they did not begin as generic tools for classification. They grew out of a much older question about population, scarcity, and what happens when growth runs into limits.

This post follows that chain of ideas from the beginning. I focus on one historical path into the logistic function: how it emerged from attempts to understand population pressure and human misery, and how it later became a much more general tool. The short version is:

  1. Malthus modeled population with constant per-capita growth.
  2. Verhulst asked how to correct that model once growth starts to slow itself down.
  3. The resulting solution is the logistic function.
  4. Later statisticians reused the same curve because it maps unrestricted values into valid probabilities.
  5. The associated transformation from probabilities to log-odds became the logit.
One curve, two meaningsA feature graphic linking Verhulst's bounded growth curve to the later inverse-logit probability model.FROM VERHULST TO LOGITOne curve, two meaningsThe same curve first explains how growth slows once crowding sets in, then returnsas the inverse logit that turns raw scores on the real line into probabilities.1838 VERHULSTPopulation growthGrowth starts out fast, then bendsas crowding and limits begin to matter.1944 BERKSONProbability mappingThe same S-curve gently turns araw score into a probability you can read.0246824TIMEPOPULATION−6−4−202460.20.40.60.81.0LINEAR SCOREPROBABILITYSAME CURVEbounded growthbecomes abounded probabilityVerhulst used the curve to describe growth with limits; later statisticians reused it to turn raw scores into probabilities.

From Verhulst to Logit

One curve, two meanings

The logistic curve first described self-limiting growth and later became the inverse logit for probabilities.

1838 Verhulst

Population growth

Growth starts fast, then bends as crowding and finite capacity begin to matter.

Growth Law

p'(t) = mp(1 - p/P)

1944 Berkson

Probability mapping

The same S-curve turns an unrestricted score on the real line into a readable probability.

Inverse Logit

σ(x) = 1 / (1 + e-x)

Verhulst used the curve to describe growth with limits; later statisticians reused it to turn raw scores into probabilities.

Why Population Became a Mathematical Problem

In the late eighteenth and early nineteenth centuries, population was not an abstract academic topic. It sat directly inside political economy. Governments and scholars were worried about food supply, wages, poverty, poor relief, taxation, military strength, and crowding. The core question was not simply whether the population was large, but whether population growth could outrun what a country was able to support1,2.

There were competing explanations for the causes of human misery. William Godwin represented a more optimistic view. He argued that people are highly improvable through reason, education, and freer social arrangements. On that view, many social problems came less from natural limits than from bad institutions and coercive social organization3,1.

Thomas Malthus took a harder line. He thought this optimism underestimated a basic constraint: population pressure on subsistence. Food is necessary for life, people continue to reproduce when conditions allow, and agriculture cannot expand without limit. For Malthus, many forms of suffering were therefore not only institutional, but also the result of a population pressing against the resources available to sustain it1.

Malthus and the Problem of Exponential Growth

Malthus’s starting point can be written as

This says the per-capita growth rate is constant. If every individual contributes, on average, the same amount to future growth, then larger populations grow faster in absolute terms.

The solution is the familiar exponential law

This is a clean and sensible first approximation. When population is small relative to the available land and food supply, exponential growth is exactly the kind of behavior one might expect. But it also reveals the weakness of the model: if the per-capita growth rate remains constant forever, then the total population grows without bound1.

Malthus used this argument to push back against the more optimistic view that social improvement alone could remove misery. Even if institutions become better, unchecked population growth still runs into material limits. Verhulst accepted that basic concern, but asked a more precise mathematical question: what happens once growth begins to slow because the size of the population itself starts to matter?

Verhulst’s Correction: Growth That Slows Itself Down

Pierre-Francois Verhulst accepted the Malthusian picture as a useful approximation at low population density, but argued that it could not remain true indefinitely. As population rises, crowding, competition, and limited resources feed back into the growth process itself4,6.

So instead of assuming a constant per-capita growth rate, he introduced a retardation term:

The key modeling question is then: what shouldbe?

Verhulst’s simplest useful choice was to let the reduction in per-capita growth be proportional to the size of the population itself:

That gives

This is where the quadratic term comes from. It is not a mysterious extra ingredient. It appears because total growth is population times per-capita growth, and the per-capita growth rate has been assumed to decline linearly with population size 4.

Whenis small, theterm is negligible compared with the linear term, so we recover the Malthusian model as a first approximation. But asgrows, the quadratic term matters more and more, and growth begins to slow.

Interactive Figure

Exponential growth versus logistic growth

The exponential curve keeps rising because its per-capita growth rate never changes. The logistic curve bends because the same growth process is discounted by the room left before the system hits its carrying capacity.

0.002.004.006.008.0010.0012.0014.0016.0002468101214carrying capacity P = 1.80exponentiallogistictimepopulation size p(t)

Exponential

\(p(t) = p_0 e^{mt}\)

Logistic

\(p(t) = \frac{P}{1 + A e^{-mt}}\)

Current contrast

At t = 14, the exponential model reaches 16.35, while the logistic model settles near 1.63.

The interactive figure makes the modeling contrast visible. The exponential model never asks whether there is still room to grow. The logistic model does. As a result, it bends and eventually saturates.

Solving the Logistic Equation

The logistic differential equation can be written as

A useful substitution is

Differentiating gives

so

Substituting into the original equation yields

Multiplying through bygives a first-order linear equation:

This is much easier to solve. First solve the homogeneous equation

If we try an exponential ansatz, then

So the homogeneous solution is

Now find a particular solution. Since the forcing term on the right-hand side is the constant, the simplest guess is a constant solution. Substituting gives

So

Combining both pieces,

Since, we have

If we now define

then

and the solution becomes

Renaming the constantgives the standard form

This is the logistic function. In the original growth setting it maps time to a bounded interval betweenand the carrying capacity.

Why Is It Called the Logistic Function?

The name is older than the statistical use. Verhulst introduced it for the growth curve itself, not for a probability transformation. What is less satisfying is that the exact etymology is not fully settled 5,9.

My own reading is that the name may be connected to the way Verhulst first wrote the solution:

In that form, time appears as a logarithmic function of population. That makes a log-related name plausible, but it is still only an inference; Verhulst did not spell out the etymology. The important point is simply that the name belongs to the nineteenth-century growth model, not to the later statistical use.

From Growth Curves to Probability Models

Much later, statisticians faced a different problem. In dose-response studies and binary outcome models, the quantity of interest is often a probability. A straight-line model is awkward in that setting because it can predict values outside the interval fromto7,8.

What they needed instead was a smooth monotone curve that takes an unrestricted input and turns it into a valid probability. The probit model was one early answer. Berkson’s move was important because the logistic curve had simple algebra and a direct interpretation in terms of odds7,8.

The standard logistic function in statistics is

This is the same S-shaped functional form as the normalized growth curve, but now the output lives inand is read as a probability rather than a population level.

The Logit and the Inverse Logit

The logit function maps a probability to the real line:

The quantity

is the odds, so the logit is the log-odds. Its inverse is exactly the standard logistic function:

That is why the same curve appears in two rather different settings:

  • in population models, it comes from self-limiting growth;
  • in statistics, it comes from mapping unrestricted scores into valid probabilities.

Interactive Figure

Linking the real line, probabilities, and odds

Move the linear score and watch the two transformations stay synchronized. The sigmoid takes an unrestricted score and turns it into a probability; the logit takes that probability back to the real line.

Sigmoid / inverse logit

Maps any real-valued score x into a valid probability p.

0.000.200.400.600.801.00−6.00−4.00−2.000.002.004.006.00linear score xprobability p

Logit

Maps a probability p back into a log-odds score on the real line.

−6.00−4.00−2.000.002.004.006.000.200.400.600.80probability plogit(p)

Probability

0.500

Odds

1.00

Identity

\(\operatorname{logit}(p) = \log\left(\frac{p}{1-p}\right) = 0.00\)

Why the Logit Is Useful

The practical strength of the logit is that probabilities are constrained while linear predictors are not. A regression model often wants to work on an unrestricted scale:

But if the outcome is a probability, then we need

The logit provides exactly the bridge needed to move between those two domains:

This is why logistic regression is built the way it is. Effects are modeled additively on the log-odds scale, and then transformed back into probabilities through the inverse logit.

In that sense, the modern use of the logistic function is conceptually different from Verhulst’s original problem, but mathematically it reuses the same form because the same shape solves a similar structural requirement: bounded output with smooth, monotone behavior.

Closing Thoughts

What I like about this story is that it makes the mathematics feel less arbitrary. The logistic function was not invented because somebody wanted an S-shaped curve for its own sake. It emerged from a specific attempt to correct exponential growth by building resource limits into the model itself. Later, the same curve reappeared because statisticians needed a disciplined way to turn real-valued predictors into probabilities.

That kind of migration across disciplines is one of the most interesting things about mathematics. An idea begins inside one concrete problem, shaped by one historical setting, and then later becomes part of a much wider conceptual toolkit. The logistic function began as a model of self-limiting growth. Today it helps express risk, probability, and decision-making across medicine, ecology, and statistical forecasting.

Remembering that history makes the formulas easier to hold onto. More importantly, it gives them back some of their original logic. The equations stop looking like isolated symbolic tricks and start looking like answers to real questions people once had to take seriously.

Source Notes

References

  1. 1.

    Malthus, Thomas Robert (1798). An Essay on the Principle of Population . Project Gutenberg edition . Link

  2. 2.

    UK Parliament (n.d.). The 1800 Population Act . UK Parliament . Link

  3. 3.

    Encyclopaedia Britannica (n.d.). William Godwin . Encyclopaedia Britannica . Link

  4. 4.

    Verhulst, Pierre-François (1838). Notice on the Law That the Population Follows in Its Increase . English translation PDF . Link

  5. 5.

    Verhulst, Pierre-François (1845). Recherches Mathématiques sur la Loi d'Accroissement de la Population . Nouveaux mémoires de l'Académie Royale des Sciences et Belles-Lettres de Bruxelles . Link

  6. 6.

    Encyclopaedia Britannica (n.d.). Logistic Function . Encyclopaedia Britannica . Link

  7. 7.

    Cramer, J. S. (2004). The Early Origins of the Logit Model . Studies in History and Philosophy of Science Part C: Studies in History and Philosophy of Biological and Biomedical Sciences . Link

  8. 8.

    Berkson, Joseph (1944). Application of the Logistic Function to Bio-Assay . Journal of the American Statistical Association . Link

  9. 9.

    Semushin, Igor (n.d.). Original Sources for Logistic Equations . Teaching note with translated excerpts from Verhulst . Link

Back to Blog

Related Posts

View All Posts »