A smooth, S-shaped function that maps any real number to the interval , serving as the bridge between unbounded linear scores and valid probabilities.
Definition
The sigmoid (logistic) function is defined as:
Equivalently: .

Key Properties
| Property | Statement |
|---|---|
| Range | |
| Midpoint | |
| Symmetry | |
| Monotonicity | Strictly increasing for all |
| Limits | , |
| Derivative |
The symmetry property is particularly useful. In logistic-regression, it means:
so both class probabilities can be expressed through the same function.
Role in Logistic Regression
In logistic-regression, the linear combination can be any real number, but we need a probability in . The sigmoid provides exactly this mapping:
The sigmoid is the inverse of the logit function: if , then .
Shape and Behaviour
The S-shape means the function is steepest around (where it equals ) and flattens out at the extremes — saturating toward 0 and 1. Practically:
- : is effectively 0 or 1.
- : is roughly linear, centred on .
This saturation is important: points far from the decision boundary (large ) get near-certain probabilities, while points near the boundary get probabilities close to .
Related
- logistic-regression — primary user of the sigmoid in this module
- decision boundary — the locus where
Active Recall
Prove that .
. Multiply numerator and denominator by : .
If , what probability does the sigmoid assign, and what does this mean for classification?
. The model is maximally uncertain — it assigns equal probability to both classes. This point lies exactly on the decision boundary.
Why does the sigmoid saturate (flatten) for large , and what is the practical consequence for classification confidence?
As , , so ; as , , so . The function approaches but never reaches 0 or 1. Practically, points with large — far from the decision boundary — get near-certain class probabilities, while points near the boundary (small ) stay close to .