Why exp(+x) Means Complete
Tier: THEOREM (T26–T28, structural proofs; T12 updated from Trichotomy to full characterization)
There are exactly 16 natural ways to combine $\exp(\pm x)$ with $\ln(y)$ using one arithmetic operation. Classifying all 16 for completeness — the ability to represent every elementary function as a finite tree — reveals a single structural rule that explains every case.
The Rule
An exp-ln operator is exactly complete if and only if its exponential term is exp(+x) and the combining operation does not introduce a domain restriction.
That’s it. One sentence. All 16 operators follow.
The Full Classification
| Completeness | Operators | Structural feature |
|---|---|---|
| Exactly complete (8) | EML, EAL, EXL, EDL, EPL, LEAd, ELAd, ELSb | exp(+x), no domain restriction |
| Approximately complete (1) | EMN | −exp(+x) (negation outside) |
| Incomplete (7) | DEML, DEMN, DEAL, DEXL, DEDL, DEPL, LEX | exp(−x) or domain-restricted |
The old “Completeness Trichotomy” (T12) described this as “1 complete / 1 approximate / 6 incomplete” because only EML was known to be complete at the time. The full census (T25) reveals 8 complete operators — all of them structurally equivalent in the dimension that matters.
Why exp(+x) Works: The Forward Direction (T26)
Every complete operator can construct exp(x) in 1 node and a slope-−1 linear function in 2 nodes. The slope-−1 construction is the key: it gives negation up to a constant offset.
For EML specifically: eml(c, exp(x)) = exp(c) − x. At c = 0: eml(0, exp(x)) = 1 − x.
That’s slope −1 with offset 1.
To get exact neg(x) = −x, we use the SuperBEST cross-family bridge (T09):
neg(x) = exl(0, deml(x,1)) = exp(0) · ln(exp(−x) − ln(1)) = 1 · (−x) = −x
Two nodes. Exact.
The common mechanism: exp(+x) grows without bound as x → +∞. This unbounded upward growth, combined with ln (which maps ℝ⁺ → ℝ), gives the operator the full real line as output range. From full range comes identity; from identity comes negation; from negation comes completeness (Ritt’s theorem).
Why exp(−x) Fails: The Reverse Direction (T27)
The exp(−x) operators each fail by a different mechanism:
| Operator | Mechanism | Core barrier |
|---|---|---|
| DEML | Slope barrier | Self-composition always has slope +1; neg requires −1 |
| DEMN | Domain failure | deml(x,1) < 0 for all x; can’t feed into ln |
| DEAL | Domain collapse | Achieves slope −1 but offset e⁻¹ is irremovable; deeper compositions lose domain |
| DEXL | Dead constant | dexl(x,1) = exp(−x)·ln(1) = 0 for all x |
| DEDL | Decay barrier | Self-composition gives −e⁻¹/x → 0; no linear growth |
| DEPL | Decay barrier | exp(−x)^(ln y) ≤ 1 for x ≥ 0; bounded above |
The unifying feature: exp(−x) decays to 0 as x → +∞. A decaying term cannot provide unbounded growth in either direction. The only source of growth in h(exp(−x), ln(y)) is the ln(y) branch — which grows logarithmically, not linearly. So slope ±1 is either unreachable or achievable only with irremovable constant offsets.
DEAL is the subtlest case. It achieves slope −1 (DEAL(1, DEAL(x,1)) = e⁻¹ − x). But the offset e⁻¹ cannot be removed: doing so requires neg (circular), and using e⁻¹ − x as a right-branch input requires it to be positive, which restricts x < e⁻¹ ≈ 0.368. The domain collapses at the second level of nesting.
The Exception: LEX (T28)
LEX(x, y) = ln(exp(x) − y) has exp(+x), yet it is incomplete.
The reason is a domain restriction in the combining operation: ln(exp(x) − y) requires exp(x) > y. With y = 1, this means x > 0. With self-composition at depth 2, the domain shrinks to x < 2.81. At depth 3, it shrinks to x < 1.8. The pattern continues: the domain approaches the empty set under iteration.
No LEX tree with depth ≥ 2 is defined on all of ℝ, so LEX cannot represent any globally defined function. The incompleteness theorem for LEX (T28) is:
LEX is incomplete because its domain under self-composition collapses to a set of measure approaching 0 as depth increases.
This is a different mechanism from exp(−x) incompleteness. LEX fails at the domain level, not the growth level.
The Bridge Case: EMN
EMN(x, y) = ln(y) − exp(x) has −exp(x) (negation outside, not inside).
The difference from exp(−x) is crucial:
- exp(−x) decays: bounded above, approaches 0
- −exp(x) grows without bound (negatively): → −∞ as x → +∞
This unbounded negative growth lets EMN approximate any target to arbitrary precision.
For approximate negation: emn(0, e^(e^k)) = k − e^0 → k for large k, and variations
get arbitrarily close to −x. The error is always of order exp(−e^k), which converges
doubly-exponentially to 0 but never reaches 0 in finite depth.
The obstruction to exact completeness: every EMN tree has an exp(·) residual in the subtracted term that cannot be made exactly zero. Approximate completeness is the ceiling for EMN.
The Exponential Position Theorem
Statement: An exp-ln operator’s completeness class is determined entirely by the position of negation relative to exp:
| Negation position | Growth type | Completeness class |
|---|---|---|
| exp(+x) (no negation, no domain restriction) | Unbounded upward | Exactly complete |
| −exp(x) (negation outside) | Unbounded downward | Approximately complete |
| exp(−x) (negation inside) | Bounded, decaying | Incomplete |
| exp(+x) with domain restriction | Restricted domain | Incomplete |
This is the updated T12 (was: Completeness Trichotomy with 1/1/6 split; now: 8/1/7 split with structural explanation).
What Changed from the Trichotomy
The Completeness Trichotomy post identified three classes with 1 exactly complete operator (EML), 1 approximately complete (EMN), and 6 incomplete. The full census reveals:
- 8 complete operators, not 1 — EML was never special in the dimension that matters
- 7 incomplete operators, not 6 — LEX adds a new incompleteness mechanism
The theorem count goes from T07 to T28 with this sprint. T12 is now the Completeness Characterization Theorem (Exponential Position Theorem), citing T26 (forward), T27 (reverse), and T28 (LEX domain).
New Result: Softplus = 1 LEAd Node (T19)
The softplus activation function ln(1 + exp(x)) is exactly LEAd(x, 1):
LEAd(x, y) = ln(exp(x) + y)
LEAd(x, 1) = ln(exp(x) + 1) = softplus(x)
One node. The entire softplus function — ubiquitous in neural networks and smooth approximations of ReLU — is a single LEAd application with constant 1.
Corollary: log-sum-exp of N terms costs N−1 LEAd nodes. The denominator of a softmax over N logits: N−1 nodes.
Monogate Research (2026). “Why exp(+x) Means Complete: The Structural Theorem for Exponential-Logarithmic Operators.” monogate research blog. https://monogate.org/blog/completeness-characterization
Session COMP-1 through COMP-5 · T12 updated, T26–T28 added · Theorem count: 28