On the Mathematical Relationship Between Layer Normalization and Dynamic Activation Functions

Felix Stollenwerk


Abstract
Layer normalization (LN) is an essential component of modern neural networks. While many alternative techniques have been proposed, none of them have succeeded in replacing LN so far. The latest suggestion in this line of research is a dynamic activation function called Dynamic Tanh (DyT). Although it is empirically well-motivated and appealing from a practical point of view, it lacks a theoretical foundation. In this work, we shed light on the mathematical relationship between LN and dynamic activation functions. In particular, we derive DyT from the LN variant RMSNorm, and show that a well-defined decoupling in derivative space as well as an approximation are needed to do so. By applying the same decoupling procedure directly in function space, we are able to omit the approximation and obtain the exact element-wise counterpart of RMSNorm, which we call Dynamic Inverse Square Root Unit (DyISRU). We demonstrate numerically that DyISRU reproduces the normalization effect on outliers more accurately than DyT does.
Anthology ID:
2026.eacl-short.48
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
674–681
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-short.48/
DOI:
Bibkey:
Cite (ACL):
Felix Stollenwerk. 2026. On the Mathematical Relationship Between Layer Normalization and Dynamic Activation Functions. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers), pages 674–681, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
On the Mathematical Relationship Between Layer Normalization and Dynamic Activation Functions (Stollenwerk, EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-short.48.pdf
Checklist:
 2026.eacl-short.48.checklist.pdf