Embeddings for Numerical Features Using tanh Activation

Bingyan Liu, Charles Elkan, Anil N. Hirani


Abstract
Recent advances in tabular deep learning have demonstrated the importance of embeddings for numerical features, where scalar values are mapped to high-dimensional spaces before being processed by the main model. Here, we propose an embedding method using the hyperbolic tangent (tanh) activation function that allows neural networks to achieve better accuracy on tabular data via an inductive bias similar to that of decision trees. To make training with the new embedding method reliable and efficient, we additionally propose a principled initialization method. Experiments demonstrate that the new approach improves upon or matches accuracy results from previously proposed embedding methods across multiple tabular datasets and model architectures.
Anthology ID:
2025.trl-1.18
Volume:
Proceedings of the 4th Table Representation Learning Workshop
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Shuaichen Chang, Madelon Hulsebos, Qian Liu, Wenhu Chen, Huan Sun
Venues:
TRL | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
208–216
Language:
URL:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.trl-1.18/
DOI:
Bibkey:
Cite (ACL):
Bingyan Liu, Charles Elkan, and Anil N. Hirani. 2025. Embeddings for Numerical Features Using tanh Activation. In Proceedings of the 4th Table Representation Learning Workshop, pages 208–216, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Embeddings for Numerical Features Using tanh Activation (Liu et al., TRL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.trl-1.18.pdf