[Submitted on 28 Oct 2025]
Adaptive Range SiLU: An Improved Activation Function
View PDFAbstract:We present Adaptive Range SiLU (AR-SiLU), a novel activation function that introduces learnable range expansion and temperature scaling for Transformer feedforward networks. Experiments show AR-SiLU achieves a 0.4\% reduction in validation perplexity compared to SwiGLU baselines (p < 0.05) while maintaining computational efficiency.
Submission history
[v1] Tue, 28 Oct 2025 23:44 UTC