PulseAugur
LIVE 06:59:51
research · [1 source] ·
0
research

New GEM activation functions offer smoother, rational alternatives to ReLU

Researchers have introduced Geometric Monomial (GEM), a new family of activation functions designed for deep neural networks. These functions utilize purely rational arithmetic and offer $C^{2N}$-smoothness, aiming to improve upon the limitations of the standard ReLU. Experiments show that GEM variants can match or exceed the performance of established functions like GELU on various benchmarks, including CIFAR-10, CIFAR-100, MNIST, GPT-2, and BERT-small. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT Introduces a new activation function family that shows competitive performance against GELU on various benchmarks, potentially improving deep learning model optimization.

RANK_REASON This is a research paper introducing a novel activation function family with experimental results.

Read on arXiv cs.LG →

New GEM activation functions offer smoother, rational alternatives to ReLU

COVERAGE [1]

  1. arXiv cs.LG TIER_1 · Eylon E. Krause ·

    Geometric Monomial (GEM): a family of rational 2N-differentiable activation functions

    The choice of activation function plays a crucial role in the optimization and performance of deep neural networks. While the Rectified Linear Unit (ReLU) remains the dominant choice due to its simplicity and effectiveness, its lack of smoothness may hinder gradient-based optimiz…