PulseAugur
LIVE 13:04:56
tool · [1 source] ·
0
tool

PolyStep optimizer trains non-differentiable neural networks using optimal transport

Researchers have developed PolyStep, a novel gradient-free optimizer designed to train neural networks with non-differentiable components. This method bypasses the need for backpropagation and surrogate gradients by evaluating loss at polytope vertices in a compressed subspace and using softmax-weighted assignments. PolyStep demonstrates superior performance on various non-differentiable architectures, including spiking networks and quantized layers, outperforming existing gradient-free methods and approaching the accuracy of gradient-based approaches. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT Introduces a new optimization technique for training complex neural network architectures that were previously difficult to optimize.

RANK_REASON This is a research paper detailing a new method for training non-differentiable neural networks. [lever_c_demoted from research: ic=1 ai=1.0]

Read on arXiv cs.LG →

COVERAGE [1]

  1. arXiv cs.LG TIER_1 · An T. Le ·

    Training Non-Differentiable Networks via Optimal Transport

    arXiv:2605.01928v1 Announce Type: new Abstract: Neural networks increasingly embed non-differentiable components (spiking neurons, quantized layers, discrete routing, blackbox simulators, etc.) where backpropagation is inapplicable and surrogate gradients introduce bias. We prese…