AI Navigate

Beyond Weighted Summation: Learnable Nonlinear Aggregation Functions for Robust Artificial Neurons

arXiv cs.AI / 3/23/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that fixed linear aggregation is a limiting design for artificial neurons and proposes learnable nonlinear aggregation mechanisms (F-Mean and Gaussian Support) to improve robustness to noisy inputs.
  • It introduces hybrid neurons that interpolate between linear and nonlinear aggregation via a learnable blending parameter to preserve optimization stability during training.
  • On MLPs and CNNs with CIFAR-10 and a noisy CIFAR-10 variant, the hybrids show robustness gains under Gaussian noise, with three-way hybrids achieving up to 0.991 robustness versus 0.890 for the baseline.
  • The learned aggregation parameters converge to sub-linear aggregation (p around 0.43-0.50) and high novelty utilization (alpha around 0.69-0.79), suggesting neuron-level aggregation as a meaningful design dimension for noise-tolerant networks.

Abstract

Weighted summation has remained the default input aggregation mechanism in artificial neurons since the earliest neural network models. While computationally efficient, this design implicitly behaves like a mean-based estimator and is therefore sensitive to noisy or extreme inputs. This paper investigates whether replacing fixed linear aggregation with learnable nonlinear alternatives can improve neural network robustness without sacrificing trainability. Two differentiable aggregation mechanisms are introduced: an F-Mean neuron based on a learnable power-weighted aggregation rule, and a Gaussian Support neuron based on distance-aware affinity weighting. To preserve the optimisation stability of standard neurons, hybrid neurons are proposed that interpolate between linear and nonlinear aggregation through a learnable blending parameter. Evaluated in multilayer perceptrons and convolutional neural networks on CIFAR-10 and a noisy CIFAR-10 variant with additive Gaussian corruption, hybrid neurons consistently improve robustness under noise while F-Mean hybrids also yield modest gains on clean data. The three-way hybrid achieves robustness scores of up to 0.991 compared to 0.890 for the standard baseline, and learned parameters converge consistently to sub-linear aggregation (p \approx 0.43--0.50) and high novelty utilisation (\alpha \approx 0.69--0.79). These findings suggest that neuron-level aggregation is a meaningful and underexplored design dimension for building more noise-tolerant neural networks.