Quantitative Approximation Rates for Group Equivariant Learning
arXiv:2602.20370v1 Announce Type: new Abstract: The universal approximation theorem establishes that neural networks can approximate any continuous function on a compact set. Later works in approximation theory provide quantitative approximation rates for ReLU networks on the class of $alpha$-H”older functions $f: [0,1]^N to mathbb{R}$. The goal of this paper is to provide similar quantitative approximation results in the context of group equivariant learning, where the learned $alpha$-H”older function is known to obey certain group symmetries. While there has […]