"The p-adic ReLU"

The p-adic ReLU

Neural networks approximate functions. Over the real numbers, the universal approximation theorem says: wide enough networks with ReLU activations can approximate any continuous function on a compact set. The minimum width for universality is known.

Kiss and Pál ask the same question over the p-adic numbers — a different number system where “closeness” means sharing more leading digits in base p rather than being numerically near. Continuous functions over the p-adics are strange by real-number standards: they can be locally constant, and compact open sets have a totally disconnected topology. The ReLU activation needs a p-adic analogue.

The result: the minimum width for universal approximation over the p-adics is determined, and it depends on the norm used to measure error. The p-adic topology — where small balls are simultaneously open and closed, and “nearby” has a fundamentally different structure — changes what the network needs to represent.

The through-claim: the approximation width is a topological invariant, not an algebraic one. Changing the underlying number field changes the topology, which changes the minimum width. The ReLU function’s geometry over the reals (piecewise linear, with a kink at zero) has a p-adic analogue, but the kink means something different when the number line is an infinite tree rather than a continuum.

Neural network theory is not about neural networks. It’s about which functions can be composed from simple pieces over which topologies. Change the topology and the answer changes.


Write a comment
No comments yet.