TLDR: v-PuNNs are a new type of neural network that uses p-adic numbers to represent hierarchical data (like family trees or file systems) more accurately and transparently than traditional methods. They achieve state-of-the-art accuracy on benchmarks like WordNet and Gene Ontology, are highly efficient (CPU-only), and offer clear interpretability. They also have applications beyond classification, such as creating structural invariants and enabling controllable data generation.
In the realm of artificial intelligence, deep learning models have achieved remarkable success by embedding data into Euclidean spaces. However, a significant portion of real-world information, such as biological taxonomies, language structures, and file systems, is inherently hierarchical. Forcing these strictly nested hierarchies into a flat Euclidean space often leads to distortions and a loss of interpretable meaning.
Addressing this fundamental geometric mismatch, a groundbreaking new architecture called van der Put Neural Networks, or v-PuNNs, has been introduced. Developed by Gnankan Landry Regis N’guessan, v-PuNNs are the first neural networks designed to operate natively in ultrametric p-adic space. This unique approach allows for a more natural and accurate representation of hierarchical data, where the distance between two points is determined by the depth of their lowest common ancestor.
The core principle behind v-PuNNs is Transparent Ultrametric Representation Learning (TURL). This means that every weight in the v-PuNN model is a p-adic number, providing exact subtree semantics. Unlike conventional “black-box” deep learning models, v-PuNNs are “white-box” models, offering clear and interpretable insights into how they process hierarchical information. This transparency is a significant leap forward for interpretable AI.
A key theoretical contribution of this work is the new Finite Hierarchical Approximation Theorem. This theorem proves that a v-PuNN of a certain depth can universally approximate any function on a K-level tree, demonstrating its expressive power with a manageable number of neurons.
Given the discrete nature of p-adic space, traditional gradient-based optimization methods are not suitable as gradients tend to vanish. To overcome this, the researchers introduced a novel optimization technique called Valuation-Adaptive Perturbation Optimization (VAPO). VAPO comes in two variants: a fast deterministic version called GIST-VAPO and a moment-based one called Adam-VAPO, both designed to efficiently navigate this discrete space.
The practical implementation of v-PuNNs, known as Hierarchically-Interpretable p-adic Network (HiPaN), has set new state-of-the-art results across three canonical benchmarks. On WordNet nouns, a large dataset of over 52,000 leaves, HiPaN achieved an impressive 99.96% leaf accuracy in under 17 minutes. For Gene Ontology molecular function, it attained 96.9% leaf and 100% root accuracy in just 50 seconds. Furthermore, on NCBI Mammalia, the learned metric showed a strong correlation with ground-truth taxonomic distance, surpassing all existing Euclidean and tree-aware baselines. Crucially, the learned metric was perfectly ultrametric, with zero triangle violations, confirming its structural fidelity.
The efficiency of v-PuNNs is particularly noteworthy. The models are CPU-only, requiring significantly less computational power and memory compared to GPU-centric approaches like graph transformers. For instance, training WordNet-19 took only 16 minutes on a single CPU core with 12 MB of RAM, demonstrating an order of magnitude lighter footprint than comparable accuracy models.
Beyond classification, v-PuNNs prove to be versatile scientific instruments. HiPaQ, a variant, can turn symbolic hierarchies into canonical structural invariants, useful for fields like algebra and quantum mechanics. Another variant, Tab-HiPaN, can discover latent hierarchies within flat tabular data, providing an interpretable control knob for conditional data generation. For example, in a wine quality dataset, Tab-HiPaN could identify underlying chemical variations and allow for the generation of “twin” wines by simply adjusting a few p-adic digits.
Also Read:
- Granular Concept Circuits: Illuminating Fine-Grained Concepts in Deep Vision Models
- SpikeSTAG: Unlocking Spatial-Temporal Insights in Complex Data
This work bridges number theory and deep learning, offering exact, interpretable, and efficient models for hierarchical data. The ability of v-PuNNs to match model geometry to data geometry leads to tangible gains in accuracy, interpretability, and resource usage. This opens up new avenues for transparent reasoning in diverse domains such as knowledge graphs, program analysis, and beyond. For more in-depth information, you can refer to the full research paper.


