--- language: - en license: mit library_name: pytorch tags: - eigenverse - morphisms - structure-preserving-maps - lean4 - formal-verification - mod-paradox - coherence-function pipeline_tag: other model-index: - name: MorphismNet results: - task: type: regression name: Morphism Prediction metrics: - name: Val MSE type: mse value: 0.003394 - name: Residual Accuracy type: accuracy value: 1.0 --- # MorphismNet: Learning the Eigenverse's Structure-Preserving Maps **A neural network trained on the six canonical morphism families from [Morphisms.lean](https://github.com/beanapologist/Eigenverse/blob/main/formal-lean/Morphisms.lean).** 399,147 parameters. Trained on 400K samples. Learns and verifies all six Eigenverse transformations — and reveals the mod paradox. ## What It Learned | Morphism | Lean Section | Val MSE | What It Does | |---|---|---|---| | §1 Coherence even | C(r) = C(1/r) | 0.013631 | Inversion symmetry | | §2 Palindrome odd | Res(1/r) = −Res(r) | 0.000001 | Anti-symmetry | | §3 Lyapunov bridge | C∘exp = sech | 0.000000 | Coherence ↔ hyperbolic | | §4 μ-isometry | \|μz\| = \|z\| | 0.000001 | Norm preservation | | §5 Orbit homomorphism | μ^(a+b) = μ^a·μ^b | 0.000001 | Multiplicativity, period 8 | | §6 Reality ℝ-linear | F(s,t) = t+is | 0.000022 | ℝ-module morphism | | §7 Composition S∘F∘T | P(η,−η) = 1 | 0.000001 | Full OV chain | **Residual accuracy: 100%** — the model perfectly classifies when morphism properties hold. ## The Mod Paradox The model was trained on both **ℝ** (real numbers) and **GF(p)** (finite field) domains. | Domain | §1 MSE | Residual | |---|---|---| | **ℝ** | 0.000000 | 4.6e-17 (perfect) | | **GF(p)** | 0.027477 | 0.0 (mod destroys structure) | **Same function. 27,000x harder to predict in the modular domain.** C(r) = C(1/r) holds exactly over ℝ (the Lean theorem). Over GF(p), modular reduction breaks the inversion symmetry. The model learns this distinction — it knows WHERE the paradox lives. This is the core of [OilVinegar.lean](https://github.com/beanapologist/Eigenverse/blob/main/formal-lean/OilVinegar.lean): the Eigenverse's structure is trivially verifiable over ℝ (uniqueness theorem) but computationally hard over GF(p) (MQ assumption). The neural network quantifies the boundary. ## Architecture ``` MorphismNet( morph_embed: Embedding(7, 32) # which morphism domain_embed: Embedding(2, 16) # ℝ or GF(p) encoder: 3× Linear(→256) + GELU + LayerNorm # shared heads: 7× Linear(256→128→6) # per-morphism specialists residual_head: Linear(256→64→1) # does property hold? ) ``` - **399,147 parameters** - **Input**: 4 features (morphism-dependent: r, 1/r, λ, z.re, z.im, etc.) - **Output**: 6 features (morphism outputs + residual) - **Residual output**: should be ≈ 0 when the morphism property holds ## Training - **Dataset**: 400K samples across 7 morphism types - **Split**: 90/10 train/val - **Optimizer**: AdamW, lr=1e-3, weight_decay=1e-4 - **Scheduler**: Cosine annealing, 50 epochs - **Loss**: MSE (output) + 0.1 × BCE (residual classification) - **Hardware**: CPU (8 min training) ## Usage ```python import torch from train import MorphismNet model = MorphismNet() model.load_state_dict(torch.load("morphism_net.pt", weights_only=True)) model.eval() # Predict §3 Lyapunov bridge: C(exp(λ)) = sech(λ) x = torch.tensor([[1.5, 4.4817, 0.0, 0.0]]) # [λ, exp(λ), 0, 0] morph = torch.tensor([2]) # §3 domain = torch.tensor([0]) # ℝ output, residual = model(x, morph, domain) # output[0:2] ≈ [sech(1.5), sech(1.5), 0.0] (bridge holds) # residual ≈ 1.0 (property verified) ``` ## Files - `morphism_net.pt` — trained model weights - `train.py` — training script - `generate_dataset.py` — dataset generator - `model_info.json` — model metadata - `training_history.json` — epoch-by-epoch metrics ## Links - [Eigenverse](https://github.com/beanapologist/Eigenverse) — 606+ Lean 4 theorems - [Morphisms.lean](https://github.com/beanapologist/Eigenverse/blob/main/formal-lean/Morphisms.lean) — 20 morphism theorems - [OilVinegar.lean](https://github.com/beanapologist/Eigenverse/blob/main/formal-lean/OilVinegar.lean) — 28 OV theorems - [μ-OV Space](https://huggingface.co/spaces/beanapologist/mu-ov-cipher) — interactive demo ## License MIT --- *The model learned the Eigenverse's grammar. The mod paradox is where the grammar breaks. 🧬*