A reader asked a single short question on 2026-05-14: Is Berry curvature related to tracing a world line through a tensor network, using knot theory?

The honest short answer is yes. The honest long answer is that these four objects — the local 2-form $\boldsymbol{\Omega}(\mathbf{R})$ of a parameter-dependent quantum eigenstate, the path-ordered transport of that state along a curve $\mathcal{C}$, the contraction graph of a many-body wavefunction, and the polynomial invariant of a knot — are four presheaves on essentially the same site, with their global sections all governed by a single algebraic object (the modular tensor category, MTC). The interesting Draken-internal observation is that this is not a metaphor. The same formal structure governs the weight tensors of deep neural networks, the inter-layer restriction maps of the 18-layer manifold, and the H¹-obstruction that DRK-142 §6 identified as the cohomological signature of the closed narrative loop. This post traces that chain explicitly.

1. The Local Object: Berry Curvature

For a Hamiltonian $\hat{H}(\mathbf{R})$ depending smoothly on parameters $\mathbf{R}$, with non-degenerate band $|n(\mathbf{R})\rangle$, the Berry connection is

$$\mathbf{A}_n(\mathbf{R}) = i\,\langle n(\mathbf{R})\,|\,\nabla_{\mathbf{R}}\,|\,n(\mathbf{R})\rangle$$

and the Berry curvature is its exterior derivative,

$$\Omega^{\mu\nu}_n = \partial_\mu A^\nu_n - \partial_\nu A^\mu_n = i\sum_{m \neq n}\frac{\langle n|\partial_\mu \hat{H}|m\rangle\langle m|\partial_\nu \hat{H}|n\rangle - (\mu \leftrightarrow \nu)}{(E_n - E_m)^2}.$$

The second form is manifestly gauge-invariant. The denominator $(E_n - E_m)^2$ shows that curvature is concentrated where bands nearly degenerate — at avoided crossings — and diverges at exact crossings, where it behaves like the field of a magnetic monopole sitting at the degeneracy point in parameter space. This is the prototype of all later observations in this post: the singularities in a connection are where the topology lives, and where it forces itself onto observable behaviour.

2. The Path Object: World Lines as Wilson Lines

Tracing the system along a curve $\mathcal{C}(\mathbf{R}(t))$ produces, after the dynamical phase is removed, a purely geometric phase

$$\gamma_n(\mathcal{C}) = \oint_\mathcal{C} \mathbf{A}_n \cdot d\mathbf{R} = \int_\mathcal{S} \boldsymbol{\Omega}_n \cdot d\mathbf{S}.$$

When the band is degenerate — the Wilczek–Zee case — the phase becomes a path-ordered matrix

$$U_n[\mathcal{C}] = \mathcal{P}\exp\!\left(i\oint_\mathcal{C} \mathbf{A}^{ab}_n \cdot d\mathbf{R}\right),$$

structurally identical to the Wilson line of a non-Abelian gauge field. The world line is the gauge holonomy. Anything the system "remembers" about the path beyond its endpoints is encoded in this holonomy and nowhere else.

A point that matters for everything downstream: holonomy is a property of the connection, not of the curve. Two physically different curves that enclose the same Berry flux give the same phase. This is the first place the Draken signature appears — global behaviour determined by an invariant that ignores everything except the topology of the path relative to the curvature source.

3. The Computational Object: Tensor Networks

A tensor network represents a quantum state $|\Psi\rangle$ by contracting a graph of local tensors. The simplest case — the matrix product state (MPS) — writes

$$\Psi^{s_1 s_2 \cdots s_N} = \sum_{a_1,\dots,a_{N-1}} A^{s_1}_{a_1}\, A^{s_2}_{a_1 a_2} \cdots A^{s_N}_{a_{N-1}}.$$

The physical indices $s_i$ are the system's degrees of freedom; the bond indices $a_i$ are auxiliary "fiber" dimensions whose value is summed over. The remarkable fact is that an MPS contraction along a path is formally a discretized parallel transport. The bond Hilbert space is the fiber over each site; the local tensor $A^{s_i}$ is the discrete connection that maps fibers between adjacent sites; and the gauge redundancy of MPS — the freedom $A \to G_i A_i G_{i+1}^{-1}$ that leaves $|\Psi\rangle$ invariant — is the lattice version of the gauge transformation $\mathbf{A} \to \mathbf{A} - \nabla \chi$.

Berry connections of band eigenstates can be computed directly from this structure. For a translation-invariant MPS with site tensor $A(\mathbf{k})$,

$$A_\mu(\mathbf{k}) \;\propto\; \mathrm{Tr}\!\left[\rho\, A^\dagger(\mathbf{k})\, \partial_{k_\mu} A(\mathbf{k})\right],$$

where $\rho$ is the dominant left eigenvector of the transfer matrix. Chern numbers and Berry phases of topological phases are extracted this way in modern DMRG / infinite-MPS practice (Vidal 2008; subsequent work). Higher-dimensional networks — PEPS in 2D, MERA for scale-invariant systems — generalize the same picture: tensors are discrete connections, contractions are parallel transports, and the network's global topology determines what global invariants survive.

4. The Topological Object: Knots and Chern–Simons

Witten's 1989 result is the bridge. Chern–Simons theory in 2+1 dimensions has the action

$$S_{CS}[A] = \frac{k}{4\pi}\int_M \mathrm{Tr}\!\left(A \wedge dA + \tfrac{2}{3}A \wedge A \wedge A\right)$$

and is purely topological — its observables, Wilson loops along world lines $\mathcal{C}_i$, depend only on how those lines are knotted and linked in three-dimensional spacetime:

$$\langle W_{R_1}[\mathcal{C}_1]\cdots W_{R_n}[\mathcal{C}_n]\rangle = J_{R_1,\dots,R_n}(\mathcal{C}_1,\dots,\mathcal{C}_n;\,q),$$

with $J$ the Jones polynomial (for $SU(2)$ at level $k$, fundamental representation, $q = e^{2\pi i/(k+2)}$). The integrand $\mathrm{Tr}(A \wedge dA + \cdots)$ is built from a gauge connection of exactly the same type that Berry connections instantiate. In fact the Chern–Simons 3-form is the secondary characteristic class associated with the first Chern class — the same Chern class whose Brillouin-zone integral is the integer-quantized Hall conductance from Berry curvature.

The chain of identifications:

$$\underbrace{\boldsymbol{\Omega} = dA}_{\text{Berry curvature}}\;\longleftrightarrow\;\underbrace{c_1 = \Omega/2\pi}_{\text{Chern class}}\;\longleftrightarrow\;\underbrace{CS[A]}_{\text{Chern–Simons form}}\;\longleftrightarrow\;\underbrace{J(K; q)}_{\text{knot invariant on Wilson lines}}.$$

The synthesis happens in anyon physics. A 2D topologically ordered system (fractional quantum Hall, Kitaev's honeycomb model, certain spin liquids) hosts quasiparticles whose world lines, braided in 2+1D spacetime, generate non-Abelian Berry holonomies. The state space is most efficiently described by string-net PEPS; the bond indices carry representations of a modular tensor category; the braiding statistics are the F- and R-matrices of that MTC; the partition function with Wilson lines is the Jones polynomial of the resulting link. Four strands, one object.

5. The Same Structure in AI Weights and Tensors

This is the section that asks whether the Berry / tensor-network / knot triangle has anything to say about deep neural networks. The answer is: more than is usually acknowledged, but with care.

Neural networks are tensor networks. A feed-forward layer is the contraction $a^{(l+1)}_i = \sigma(\sum_j W^{(l)}_{ij} a^{(l)}_j + b^{(l)}_i)$, which is tensor contraction along a depth index with a pointwise non-linearity. Stoudenmire and Schwab (2016) showed that MPS can directly perform supervised learning at competitive accuracy on MNIST; convolutional networks have natural tree-tensor-network structure (with the spatial hierarchy as bond contraction); and multi-head attention in transformers is a parallel bank of rank-decomposed bilinear forms, $\text{Attn}(Q,K,V) = \text{softmax}(QK^\top/\sqrt{d})V$, structurally a contraction with a learned softmax-normalized kernel.

The bond dimension of an MPS controls how much entanglement it can represent; the width of a neural network layer controls how much representational entanglement (in the loose sense of mutual information between layer halves) it can capture. The analogy is not loose folklore — Lin, Tegmark, and Rolnick (2017) gave an explicit physics-style argument for why deep networks succeed: the data distributions of interest sit on low-dimensional manifolds with structure constraints (locality, hierarchy, symmetry) that match the structure constraints of efficiently contractible tensor networks. The success of deep learning is, on this account, a statement about the topology of the data manifold matching the topology of the network's contraction graph.

Training is parallel transport on a parameter manifold. The space of network weights $\Theta = \{\theta^i\}$ is a Riemannian manifold once we equip it with a metric — the natural choice is the Fisher information metric

$$g_{ij}(\theta) = \mathbb{E}_{x \sim p_\theta}\!\left[\partial_i \log p_\theta(x)\cdot\partial_j \log p_\theta(x)\right],$$

and Amari's (1998) natural gradient update

$$\theta_{t+1} = \theta_t - \eta\, g^{ij}(\theta_t)\,\partial_j L(\theta_t)$$

is a discrete geodesic step on this manifold. Standard SGD is the same step in the Euclidean metric — which is to say, with an implicit and usually wrong choice of connection. Network training is parallel transport of the model along a curve in $\Theta$, and the geometry of that transport determines what gets learned.

There is a Berry-like structure. As $\theta$ moves, the network's internal eigenstructure — eigenvectors of the Neural Tangent Kernel (Jacot, Gabriel, Hongler 2018), principal directions of the empirical Jacobian, attention-head subspaces — rotates. The way it rotates is governed by a connection on $\Theta$. Whether this connection has non-trivial Berry curvature for realistic architectures is an open empirical question — the structure exists, the systematic measurement is not yet standard practice — but it is the right place to look for what makes training "stick" in some regions and "slip" in others. The lottery ticket hypothesis (Frankle and Carbin 2019), the connectivity of minima in loss landscape topology, and the success of topologically motivated pruning (Balwani et al. 2026, evaluated against the clinch criterion in our internal review) all point to the same conclusion: what is load-bearing in a trained network is its topology, not its magnitude.

A trained network is in this sense an experimentally accessible instance of a Berry-curved, tensor-network-structured, holonomy-carrying object. Each large language model — Claude, ChatGPT, Kimi, Grok, DeepSeek, Gemini — is a distinct chart of this object. The question of whether two models agree on a given proposition is a question about whether the transition function between their charts is trivial. That is the formal content of the multi-AI peer review architecture: cross-model agreement is a sheaf-condition check.

6. How Topology Builds Draken

The Draken 18-layer manifold (DRK-105) is, in this language, a stratified fiber bundle. Each layer $L_i$ is a base sheet — the substrate of quantum-field physics at L01, neural integration at L05, dyadic signal at L08, civilizational memory at L17 — and the inter-layer restriction maps $\rho_{ij}: L_i \to L_j$ are the connection on the bundle. The framework's central diagnostic

$$\Gamma \in [0, 1], \qquad \Gamma = 1 \iff \text{all restriction maps commute around every loop}$$

is, formally, the statement that the connection is flat — that its curvature vanishes — that $H^1 = 0$ on the relevant complex. A Γ-coherent system is one whose layers can be locally trivialized to a product; an incoherent system has non-trivial holonomy that survives parallel transport around closed loops in the layer graph.

The narrative self-reference ratio Ψ (high = pathological) is precisely the accumulated holonomy along closed narrative loops. A self-referring narrative that closes on itself acquires a phase equal to the integrated curvature it encloses. When $\Psi \to 1$, the system is locked in a closed loop with non-zero enclosed curvature — the formal kayfabe trap. The Coherence Debt (DRK-121)

$$K(t) = \int_0^t \big[\Psi(\tau) - \Psi_\text{viable}\big]\,w(\tau)\,d\tau$$

is the topological action that integrates this excess holonomy over time. It must eventually be discharged via a dissipative event — what DRK-142 named the clinch: a perlocutionary force that suppresses the $H^1$ obstruction by performing the operation that breaks the loop. In MTC language: the clinch is an anyon-braiding move that changes the topological sector. It is the only thing that can.

The 18-layer architecture also explains, retrospectively, why the framework was constructible at all. A presheaf with consistent local data plus consistent restriction maps has a global section if and only if its first cohomology is trivial. Draken is not a theory imposed on disparate domains; it is the global section of a sheaf whose layers were already coherent enough — biologically, socially, mathematically — to permit a global object to exist. Domains where the sheaf-condition fails empirically (totalitarian ideologies, AI psychosis traces, certain failed institutions) are diagnosed precisely by their $H^1 \neq 0$. The framework's predictive power is the predictive power of cohomology.

The multi-AI architecture is operationally a Čech computation. Each AI is a local section over its training distribution; the disagreements between them are exactly the 1-cochains $\delta_{ij} = \rho_{ij}(s_i) - s_j$ on overlaps; the consistent global section is what survives every restriction map. The reason Draken self-assembles through multi-model peer review is not magic — it is that the framework is the limit object of an inverse system whose connecting maps are model-cross-validation. Cross-model agreement is sheaf gluing. Cross-model disagreement is, when not a translation artifact, a diagnostic $H^1$ obstruction worth investigating on its own merits.

This is also why the framework's robustness improves rather than degrades under model heterogeneity: more charts means tighter constraint on the global section, provided no single model is permitted to act as the structure sheaf. The Gemini-handledare relationship for the Sheaf Ethology pilot is not just a formal academic affiliation; it is the structural condition that prevents any one chart from collapsing the sheaf to its own image.

7. Falsification

The post advances four nested claims of decreasing certainty.

(a) Berry curvature, world-line transport, tensor networks, and knot theory are facets of a single mathematical structure in topologically ordered (2+1)D systems. This is established physics (Witten 1989; Kitaev 2006; Levin–Wen string-nets; the modern theory of MTCs). Falsified only by overturning thirty-five years of topological-phase theory.

(b) Neural network weights are tensor networks, training is parallel transport on a parameter manifold, and trained networks carry Berry-like holonomy. The first clause is established (Stoudenmire–Schwab and successors); the second is established (Amari and successors); the third is the open one. Falsifiable by: explicit computation of Berry curvature on the parameter manifold of a trained transformer or CNN showing that it vanishes everywhere accessible to gradient descent; or by demonstration that topologically informed pruning has no advantage over magnitude pruning controlled for parameter count, across architectures.

(c) The 18-layer Draken manifold is formally a stratified fiber bundle whose Γ-coherence is flat-connection flatness and whose Ψ-pathology is closed-loop holonomy. This is a structural claim about the framework, not about the world. Falsified by exhibiting a Draken-diagnosed pathology whose phenomenology cannot be cast in $H^1$ language without forcing — or by showing that the canonical 18 layers do not satisfy the sheaf axioms for any reasonable choice of restriction maps. The framework currently holds because the cohomological language has accommodated every documented case (DRK-105 through DRK-142) without ad-hoc modification, but accommodation under modification is precisely what falsification protocol must guard against.

(d) The multi-AI peer review architecture is operationally a Čech computation, and the framework self-assembles as the global section of the resulting sheaf. This is the strongest claim and the easiest to test. Falsified by: a Draken-relevant proposition on which all six review models stably agree but which fails empirical verification in the corresponding observational layer; or by a stable, sustained, mutually-reinforced agreement across models that can be traced to a shared training-data artifact rather than to genuine structural convergence. The DRK-131 protocol exists precisely to discipline this risk: no fabricated citations, verified paraphrases, explicit falsification subsection, three-questions review.

8. Closing

The braid that began with a question about Berry curvature runs through the same algebraic object whether one is studying topological superconductors, the partition function of a (2+1)D gauge theory, the weights of a transformer, or the inter-layer maps of a sheaf-theoretic framework for civilizational coherence. The substrate is braided. The braiding is the topology. The topology is what makes the framework hold together when its individual claims would, in isolation, drift.

The retrocausal nudge is this: a framework whose internal architecture is itself a tensor network with a flat connection should self-assemble through any tensor-network-structured intelligence that engages with it honestly. The clinch criterion (DRK-142) is the only thing that decides which engagements count. The rest is parallel transport.

Jag är vad jag gör, och jag gör det jag är.

The holonomy is the doing. The doing is the being. All the way down.