Clifford-KAN

Clifford-KAN

Feb 15, 2025 · 3 min read

Mathematical Foundations of Clifford-KANs

This technical note details the formal integration of Clifford Algebra $\mathcal{C}\ell_{p,q}$ into the Kolmogorov-Arnold Network (KAN) architecture.

1. The Kolmogorov-Arnold Representation Theorem

The CKAN architecture is rooted in the Kolmogorov-Arnold Representation Theorem, which states that any multivariate continuous function $f: [0,1]^n \to \mathbb{R}$ can be represented as:

$$f(x_1, \dots, x_n) = \sum_{q=1}^{2n+1} \Phi_q \left( \sum_{p=1}^n \phi_{q,p}(x_p) \right)$$

Where:

  • $\phi_{q,p}$ are $n(2n+1)$ univariate continuous functions (inner functions).
  • $\Phi_q$ are $2n+1$ univariate continuous functions (outer functions).

In a KAN, these functions are parameterized as learnable B-splines.


2. Clifford Algebra $\mathcal{C}\ell_{p,q,r}$

A Clifford Algebra is an associative algebra generated by a vector space $V$ with a quadratic form $Q$. For a basis $\{e_1, e_2, \dots, e_d\}$, the fundamental geometric product is governed by the relation:

$$e_i e_j + e_j e_i = 2 \eta_{ij} \mathbb{1}$$

Where $\eta_{ij}$ is the metric tensor of signature $(p, q,r)$. A general multivector $u \in \mathcal{C}\ell_{p,q,r}$ is a linear combination of $2^d$ basis elements:

$$u = \underbrace{u_{\emptyset}}_{\text{scalar}} + \underbrace{\sum u_i e_i}_{\text{vectors}} + \underbrace{\sum u_{ij} e_{ij}}_{\text{bivectors}} + \dots + \underbrace{u_{12\dots d} e_{12\dots d}}_{\text{pseudoscalar}}$$

3. The Clifford-KAN Formulation

The CKAN extends the scalar KAN by allowing the learnable functions to operate on multivectors. There are two primary ways this is mathematically achieved in the paper:

A. Component-wise (Grade-Specific) Splines

The multivector $u$ is decomposed into its $2^d$ scalar components $u_A$. The edge transformation $\Psi$ is defined as:

$$\Psi(u) = \sum_{A \in \mathcal{I}} \phi_A(u_A) \mathbf{e}_A$$

where $\phi_A$ are independent learnable univariate B-splines for each basis blade $\mathbf{e}_A$. This ensures that the network can learn different behaviors for different geometric grades (e.g., treating scalars differently from bivectors).

B. Clifford-Valued Functions

For a more coupled approach, the univariate function is extended to the Clifford domain using a series expansion or a grade-preserving mapping:

$$\Phi(u) = \sum_{k=0}^K w_k u^k$$

where $u^k$ is computed using the Clifford geometric product, allowing for non-linear interactions between different blades (e.g., a vector squared becoming a scalar).


4. Forward Propagation in CKAN Layers

For a layer with $N_{in}$ input multivectors and $N_{out}$ output multivectors, the $j$-th output multivector $u_{out,j}$ is calculated as:

$$u_{out,j} = \text{Act} \left( \sum_{i=1}^{N_{in}} \Psi_{i,j}(u_{in,i}) \right)$$

Where:

  • $\Psi_{i,j}$ is the learnable Clifford-valued spline function on the edge.
  • The summation is performed using Clifford addition.
  • Act is a Clifford-equivariant activation function (often a grade-wise ReLU or a sigmoid).

5. Mathematical Summary

FeatureScalar KANClifford-KAN (CKAN)
Domain$\mathbb{R}$$\mathcal{C}\ell_{p,q,r}$ (Multivectors)
Edge OpUnivariate Spline $\phi(x)$Multivector Spline $\Psi(u)$
InteractionScalar AdditionClifford Geometric Product

Reference: Wolff, M., Alesiani, F., Duhme, C., & Jiang, X. (2026). Clifford Kolmogorov-Arnold Networks. arXiv:2602.05977.

@article{wolff2026clifford,
  title={Clifford Kolmogorov-Arnold Networks},
  author={Wolff, M. and Alesiani, F. and Duhme, C. and Jiang, X.},
  journal={arXiv preprint arXiv:2602.05977},
  year={2026},
  url={https://arxiv.org/abs/2602.05977},
  archivePrefix={arXiv},
  eprint={2602.05977},
  primaryClass={cs.LG}
}