Documentation

Linglib.Theories.Processing.Lexical.Discriminative.Normed

DLM in normed carriers — Lipschitz continuity and approximate isomorphism #

@cite{baayen-2019} @cite{chuang-bell-tseng-baayen-2026} @cite{lu-chuang-baayen-2026} @cite{heitmeier-chuang-baayen-2026}

Sibling to Defs.lean. Introduces the normed-carrier layer of the DLM substrate: when the form and meaning carriers are finite-dimensional real normed vector spaces, the production and comprehension maps are automatically continuous (mathlib's LinearMap.toContinuousLinearMap), and we can talk about their operator norms, Lipschitz constants, and approximate-inverse properties.

Why this is the substantive substrate layer #

The Defs.lean substrate establishes only architectural capacity — DLMs can distinguish meanings (via linear_dlm_distinguishes_meanings) and DLMs can neutralize them (via dlm_neutralizes_meanings_in_kernel). Both theorems are about exact equality / inequality of form vectors.

The headline empirical claim of @cite{chuang-bell-tseng-baayen-2026} (predictions iii / iv) and @cite{lu-chuang-baayen-2026} (paper §4 form- meaning isomorphism) is quantitative: that the form space and meaning space exhibit measurable isomorphism — cosine similarity 0.93, Pearson correlation 0.98 between GAMM-derived gold-standard contours and DLM-predicted contours (Lu §4.4). That quantitative content needs norms to state.

This file provides the structural primitives:

Carrier requirements #

Theorems are polymorphic over normed carriers F, M with [NormedAddCommGroup] + [NormedSpace ℝ ·] instances. Lipschitz theorems additionally require the source of the relevant linear map to be [FiniteDimensional ℝ ·] (so mathlib's toContinuousLinearMap applies). For our concrete consumer carriers:

Choice of carrier determines which norm appears in the operator-norm bound. Studies that need the L² (Euclidean) norm — e.g. for cosine- similarity statements — should use EuclideanSpace. The bare-Pi sup norm suffices for direction-of-effect arguments (kernel proximity → contour proximity).

theorem Theories.Processing.Lexical.Discriminative.dlm_lipschitz_production {F : Type u_1} {M : Type u_2} [NormedAddCommGroup F] [NormedAddCommGroup M] [NormedSpace F] [NormedSpace M] [FiniteDimensional M] (D : LinearDiscriminativeLexicon F M) (e₁ e₂ : M) :
D.production e₁ - D.production e₂ LinearMap.toContinuousLinearMap D.production * e₁ - e₂

Lipschitz continuity of production. For any LinearDiscriminativeLexicon over real finite-dimensional normed carriers, the production map sends meaning differences to form differences bounded by ‖production‖ * ‖meaning difference‖.

This is the headline normed-substrate theorem. It captures structurally what the consumer papers report empirically: that "centroid neighbours yield contour neighbours" in the trained DLM. The empirical claim is that the operator norm of the trained production map is small enough to make this bound informative on real data; the structural capacity for the bound is established here.

Derived from map_sub + ContinuousLinearMap.le_opNorm.

theorem Theories.Processing.Lexical.Discriminative.dlm_lipschitz_comprehension {F : Type u_1} {M : Type u_2} [NormedAddCommGroup F] [NormedAddCommGroup M] [NormedSpace F] [NormedSpace M] [FiniteDimensional F] (D : LinearDiscriminativeLexicon F M) (f₁ f₂ : F) :
D.comprehension f₁ - D.comprehension f₂ LinearMap.toContinuousLinearMap D.comprehension * f₁ - f₂

Lipschitz continuity of comprehension — dual of dlm_lipschitz_production for the form → meaning direction.

theorem Theories.Processing.Lexical.Discriminative.dlm_neighbor_centroids_imply_neighbor_contours {F : Type u_1} {M : Type u_2} [NormedAddCommGroup F] [NormedAddCommGroup M] [NormedSpace F] [NormedSpace M] [FiniteDimensional M] (D : LinearDiscriminativeLexicon F M) {e₁ e₂ : M} {ε : } (h : e₁ - e₂ ε) :
D.production e₁ - D.production e₂ LinearMap.toContinuousLinearMap D.production * ε

Neighbor centroids → neighbor contours. Quantitative form of the empirical observation in @cite{chuang-bell-tseng-baayen-2026} Fig. 18 and @cite{lu-chuang-baayen-2026} Fig. 9 that the DLM- predicted pitch contour from a tone-pattern centroid CE matches the GAMM-derived gold-standard contour for that pattern.

The structural form: if two meaning vectors are within ε of each other, their production-images are within ‖production‖ * ε of each other. The empirical content of the consumer papers is that this bound is small for real centroid pairs because the trained production map's operator norm is moderate.

Specialises dlm_lipschitz_production to a fixed bound.

def Theories.Processing.Lexical.Discriminative.LinearDiscriminativeLexicon.IsMeaningApproxIso {F : Type u_1} {M : Type u_2} [NormedAddCommGroup F] [NormedAddCommGroup M] [NormedSpace F] [NormedSpace M] (D : LinearDiscriminativeLexicon F M) (ε : ) :

Meaning-side approximate isomorphism. The DLM is an ε-approximate isomorphism on the meaning side if, for every meaning vector, the round-trip comprehension ∘ production returns within ε of the input.

The empirical content of @cite{chuang-bell-tseng-baayen-2026} predictions (iii) and (iv) — that the trained DLM predicts meaning from pitch above chance and pitch from meaning above chance — is the existence of a small ε for which the trained DLM satisfies this property. The structural definition is here; the empirical witness is paper-supplied.

Equations
Instances For
    def Theories.Processing.Lexical.Discriminative.LinearDiscriminativeLexicon.IsFormApproxIso {F : Type u_1} {M : Type u_2} [NormedAddCommGroup F] [NormedAddCommGroup M] [NormedSpace F] [NormedSpace M] (D : LinearDiscriminativeLexicon F M) (ε : ) :

    Form-side approximate isomorphism. Dual to IsMeaningApproxIso: the round-trip production ∘ comprehension returns within ε of the input form vector.

    Equations
    Instances For
      theorem Theories.Processing.Lexical.Discriminative.LinearDiscriminativeLexicon.isMeaningApproxIso_zero_iff {F : Type u_1} {M : Type u_2} [NormedAddCommGroup F] [NormedAddCommGroup M] [NormedSpace F] [NormedSpace M] (D : LinearDiscriminativeLexicon F M) :
      D.IsMeaningApproxIso 0 ∀ (e : M), D.comprehension (D.production e) = e

      The ε = 0 case of IsMeaningApproxIso is exact isomorphism on the meaning side: comprehension is a left inverse of production.