Documentation

Linglib.Theories.Pragmatics.InformationTheory.Channel

Communication Channels #

@cite{cover-thomas-2006} @cite{shannon-1948} @cite{zaslavsky-etal-2019}

A CommChannel C W is a finite-alphabet stochastic conditional distribution p(w | c) — the basic Shannon channel restricted to finite input/output types. Used as substrate by:

The derived quantities (marginalWord, posterior, commPrecision, mutualInfo) live here because they are channel-and-prior functions of purely Shannon character. Capacity-specific theorems live in the sibling ChannelCapacity.lean.

Main definitions #

structure Pragmatics.InformationTheory.CommChannel (C W : Type) [Fintype C] [Fintype W] :

A finite-alphabet communication channel: a row-stochastic conditional encode c w = p(w | c). The Shannon-channel primitive shared by information-theoretic and pragmatic communication models.

Originally NamingChannel in @cite{zaslavsky-etal-2019}; lifted here because the same primitive serves color-naming, lexicalization, asymmetric-lexicon models, and RSA literal-speaker semantics.

  • encode : CW

    p(w|c): probability of word w given meaning c.

  • encode_nonneg (c : C) (w : W) : 0 self.encode c w
  • encode_sum_one (c : C) : w : W, self.encode c w = 1
Instances For
    noncomputable def Pragmatics.InformationTheory.marginalWord {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (prior : C) (w : W) :

    Marginal word probability p(w) = Σ_c p(c) · p(w|c).

    Equations
    Instances For
      noncomputable def Pragmatics.InformationTheory.posterior {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (prior : C) (w : W) (c : C) :

      Posterior p(c|w) via Bayes' rule. The listener's belief about the meaning after hearing word w (≡ RSA literal listener L₀).

      Equations
      Instances For
        noncomputable def Pragmatics.InformationTheory.commPrecision {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (prior : C) (c : C) :

        Communicative precision (expected surprisal) of meaning c: S(c) = -Σ_w p(w|c) · log p(c|w). Lower means the channel communicates c more precisely. Defined in @cite{zaslavsky-etal-2019}.

        Equations
        Instances For
          noncomputable def Pragmatics.InformationTheory.mutualInfo {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (prior : C) :

          Mutual information I(W;C) = Σ_{c,w} p(c) · p(w|c) · log(p(c|w)/p(c)).

          Equations
          Instances For

            Basic structural lemmas #

            theorem Pragmatics.InformationTheory.CommChannel.encode_le_one {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (c : C) (w : W) :
            nc.encode c w 1

            Each encode probability is at most 1 (from row-stochastic constraint).

            theorem Pragmatics.InformationTheory.marginalWord_nonneg {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (prior : C) (hp : ∀ (c : C), 0 prior c) (w : W) :
            0 marginalWord nc prior w

            Marginal word probability is non-negative under a non-negative prior.

            theorem Pragmatics.InformationTheory.marginalWord_sum_one {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (prior : C) (hsum : c : C, prior c = 1) :
            w : W, marginalWord nc prior w = 1

            The marginal word distribution sums to 1 under a normalized prior.

            theorem Pragmatics.InformationTheory.marginalWord_pos_of {C W : Type} [Fintype C] [Fintype W] (nc : CommChannel C W) (prior : C) (hp : ∀ (c : C), 0 prior c) {c : C} {w : W} (hpc : 0 < prior c) (hew : 0 < nc.encode c w) :
            0 < marginalWord nc prior w

            When prior c > 0 and encode c w > 0, the marginal p(w) > 0.