Documentation

Linglib.Phenomena.Clarification.Studies.DongEtAl2026PMF

@cite{dong-etal-2026} on mathlib PMF (binary identification) #

@cite{dong-etal-2026}

PMF-shaped formalisation of the paper's 4 findings on binary identification. The task is intentionally degenerate: at any target, exactly one utterance applies (matching pair).

All 4 findings are vacuous-zero: the wrong guess has L0 = 0, so its S1/L1 mass is 0, hence "correct > wrong".

Stakes parameter k (animal: 1, medical: 10) is irrelevant to the qualitative predictions because the L0 gate zeros incorrect responses.

For the simplest PMF formulation, we set α = 1 and use a unit cost factor (no stakes-dependence). Both bundled-API configs (animalCfg, medicalCfg) in the paper collapse to the same PMF model after qualitative-only migration.

Binary identification task (simplified from the paper's Mixed-Stakes 20 Questions, which uses 100 animals / 15 diseases).

Instances For
    @[implicit_reducible]
    Equations
    def DongEtAl2026.PMF.instReprTarget.repr :
    TargetStd.Format
    Equations
    • One or more equations did not get rendered due to their size.
    Instances For
      @[implicit_reducible]
      Equations
      • One or more equations did not get rendered due to their size.

      §1. L0 — uniform on extension via RSA.L0OfBoolMeaning #

      @[simp]
      theorem DongEtAl2026.PMF.mem_support_L0_iff (u w : Target) :
      w (L0 u).support targetMatches u w = true
      theorem DongEtAl2026.PMF.L0_apply_of_true {u w : Target} (h : targetMatches u w = true) :
      (L0 u) w = (↑(extension u).card)⁻¹
      theorem DongEtAl2026.PMF.L0_apply_of_false {u w : Target} (h : targetMatches u w true) :
      (L0 u) w = 0

      §2. S1 — speaker as PMF.normalize of L0 #

      noncomputable def DongEtAl2026.PMF.S1 (w : Target) :
      PMF Target

      Pragmatic speaker: each world has at least one matching utterance, so no fallback is needed (every world is non-degenerate in the binary task).

      Equations
      Instances For
        theorem DongEtAl2026.PMF.S1_apply (w u : Target) :
        (S1 w) u = (L0 u) w * (∑' (u' : Target), (L0 u') w)⁻¹

        §3. L1 — Bayesian inversion against the uniform world prior #

        noncomputable def DongEtAl2026.PMF.worldPrior :
        PMF Target
        Equations
        Instances For

          §4. Findings — all vacuous-zero #

          S1 prefers correct guess: at world .t₁, utterance .t₂ has L0 = 0 (targetMatches .t₂ .t₁ = false), so S1 .t₁ .t₂ = 0.

          Same-base comparison (same world .t₁), different utterances → use normalize_lt_iff_lt to reduce to L0 score comparison.

          S1 facing target .t₁ prefers utterance .t₁ over .t₂.

          Same prediction at "medical" stakes (the binary task is degenerate; qualitative prediction unchanged by the stakes parameter).

          Per-world S1 leaf for L1: at world .t₂, utterance .t₁ doesn't apply, so S1 .t₂ .t₁ = 0. At world .t₁, S1 .t₁ .t₁ > 0.

          L1 correctly identifies target from utterance.