@cite{dong-etal-2026} on mathlib PMF (binary identification) #
@cite{dong-etal-2026}
PMF-shaped formalisation of the paper's 4 findings on binary identification. The task is intentionally degenerate: at any target, exactly one utterance applies (matching pair).
All 4 findings are vacuous-zero: the wrong guess has L0 = 0, so its S1/L1 mass is 0, hence "correct > wrong".
Stakes parameter k (animal: 1, medical: 10) is irrelevant to the
qualitative predictions because the L0 gate zeros incorrect responses.
For the simplest PMF formulation, we set α = 1 and use a unit cost factor
(no stakes-dependence). Both bundled-API configs (animalCfg, medicalCfg)
in the paper collapse to the same PMF model after qualitative-only migration.
Binary identification task (simplified from the paper's Mixed-Stakes 20 Questions, which uses 100 animals / 15 diseases).
Instances For
Equations
- DongEtAl2026.PMF.instDecidableEqTarget x✝ y✝ = if h : x✝.ctorIdx = y✝.ctorIdx then isTrue ⋯ else isFalse ⋯
Equations
- DongEtAl2026.PMF.instReprTarget = { reprPrec := DongEtAl2026.PMF.instReprTarget.repr }
Equations
- One or more equations did not get rendered due to their size.
Instances For
Equations
- One or more equations did not get rendered due to their size.
Boolean match: does guess match target?
Equations
Instances For
§1. L0 — uniform on extension via RSA.L0OfBoolMeaning #
Instances For
Equations
Instances For
§2. S1 — speaker as PMF.normalize of L0 #
Pragmatic speaker: each world has at least one matching utterance, so no fallback is needed (every world is non-degenerate in the binary task).
Equations
- DongEtAl2026.PMF.S1 w = PMF.normalize (fun (u : DongEtAl2026.PMF.Target) => (DongEtAl2026.PMF.L0 u) w) ⋯ ⋯
Instances For
§3. L1 — Bayesian inversion against the uniform world prior #
Equations
- DongEtAl2026.PMF.worldPrior = PMF.uniformOfFintype DongEtAl2026.PMF.Target
Instances For
Equations
Instances For
§4. Findings — all vacuous-zero #
S1 prefers correct guess: at world .t₁, utterance .t₂ has L0 = 0 (targetMatches .t₂ .t₁ = false), so S1 .t₁ .t₂ = 0.
Same-base comparison (same world .t₁), different utterances → use
normalize_lt_iff_lt to reduce to L0 score comparison.