Documentation

Linglib.Phenomena.Gradability.Studies.LassiterGoodman2017PMF

@cite{lassiter-goodman-2017} on mathlib PMF — structural shell #

@cite{lassiter-goodman-2017}

L&G 2017 ("Adjectival vagueness in a Bayesian model of interpretation", Synthese 194:3801-3836) gives a Bayesian/RSA account of vague gradable adjectives. This file formalises the structural skeleton of that account on mathlib's PMF, with explicit acknowledgment of what the file does and does not capture.

Scope (honest reckoning, post-audit) #

L&G 2017's novel contribution is the free-variable inference of §4 (Eqs. 26-29): the L1 listener jointly infers the world A and the threshold variable V, with marginalisation over thresholds (Eq. 30) giving the height posterior. The paper's central empirical claims — information transmission for vague terms (§4.3, Figs. 5-6), context-sensitivity via prior shape (§4.4, Fig. 7, "tall building" vs. "tall person"), antonym symmetry (Fig. 6), and the MC/PC/free-variable sorites trichotomy (§5, Eqs. 36/38/41-44) — all live in that joint posterior.

What this file captures:

  1. §3.5 (NEW): The joint (world × threshold) posterior architecture (L&G Eqs. 26-30) — the paper's novel contribution. Built on PMF.posterior instantiated at product type α := W × T, with marginalization formulas via PMF.posterior_fst_apply / posterior_snd_apply from Core/Probability/JointPosterior.lean. worldPosterior_apply is L&G's Eq. 30 (height marginalization); thresholdPosterior_apply is the Fig. 5 left-panel marginal.
  2. L&G-anchored interpretive wrappers for generic PMF theorems already promoted to Core/Probability/Posterior.lean (the structural sorites bound, the borderline-as-intermediate-measure schema, posterior concentration).
  3. L&G's Eq. 32 outer-measure metalinguistic-probability formula directly via PMF.toOuterMeasure.
  4. The Frank-&-Goodman 2012 scalar-implicature mechanism that L&G use as a §3 warmup — clearly labelled as not L&G's novel contribution.

Not yet captured (paper-specific evaluative content, not structural gaps):

Connection to the bundled formalisation #

The bundled-RSAConfig formalisation in LassiterGoodman2017.lean (sibling file, in this directory) handles the empirical-fit reproduction via rsa_predict reflection. Its defaultCfg.L1 : Utterance → Height → ℝ returns unnormalised real values via the bundled API; it is not PMF-shaped.

The two files are deliberately complementary, not bridged: this one for the structural skeleton + cross-framework positioning, the bundled one for numeric simulations of Figs. 5-10. A formal bridge would require a PMF-shaped projection of defaultCfg.L1 (currently absent from the bundled API) and isn't load-bearing for either file's contribution. The structural bounds proved here apply to any PMF over thresholds, so any future PMF projection of defaultCfg.L1 would inherit them by construction.

Geometry of the sorites bound #

L&G 2017 Eq. 37 defines premise probability as an integral over an interval:

P(x_m tall ∧ x_{m-1} not tall) = ∫_{h(x_{m-1})}^{h(x_m)} L1_latent(.tall) dθ

With grid spacing ε > 0 (e.g., ε = 0.5 inch in the Fig. 8 simulation), each sorites gap corresponds to a set of thresholds, not a single value. The discrete sum-over-singletons bound proved here (sorites_borderline_sum_le_one) is the single-point discretisation of Eq. 37; the full interval-additive form requires sigma-additivity of PMF.toOuterMeasure on disjoint sets, which is true but not yet factored out as a lemma. Stated as sorites_premise_interval_sum_le_one below as a deferred sorry.

Cross-framework positioning (linglib's "make incompatibilities visible") #

L&G's probabilistic resolution of the sorites and characterisation of borderline cases are one of several formalised positions in linglib:

Per linglib's "no bridge files" discipline, framework-comparison content is anchored here (the chronologically-later paper) rather than in a dedicated comparison file. The §7 theorem below proves the L&G prediction, contrasts it with the @cite{alxatib-pelletier-2011} borderline-contradiction data, and so makes the empirical incompatibility between L&G's literal-meaning rule and the observed acceptance rate visible at theorem level.

Note on Adams's bound #

L&G p. 25 cites Adams (1966) on cumulative uncertainty: "the uncertainty of the consequent cannot exceed the sum of the uncertainties of the premises." This is a generic probability theorem (P(⋂ A_i) ≥ 1 - ∑ (1 - P(A_i))) with no RSA-specific content. Stated as adams_uncertainty_bound (sorry'd) for completeness; the proof is a verbose induction that belongs in mathlib's outer-measure library, not here. L&G p. 27 (Eq. 38) also invokes a different Adams thesis ("The Equation": P(if A then B) = P(B|A)), used for PC-sorites premise strengths — neither captured.

§1. Sorites bound (singleton discretisation) #

The discrete singleton sorites bound. Faithful to L&G 2017 §5 ARITHMETICALLY but uses a single-point approximation of Eq. 37's interval integral. The general interval-additive form is below (deferred).

theorem LassiterGoodman2017.PMF.sorites_borderline_sum_le_one {Threshold : Type u_1} (L1_latent : PMF Threshold) (s : Finset Threshold) :
θs, L1_latent θ 1

Sorites resolution (singleton discretisation): for any threshold posterior PMF and any finite set of distinct threshold values, the sum of single-threshold probabilities is bounded by 1.

This is the discrete-grid approximation of L&G 2017 Eq. 37 with grid-spacing ε such that each sorites gap corresponds to exactly one threshold value. Faithful to the §5 arithmetic of the resolution ("the cumulative probability budget is bounded") but not to the geometry (Eq. 37 sums measures of disjoint intervals, not singletons).

Wraps the generic PMF.sum_finset_le_one with L&G framing.

theorem LassiterGoodman2017.PMF.sorites_premise_interval_sum_le_one {Threshold : Type u_1} (L1_latent : PMF Threshold) {β : Type u_2} (s : Finset β) (I : βSet Threshold) (h_disjoint : bs, b's, b b'Disjoint (I b) (I b')) :
bs, L1_latent.toOuterMeasure (I b) 1

Sorites resolution (interval-additive form, faithful to Eq. 37): for pairwise disjoint sorites-gap intervals I θ, the sum of premise-event probabilities is bounded by 1.

This is the form L&G actually use in §5 — premise probability is L1_latent.toOuterMeasure (gap interval). The bound follows from sigma-additivity of PMF.toOuterMeasure on disjoint sets + toOuterMeasure_apply_le_one on the union.

Not yet proved: requires factoring out a sigma-additivity lemma for PMF.toOuterMeasure on Finset-indexed disjoint sets, which deserves its own Core/Probability/ slot. The discrete singleton form (sorites_borderline_sum_le_one) carries the §5 arithmetic; this form captures the §5 geometry.

§2. Borderline as intermediate measure (L&G §4.4 closing argument) #

The probabilistic characterisation of "borderline": the metalinguistic probability P_T(a is tall) = L1_latent.toOuterMeasure {θ | θ ≤ h(a)} is intermediate exactly when the threshold posterior straddles the height being judged.

Cross-framework note: this characterisation is contested. Supervaluation (Fine1975.lean) maps borderline to Truth3.indet; epistemicism denies the very framing (borderline cases have determinate Boolean truth values we don't know); TCS predicts borderline contradictions are tolerantly true, with empirical support from @cite{alxatib-pelletier-2011}. See Phenomena/Gradability/Compare.lean for the comparison taxonomy.

theorem LassiterGoodman2017.PMF.borderline_intermediate {S : Type u_2} (L1_latent : PMF S) (s : Set S) (h_witness_in : θs, θ L1_latent.support) (h_witness_out : θs, θ L1_latent.support) :
0 < L1_latent.toOuterMeasure s L1_latent.toOuterMeasure s < 1

Borderline-case theorem: when both some threshold below h AND some threshold ≥ h are in the posterior support, the metalinguistic probability P_T is intermediate (0 < P_T < 1).

L&G's structural form of vagueness — but contested across frameworks. See module docstring for cross-framework positioning.

§3. Pragmatic strengthening — Frank-Goodman 2012 mechanism #

L&G 2017 §3 introduces this mechanism (Eqs. 14-20 with SOME/ALL) as a warmup illustration of iterated rational reasoning. The genuinely novel L&G L1 architecture is the §4 free-variable inference (Eqs. 28-29 — joint posterior over (world, threshold)), NOT formalised here.

The theorem below captures the FG2012 scalar-implicature mechanism that L&G use as exposition. It is genuinely RSA-architectural (depends on rational speaker over alternatives + Bayesian listener), but does not capture L&G's novel contribution. Anchored to L&G only because they present this mechanism in §3; the canonical reference is @cite{frank-goodman-2012}.

theorem LassiterGoodman2017.PMF.pragmatic_strengthening {Utt : Type u_2} {World : Type u_3} (score : WorldUttENNReal) (h_score_top : ∀ (w : World), ∑' (u : Utt), score w u ) (h_score_pos : ∀ (w : World), ∑' (u : Utt), score w u 0) (S1 : WorldPMF Utt) (h_S1 : ∀ (w : World), S1 w = PMF.normalize (score w) ) (worldPrior : PMF World) (u_weak : Utt) (w_strong w_only_weak : World) (h_marg : PMF.marginal S1 worldPrior u_weak 0) (h_eq : score w_strong u_weak = score w_only_weak u_weak) (h_pos_strong : score w_strong u_weak 0) (h_pos_strong_top : score w_strong u_weak ) (h_partition_strict : ∑' (u : Utt), score w_only_weak u < ∑' (u : Utt), score w_strong u) (h_prior_eq : worldPrior w_strong = worldPrior w_only_weak) (h_prior_pos : worldPrior w_strong 0) :
(PMF.posterior S1 worldPrior u_weak h_marg) w_strong < (PMF.posterior S1 worldPrior u_weak h_marg) w_only_weak

Pragmatic strengthening (FG2012 mechanism): when u_weak applies at both w_strong and w_only_weak but a stronger alternative u_strong applies only at w_strong, the listener posterior at u_weak underspecifies w_strong. Generalised over arbitrary world prior: requires the prior to assign equal weight to the two compared worlds (the "neutral prior" assumption isolates the speaker-side contribution).

The proof composes the promoted posterior_lt_of_kernel_lt_of_prior_eq with normalize_lt_of_apply_eq_of_sum_lt. This is the FG2012 scalar-implicature mechanism, reframed at the L&G layer.

theorem LassiterGoodman2017.PMF.iteration_strengthens {Utt : Type u_2} {World : Type u_3} (score : WorldUttENNReal) (h_score_top : ∀ (w : World), ∑' (u : Utt), score w u ) (h_score_pos : ∀ (w : World), ∑' (u : Utt), score w u 0) (S1 : WorldPMF Utt) (h_S1 : ∀ (w : World), S1 w = PMF.normalize (score w) ) (worldPrior : PMF World) (u_weak : Utt) (w_strong w_only_weak : World) (h_marg : PMF.marginal S1 worldPrior u_weak 0) (h_eq : score w_strong u_weak = score w_only_weak u_weak) (h_pos_strong : score w_strong u_weak 0) (h_pos_strong_top : score w_strong u_weak ) (h_partition_strict : ∑' (u : Utt), score w_only_weak u < ∑' (u : Utt), score w_strong u) (h_prior_eq : worldPrior w_strong = worldPrior w_only_weak) (h_prior_pos : worldPrior w_strong 0) :
(PMF.posterior S1 worldPrior u_weak h_marg) w_strong (PMF.posterior S1 worldPrior u_weak h_marg) w_only_weak

Iteration strictly strengthens L0: corollary of pragmatic_strengthening. The same asymmetric extensions that produce the strengthening also distinguish L1 from L0.

§3'. Note on context-sensitivity (L&G §4.4) #

L&G's §4.4 "skyscraper" passage characterises context-sensitivity via prior shape transfer to posterior shape — the height prior for "buildings" has a different mean and standard deviation than for "people", and this propagates through the joint inference (Eqs. 28-29) to a different threshold posterior, hence a different metalinguistic probability of "tall".

A previous version of this file ("prior_dominates_implicature") captured a much weaker claim: that a strongly-skewed prior on a single world can override the kernel's pragmatic ranking on that world. That is generic Bayes weighting (posterior_lt_iff_score_lt), not L&G's §4.4 claim about prior-shape transfer. The honest version requires the joint posterior of Eqs. 28-29, which lives in the substrate gap (§4 of the module docstring).

The generic Bayes-weighting result is PMF.posterior_lt_iff_score_lt in Core/Probability/Posterior.lean and is reusable for any consumer; no L&G-specific wrapper here.

§3.5. The joint (world × threshold) posterior — L&G's novel contribution #

L&G 2017's central architectural innovation (§4.2, Eqs. 26-29): the pragmatic listener jointly infers the world A and the free threshold variable V, with marginalisation over V (Eq. 30) yielding the height posterior of §4.3 (Fig. 5).

No new definition needed: the joint posterior is PMF.posterior instantiated at the product type α := W × T. The structural content is the marginalisation formulas — which are direct instantiations of PMF.posterior_fst_apply / posterior_snd_apply from Core/Probability/JointPosterior.lean.

This section captures the architectural skeleton. Per-paper numerical simulations (Fig. 5/6/7/8) are deliberately out of scope — they are MCMC approximations of the integrals, not structural theorems about the model class. The structural theorems below apply to ANY choice of (W, T, Utt, S1, prior) instantiating L&G's joint chain.

noncomputable def LassiterGoodman2017.PMF.jointL1 {W : Type u_2} {T : Type u_3} {Utt : Type u_4} (S1 : W × TPMF Utt) (worldThresholdPrior : PMF (W × T)) (u : Utt) (h_marg : PMF.marginal S1 worldThresholdPrior u 0) :
PMF (W × T)

L&G 2017 Eq. 29: the joint posterior over (world, threshold). P_L1(W, T | u) ∝ P_S1(u | W, T) · P(W, T).

This is just PMF.posterior at α := W × T — no new definition. The structural content lives in the marginalization theorems below.

Equations
Instances For
    theorem LassiterGoodman2017.PMF.worldPosterior_apply {W : Type u_2} {T : Type u_3} {Utt : Type u_4} [Fintype W] [Fintype T] [Fintype Utt] [DecidableEq W] (S1 : W × TPMF Utt) (worldThresholdPrior : PMF (W × T)) (u : Utt) (h_marg : PMF.marginal S1 worldThresholdPrior u 0) (w : W) :
    (jointL1 S1 worldThresholdPrior u h_marg).fst w = (∑ θ : T, worldThresholdPrior (w, θ) * (S1 (w, θ)) u) / PMF.marginal S1 worldThresholdPrior u

    L&G 2017 Eq. 30: the world (height) marginal of the joint posterior. P_L1(w | u) = ∑_θ P_L1(w, θ | u) = (∑_θ P(w, θ) · S1(u | w, θ)) / Z.

    The height-marginalisation formula (paper Fig. 5 right panel — height posterior). Direct corollary of PMF.posterior_fst_apply.

    theorem LassiterGoodman2017.PMF.thresholdPosterior_apply {W : Type u_2} {T : Type u_3} {Utt : Type u_4} [Fintype W] [Fintype T] [Fintype Utt] [DecidableEq W] [DecidableEq T] (S1 : W × TPMF Utt) (worldThresholdPrior : PMF (W × T)) (u : Utt) (h_marg : PMF.marginal S1 worldThresholdPrior u 0) (θ : T) :
    (jointL1 S1 worldThresholdPrior u h_marg).snd θ = (∑ w : W, worldThresholdPrior (w, θ) * (S1 (w, θ)) u) / PMF.marginal S1 worldThresholdPrior u

    L&G 2017 Fig. 5 left panel: the threshold marginal of the joint posterior. P_L1(θ | u) = ∑_w P_L1(w, θ | u) = (∑_w P(w, θ) · S1(u | w, θ)) / Z.

    The threshold-marginalisation formula. The threshold posterior is what yields the metalinguistic probability of Eq. 32 (cf. §4 below).

    theorem LassiterGoodman2017.PMF.worldPosterior_lt_iff {W : Type u_2} {T : Type u_3} {Utt : Type u_4} [Fintype W] [Fintype T] [Fintype Utt] [DecidableEq W] (S1 : W × TPMF Utt) (worldThresholdPrior : PMF (W × T)) (u : Utt) (h_marg : PMF.marginal S1 worldThresholdPrior u 0) (w₁ w₂ : W) :
    (jointL1 S1 worldThresholdPrior u h_marg).fst w₁ < (jointL1 S1 worldThresholdPrior u h_marg).fst w₂ θ : T, worldThresholdPrior (w₁, θ) * (S1 (w₁, θ)) u < θ : T, worldThresholdPrior (w₂, θ) * (S1 (w₂, θ)) u

    Comparison decomposition for the height posterior. For two heights w₁, w₂, the L1 height posterior favours w₂ over w₁ iff the conditional-joint sums favour it. Generalises PMF.posterior_lt_iff_score_lt to the marginalised case.

    Useful for paper claims of the shape "L1 favours height h₂ > h₁ given 'tall'" — reduces to comparing per-height conditional joint masses. Direct corollary of PMF.posterior_fst_lt_iff.

    §4. L&G Eq. 32 — metalinguistic probability as outer measure #

    L&G 2017 Eq. 32 (paper p. 22, verified):

    P_T(a is tall) = ∫_0^{height(a)} P_L1(θ_tall | u = "a is tall") dθ_tall
    

    The outer measure of Set.Iio h (thresholds at most h) under the threshold posterior is the discrete analogue. No new theorem here; the formal content is just PMF.toOuterMeasure (Set.Iio h).

    §5. Adams's bound — generic probability, not RSA-specific #

    L&G p. 25 cites Adams (1966) on cumulative uncertainty:

    1 - P(⋂ A_i) ≤ ∑ (1 - P(A_i))
    

    or equivalently P(⋂ A_i) ≥ 1 - ∑ (1 - P(A_i)).

    Generic probability theorem with no RSA-specific content; stated for completeness, with proof deferred. The right home is mathlib's outer- measure library, not here.

    theorem LassiterGoodman2017.PMF.adams_uncertainty_bound {S : Type u_2} (p : PMF S) (sets : List (Set S)) :
    p.toOuterMeasure (List.foldr (fun (x1 x2 : Set S) => x1 x2) Set.univ sets) 1 - List.foldr (fun (A : Set S) (acc : ENNReal) => acc + (1 - p.toOuterMeasure A)) 0 sets

    Adams's bound (cumulative uncertainty, generic): for any indexed family of sets in a PMF, the measure of the intersection is bounded below by 1 - ∑ (1 - measure(A_i)). Generic probability theorem; deferred.

    §6. Posterior concentration as fully-informative limit #

    When only one world has positive prior · kernel mass at observation u, the posterior concentrates entirely on that world. The deterministic limit of Bayesian update.

    The theorem posterior_eq_one_of_singleton_score_support is in Core/Probability/Posterior.lean; it generalises the L&G "fully informative L1 update" intuition without L&G framing.

    §7. Empirical contrast with @cite{alxatib-pelletier-2011} #

    L&G's literal-meaning prediction for "X is tall and X is not tall" — the joint-Boolean conjunction of two complementary truth-conditional contributions — is bounded by P_T(tall) · (1 - P_T(tall)) ≤ 1/4, maximised at the maximally borderline P_T = 1/2.

    Empirical contrast: @cite{alxatib-pelletier-2011} report 44.7% acceptance for "X is tall and not tall" applied to the median (borderline) man in their visual stimulus. The data is encoded at Phenomena/Gradability/Vagueness.lean::alxatibPelletier2011Tall.

    44.7% > 25%, so a literal-meaning probabilistic account cannot reproduce the data. This is the formal expression of the empirical challenge that motivated TCS (Linglib/Theories/Semantics/Supervaluation/TCS.lean, @cite{cobreros-etal-2012}), where borderline cases tolerantly satisfy P ∧ ¬P as a tolerantly-true proposition — not via probability multiplication.

    This file does not formalise L&G's pragmatic enrichment of the literal prediction (which would route through the joint posterior of Eqs. 28-29, in the substrate gap). The bound below is the literal-rule prediction only.

    theorem LassiterGoodman2017.PMF.lg_literal_borderline_bounded {S : Type u_2} (L1_latent : PMF S) (s : Set S) :
    L1_latent.toOuterMeasure s * (1 - L1_latent.toOuterMeasure s) 1 / 4

    L&G literal-rule borderline-contradiction bound: under the literal-meaning rule P("X is P and not P") = P_T(P) · P_T(¬P), the predicted acceptance is bounded by 1/4, regardless of the underlying threshold posterior or the height being judged.

    Empirical contrast: @cite{alxatib-pelletier-2011} report 44.7% — well above 25% — so the literal-rule prediction is empirically refuted. TCS (@cite{cobreros-etal-2012}, formalised in Theories/Semantics/Supervaluation/TCS.lean) accommodates the data via non-probabilistic tolerant satisfaction.

    §8. L&G-vs-TCS framework expressivity gap (Lean-checkable reciprocation) #

    The previous version of this engagement was docstring-prose-only: §7 above mentions TCS as the framework that handles the data direction the literal rule misses, and Phenomena/Gradability/Studies/CobrerosEtAl2012.lean docstring mentions L&G's lg_literal_borderline_bounded as the empirical challenge that motivated TCS — but no Lean theorem connected the two.

    The theorem below closes that gap with a concrete PMF-typed witness of the expressivity divergence: simultaneously, L&G's literal-rule prediction is bounded (≤ 1/4) AND TCS's borderline-contradiction prediction is categorical (true, not fractional). The abstract real-arithmetic version of the same gap is CobrerosEtAl2012.tcs_categorical_vs_product_bounded. Together the two make the framework architecture difference visible at theorem level from both sides.

    L&G-vs-TCS framework gap, PMF-typed: L&G's literal-rule prediction P(P) · P(¬P) is bounded above by 1/4 for any PMF (lg_literal_borderline_bounded), while TCS's borderline-contradiction prediction is categorical — the conjunction is tolerantly true, not a fractional value to be matched.

    The empirical Alxatib-Pelletier 2011 acceptance rate (44.7%) exceeds the L&G literal upper bound (25%), refuting the literal-rule's expressivity at this dataset. TCS's categorical prediction is consistent with the empirical direction. The Lean theorem records both halves of the framework gap simultaneously.