IncrementalSemantics — Cohn-Gordon's bundle for word-by-word RSA #
@cite{cohn-gordon-goodman-potts-2019}
A scene-specific incremental RSA model factors into three pieces:
wordApplies : U → W → Bool— word-level Boolean truthcompleteUtterances : List (List U)— closed set of full utterancesworlds : List W— referents to normalize over
From these three pieces every other ingredient — utterance-level truth,
extension-based incremental semantics ⟦pfx⟧(r), the chain-rule speaker,
the literal-listener categorical L0^UTT, and the full RSAConfig — is
derived rather than re-stipulated per study.
This consolidates what was previously triplicated across CG's Figure 1
scene, the @cite{sedivy-2007} reference-game scene, and the
@cite{rubio-fernandez-2016} display: each becomes a single
IncrementalSemantics value, with toRSAConfig producing the model.
The deep theorem (§2.4) #
l0Utt_ge_inv_card proves Cohn-Gordon's weakly-informative bound
generically over the bundle: any complete utterance true of r ∈ worlds
yields a literal-listener posterior of at least 1 / worlds.length for
r. The bound follows from filter-monotonicity (numerator ≥ 1 since r
is in the filter; total ≤ worlds.length since filtering can only
shrink). Concrete scenes inherit the bound by instantiating the bundle.
Bundle of scene-specific data for an incremental RSA model.
The three fields jointly determine the entire model: toRSAConfig
builds the No-Brevity (s1Score = L0, α = 1, no cost) RSAConfig,
incrementalSem derives the extension-based meaning function
(@cite{cohn-gordon-goodman-potts-2019} §2.2), and l0Utt projects
the literal listener over complete utterances.
- wordApplies : U → W → Bool
Word-level Boolean truth: does word
uapply to worldw? - completeUtterances : List (List U)
Closed set of complete utterances available in the scene.
- worlds : List W
Referents to normalize over (e.g. the visual display).
Instances For
Utterance-level Boolean semantics: conjunction of word applicability.
Equations
- sem.uttSem utt r = utt.all fun (w : U) => sem.wordApplies w r
Instances For
Number of complete utterances extending pfx that are true of r.
Equations
- sem.trueExtCount pfx r = (List.filter (fun (u : List U) => pfx.isPrefixOf u && sem.uttSem u r) sem.completeUtterances).length
Instances For
Number of complete utterances extending pfx that are true of at
least one referent in sem.worlds.
Equations
- sem.viableExtCount pfx = (List.filter (fun (u : List U) => pfx.isPrefixOf u && sem.worlds.any fun (r : W) => sem.uttSem u r) sem.completeUtterances).length
Instances For
Extension-based incremental semantics (@cite{cohn-gordon-goodman-potts-2019} §2.2):
⟦pfx⟧(r) = trueExtCount(pfx, r) / viableExtCount(pfx)
Equations
- sem.incrementalSem pfx r = ↑(sem.trueExtCount pfx r) / ↑(sem.viableExtCount pfx)
Instances For
Literal listener over complete utterances:
L0^UTT(r | utt) = ⟦utt⟧(r) / Σ_{r'} ⟦utt⟧(r').
For complete utt with no proper extensions, ⟦utt⟧ collapses to
Boolean truth, so this is uniform-prior Bayes over worlds.
Equations
- One or more equations did not get rendered due to their size.
Instances For
Cohn-Gordon §2.4 weakly-informative bound (generic).
For any complete utterance utt that is true of r (with
r ∈ worlds), the literal listener assigns posterior at least
1 / worlds.length to r. The proof is purely combinatorial:
the numerator is 1 (since utt is true of r), and the total
counts referents satisfying utt, which is at most worlds.length
and at least 1 (since r is in the filter).
Cohn-Gordon use this bound to certify that greedy unrolling — even
without a global view of the utterance space — never produces an
utterance arbitrarily worse than uniform. Studies that build a
greedy unroller for a specific scene need only prove that the
output is in completeUtterances and is true of the target;
the bound then follows.
The No-Brevity incremental RSA built from the bundle:
chain-rule speaker, α = 1, no cost, uniform priors,
extension-based L0 meaning. The "No-Brevity" name (after
@cite{dale-reiter-1995}) flags s1Score = L0: the speaker scores
each next word by the literal listener's posterior, with brevity
emerging only via the chain-rule product over longer trajectories.
Equations
- One or more equations did not get rendered due to their size.