Documentation

Linglib.Phenomena.Negation.Studies.StankovaSimik2024.Data

Staňková & Šimík (2024): Negation in Czech Polar Questions #

@cite{stankova-2025} @cite{gartner-gyuris-2017} @cite{simik-2024}

Experimental data from three naturalness judgment experiments on negation in Czech polar questions (Staňková & Šimík, FASL 32 / JSL 33).

Main experiment (§5) #

2×2×2 within-subjects design manipulating:

75 native Czech speakers, Likert 1–7, 32 items (4 per condition).

Key findings #

  1. FALSUM is preferred in V1 (interrogative) PQs — PPIs preferred over NCIs
  2. Declarative word order (nonV1) is preferred in negatively biased contexts
  3. Czech FALSUM is compatible with any type of evidential bias (positive, negative, neutral) — broader than English
  4. The particle náhodou is licensed by FALSUM — PPIs preferred (§6.1)
  5. The particle copak requires contextual evidence — biased contexts preferred (§6.2)

A Cumulative Link Mixed Model (CLMM) regression result. z-values stored as Int × 1000 for precision without Float.

  • name : String

    Human-readable effect name

  • z1000 :

    z-value × 1000 (e.g., -15674 = z = -15.674)

  • significant : Bool

    Whether the effect is statistically significant

  • pThreshold : String

    p-value threshold (as string for display)

Instances For
    @[implicit_reducible]
    Equations
    def Data.instReprCLMMEffect.repr :
    CLMMEffectStd.Format
    Equations
    • One or more equations did not get rendered due to their size.
    Instances For
      Equations
      Instances For
        @[implicit_reducible]
        Equations

        Whether a CLMM effect's z-value is positive (higher ratings in the first level of the factor).

        Equations
        Instances For

          Whether a CLMM effect's z-value is negative.

          Equations
          Instances For

            V1: Main effect of INDEFINITE (NCI < PPI). PPIs (nějaký) are significantly more natural than NCIs (žádný) in V1 PQs. z = −15.674, p < .001.

            Interpretation: V1 negation is interpreted as FALSUM (outer negation), which licenses PPIs but blocks NCIs.

            Equations
            • Data.v1_indefinite = { name := "V1: INDEFINITE", z1000 := -15674, significant := true, pThreshold := "< .001" }
            Instances For

              V1: Main effect of CONTEXT (not significant). z = −1.374, p = 0.169.

              Interpretation: FALSUM/outer negation is insensitive to contextual evidence — V1 PQs are equally natural in negative and neutral contexts. This is because FALSUM conveys epistemic bias (speaker's possibility assessment), not evidential bias.

              Equations
              • Data.v1_context = { name := "V1: CONTEXT", z1000 := -1374, significant := false, pThreshold := "= 0.169" }
              Instances For

                V1: CONTEXT × INDEFINITE interaction. z = 2.933, p < 0.01.

                Post-hoc: the effect of INDEFINITE is more pronounced in neutral contexts. Simple effect of CONTEXT within PPI: z = −3.522, p < .001. Simple effect of CONTEXT within NCI: z = 1.104, p = .27 (n.s.).

                Equations
                • Data.v1_interaction = { name := "V1: CONTEXT × INDEFINITE", z1000 := 2933, significant := true, pThreshold := "< 0.01" }
                Instances For

                  Post-hoc: V1, simple effect of CONTEXT within PPI level. z = −3.522, p < .001. PPI V1 PQs are more natural in neutral context.

                  Equations
                  Instances For

                    Post-hoc: V1, simple effect of CONTEXT within NCI level. z = 1.104, p = .27 (not significant).

                    Equations
                    Instances For

                      nonV1: Main effect of CONTEXT. Negative contexts rated significantly more natural than neutral contexts. z = 8.674, p < 0.01.

                      Interpretation: nonV1 (declarative) PQs are sensitive to evidential bias and preferred in negatively biased contexts.

                      Equations
                      • Data.nonV1_context = { name := "nonV1: CONTEXT", z1000 := 8674, significant := true, pThreshold := "< 0.01" }
                      Instances For

                        nonV1: Main effect of INDEFINITE (NCI > PPI). NCIs (žádný) rated higher than PPIs (nějaký) in nonV1 PQs. z = 6.208, p < 0.01.

                        Interpretation: inner negation (Op¬) is the preferred reading for nonV1, licensing NCIs.

                        Equations
                        • Data.nonV1_indefinite = { name := "nonV1: INDEFINITE", z1000 := 6208, significant := true, pThreshold := "< 0.01" }
                        Instances For

                          V1 PQs: PPIs preferred over NCIs → outer (FALSUM) is the reading.

                          V1 PQs: No effect of context → FALSUM is insensitive to evidential bias.

                          nonV1 PQs: Negative context preferred → sensitive to evidential bias.

                          nonV1 PQs: NCIs preferred → inner negation is the default reading.

                          The key asymmetry: V1 is CONTEXT-insensitive (FALSUM = epistemic bias), nonV1 is CONTEXT-sensitive (inner neg = evidential bias).

                          This matches @cite{stankova-2025}'s claim that inner negation requires contextual evidence while outer negation (FALSUM) does not, and confirms the VerbPosition.requiresContextualEvidence classification.

                          Czech FALSUM with positive evidential bias.

                          In a subexperiment, V1 PQs were tested in contexts with positive evidence for p (e.g., Eva winning first place → "Didn't Eva win a prize?"). Median rating = 6 (biased) vs 5 (neutral), Likert 1–7.

                          This confirms Czech outer negation (FALSUM) is compatible with positive evidential bias, unlike English HiNQs.

                          • medianBiased :

                            Median rating in positively biased context

                          • medianNeutral :

                            Median rating in neutral context

                          Instances For
                            Equations
                            Instances For

                              V1 PQs with positive evidence rated at least as natural as neutral.

                              Czech FALSUM compatible with all three evidence types (positive, negative, neutral). This matches the broad distribution of InterNPQ in @cite{simik-2024}'s bias profile table.

                              náhodou subexperiment: 2×2 design (Context × Indefinite). All V1 PQs with náhodou. 8 items.

                              Tests whether náhodou requires FALSUM (outer negation). If so, PPIs (nějaký) should be preferred over NCIs (žádný), because FALSUM licenses PPIs but not NCIs.

                              náhodou: Main effect of INDEFINITE (NCI < PPI). PPIs strongly preferred with náhodou. z = −12.845, p < .001.

                              Interpretation: náhodou requires FALSUM, which only licenses PPIs. "based on this we suggest that náhodou could be used as an overt indicator of the covert FALSUM operator" (S&Š §6.1).

                              Equations
                              Instances For

                                náhodou's INDEFINITE effect is in the same direction as V1's, confirming both involve FALSUM (outer negation).

                                copak subexperiment: 2×2 design (Context × PQ Polarity).

                                For positive PQs: biased context has evidence for ¬p, speaker believed p. For negative PQs: biased context has evidence for p, speaker believed ¬p. In both cases, copak expresses surprise at the evidence conflicting with the speaker's prior belief.

                                copak: Main effect of CONTEXT. Biased contexts significantly more natural than neutral. z = 9.372, p < .001.

                                Interpretation: copak requires a conflict between speaker's prior belief and current contextual evidence (evidential bias). "copak is a particle which is used to express speaker's surprise about the current contextual evidence" (S&Š §6.2).

                                Equations
                                • Data.copak_context = { name := "copak: CONTEXT", z1000 := 9372, significant := true, pThreshold := "< .001" }
                                Instances For

                                  copak requires evidential bias (biased context).

                                  náhodou vs copak: opposite context sensitivity.

                                  • náhodou (FALSUM-tied): context-insensitive. FALSUM conveys epistemic bias regardless of contextual evidence — V1 PQs are equally natural in any evidential context.
                                  • copak (evidential-bias-tied): context-sensitive. Requires a biased context where speaker's prior belief conflicts with evidence.

                                  This confirms they express different types of bias: náhodou → epistemic bias; copak → evidential bias. (@cite{stankova-2025} §6)

                                  The V1 context insensitivity and náhodou's FALSUM requirement converge: both reflect that outer negation (FALSUM) is an epistemic-bias phenomenon, not an evidential-bias phenomenon. Conversely, nonV1 context sensitivity and copak's context requirement both reflect evidential bias sensitivity.

                                  • nParticipants :
                                  • nItems :
                                  • scale : String
                                  • method : String
                                  • platform : String
                                  • citation : String
                                  Instances For
                                    Equations
                                    • One or more equations did not get rendered due to their size.
                                    Instances For
                                      Equations
                                      • One or more equations did not get rendered due to their size.
                                      Instances For
                                        Equations
                                        • Data.nahodouExperiment = { nParticipants := 75, nItems := 8, scale := "Likert 1–7", method := "CLMM (ordinal package in R)", platform := "L-Rex", citation := "Staňková & Šimík 2024, §6.1" }
                                        Instances For
                                          Equations
                                          • Data.copakExperiment = { nParticipants := 75, nItems := 8, scale := "Likert 1–7", method := "CLMM (ordinal package in R)", platform := "L-Rex", citation := "Staňková & Šimík 2024, §6.2" }
                                          Instances For