Skip to content

Saying the Unsaid: Revealing the Hidden Language of Multimodal Systems Through Telephone Games

Conference: NeurIPS 2025 arXiv: 2511.10690 Code: None Area: Interpretability Keywords: Hidden language, telephone game, concept association, multimodal interpretability, test-time probing

TL;DR

By running multi-round "telephone games" (image→text→image loops), this paper exploits the preference biases of multimodal systems to quantify the connection strength between concepts in the system's implicit space (i.e., the "hidden language"). It contributes the Telescope dataset (10,000+ concept pairs) and establishes a scalable test-time "world map" of multimodal systems.

Background & Motivation

Closed-source multimodal systems (e.g., GPT-4o) have achieved remarkable progress in recent years, yet their closed features, data, and architectures preclude the use of training-based methods to study how these systems understand the world. Traditional probing approaches—attention maps, PCA, linear probes—require access to internal model representations, rendering them infeasible in the era of closed-source systems.

The root cause is as follows: multimodal systems learn concept associations by fitting textual and visual representations, but imbalances in training data lead to uneven connection strengths across concept pairs—some combinations are thoroughly trained and thus tightly coupled, while others remain weakly linked. This disparity constitutes the system's "hidden language," yet effective test-time methods for revealing it are lacking.

The paper's key insight is that multimodal systems tend to drop weakly connected concepts during image→text compression and to introduce strongly connected ones during text→image reconstruction. This preference bias can be strategically exploited: by amplifying these shifts over multiple rounds of a telephone game, the connection strength between concepts can be quantified entirely at test time.

Method

Overall Architecture

The framework consists of three components: the Telephone Game, a Co-occurrence Frequency metric, and the Telescope dataset. For fully integrated multimodal systems (e.g., GPT-4o), the same system performs both the text→image and image→text processes; for systems composed of separate components, a V-LLM and a text-to-image model from the same organization are combined.

Key Designs

  1. Telephone Game Mechanism: The framework involves two core processes. During image→text compression, the system tends to read out the more strongly connected concepts in its understanding (e.g., reading a cow as a pig). During text→image reconstruction, it tends to generate more strongly connected concepts (e.g., replacing cherries with balloons). Changes from a single reconstruction round may be subtle (e.g., the generated image may resemble both a cow and a pig), but multiple rounds of the telephone game amplify these shifts: fragile concept combinations gradually collapse, revealing their weak connection strength within the system's understanding.

  2. Co-occurrence Frequency Metric: Over \(n\) rounds of telephone games, the co-occurrence frequency of a concept pair "A and B" is defined as \(F(A,B) = \frac{\sum_{i=1}^{r}\sum_{j=1}^{n}\mathcal{I}_{i,j}(A,B)}{r \times n}\), where \(r\) is the number of repetitions and \(\mathcal{I}_{i,j}(A,B)\) indicates whether A and B co-occur in round \(j\) of repetition \(i\) (as judged by an LLM). Higher co-occurrence frequency indicates stronger conceptual connection. This metric simultaneously captures training bias and generalization ability: stronger generalization corresponds to a more uniform distribution of connection strengths.

  3. Telescope Dataset: The dataset contains 150 common visual concepts forming 11,175 simple-pattern concept pairs (two concepts placed side by side) and 450 complex-pattern concept pairs (three fusion strategies: displayed on a television screen, rendered in Van Gogh's style, and depicted in a wood texture). This dataset serves as the systematic foundation for telephone game probing.

Loss & Training

This paper does not involve model training. The entire framework is a purely test-time method that requires no access to model parameters. An LLM is used as a "concept verifier" to determine whether concepts co-occur in image descriptions, and reasoning-capable LLMs (GPT-o1, DeepSeek-R1) are used to analyze implicit associations beyond semantic and visual similarity.

Key Experimental Results

Main Results (Metric Correlation Analysis)

Metric Pair Pearson Correlation Note
Co-occurrence frequency vs. semantic similarity 0.046 Near zero
Co-occurrence frequency vs. visual similarity -0.178 Weak negative
Semantic similarity vs. visual similarity 0.041 Near zero
OpenAI vs. StepFun 0.506 Moderate positive
OpenAI vs. QWen 0.475 Moderate positive
StepFun vs. QWen 0.503 Moderate positive

Ablation Study (Concept Collapse Rate in Complex Patterns)

Pattern Concept Collapse Rate Note
Van Gogh style 0.767 Highest collapse rate, worst generalization
Displayed on TV screen 0.740 High collapse rate
Wood texture 0.560 Relatively stable
TV screen (after bridge improvement) 0.427 Significantly improved by introducing "cartoon style" as intermediate concept

Key Findings

  • Semantic and visual similarity cannot explain the hidden language. Across 246 collapsed concept pairs, the correlation between co-occurrence frequency and semantic/visual similarity is near zero, demonstrating the need for a new metric.
  • The hidden languages of different systems are moderately correlated (~0.5), supporting the Platonic Representation Hypothesis: as multimodal systems scale up, their internal representations tend to converge toward modeling the joint statistical structure of real-world events.
  • In simple patterns, after 5 rounds of telephone games, the collapse rate for original concepts is 26.4% (single concept) and 24.4% (two identical concepts), revealing significant biases in multimodal systems.
  • Complex patterns (Van Gogh style, TV display) are more fragile than simple patterns, indicating insufficient generalization by the systems to these scenarios.
  • Reasoning-capable LLMs can uncover associations beyond semantic/visual similarity—for example, the stable connection between "cow and cola" stems from cows frequently appearing on milk packaging, while milk and cola often co-occur in beverage aisles.

Highlights & Insights

  • The telephone game metaphor is highly apt: just as information becomes distorted through human whisper chains, multimodal systems introduce preference biases during compression–reconstruction cycles.
  • The framework is test-time scalable: each new round of telephone games tends to reveal new conceptual connections, progressively building a more complete "world map" of concepts as computational budget increases.
  • The discovery of "concept bridging" has strong practical value—introducing an intermediate concept (e.g., "cartoon style") can significantly strengthen fragile concept connections, offering a pathway to improve the controllability of multimodal system outputs.
  • The analogy between reasoning-capable LLMs and "MLP linear probes" for interpreting textual evolution is illuminating, representing a paradigm shift from internal inspection to behavioral analysis.

Limitations & Future Work

  • The number of telephone game rounds is finite; concept pairs with a co-occurrence frequency of 1.0 must be excluded from metric correlation analysis, as infinite rounds cannot be conducted.
  • The diversity of outputs from large-scale systems may introduce stochastic noise.
  • The current work focuses on combinations of two concepts; more complex multi-concept combinations are left for future exploration.
  • The high experimental cost (requiring calls to closed-source APIs) constrains reporting to a subset of the Telescope dataset.
  • Concept collapse/survival judgments rely on an LLM, which may introduce judgment biases.
  • This work is related to the "hidden language of diffusion models" proposed by Chefer et al., but targets closed-source systems and operates entirely at test time.
  • The Platonic Representation Hypothesis predicts that internal representations of different systems converge; the cross-system correlation analysis (~0.5) reported in this paper provides empirical support.
  • Validation experiments on the preference biases of open-source model components such as CLIP (Appendix F) strengthen the credibility of the conclusions.
  • Insight: In the absence of access to model internals, carefully designed input–output interaction loops can effectively probe a model's internal preferences and knowledge structure.

Rating

  • Novelty: ⭐⭐⭐⭐⭐ (The telephone game framework and the concept of hidden language are highly innovative)
  • Experimental Thoroughness: ⭐⭐⭐⭐ (Rich multi-system comparisons and bridging experiments; data scale is limited by API costs)
  • Writing Quality: ⭐⭐⭐⭐⭐ (Fluent narrative, apt metaphors, excellent visualizations)
  • Value: ⭐⭐⭐⭐ (Offers a novel paradigm for interpretability of closed-source systems)