Skip to content

Beyond Single-Sample: Reliable Multi-Sample Distillation for Video Understanding

Conference: CVPR2026 arXiv: 2603.11423 Code: N/A Area: Video Understanding Keywords: Knowledge Distillation, Black-Box Distillation, Video LVLM, Multi-Sample Sampling, Adversarial RL Distillation, Teacher Reliability

TL;DR

This paper identifies severe unreliability in single-sample teacher responses under black-box distillation for video LVLMs—manifested as cross-question variance (\(\sigma=0.22\)), intra-sampling variance (\(\sigma=0.07\)\(0.15\)), and format violation rates (1%–10%)—and proposes R-MSD, a framework that addresses these issues through a multi-sample teacher pool, task-adaptive matching, and two-stage SFT→RL adversarial distillation. The resulting 4B student model comprehensively outperforms the same-scale Qwen3-VL-4B on VideoMME, Video-MMMU, and WorldSense.

Background & Motivation

Background: Black-box distillation is the dominant paradigm for LVLM compression. Since large model APIs expose only text outputs—without logits or intermediate features—practitioners collect teacher text responses as training signals. This approach has been widely adopted in NLP (e.g., Alpaca, Vicuna), yet its reliability in video multimodal settings has not been thoroughly investigated.

Limitations of Prior Work: Large-scale statistical analysis of teacher models such as GPT-4o reveals three categories of unreliability in single-sample teacher responses: - High cross-question variance (\(\sigma=0.22\)): Substantial difficulty variation across questions leads to pronounced fluctuation in teacher response quality. - Non-negligible intra-sampling variance (\(\sigma_{\text{sampling}}=0.07\)\(0.15\)): Repeated sampling of the same question yields responses of varying quality. - Pervasive format violations (1%–10%): Teacher outputs frequently fail to conform to specified formats (e.g., omitting option letters in MCQ tasks).

Key Challenge: Standard SFT-based distillation treats a single teacher response as ground truth, implicitly assuming that the teacher is always correct. The above analysis demonstrates that this assumption is seriously violated for video tasks.

Core Problem: Closed-form tasks (MCQ, temporal ordering, bounding-box localization) admit well-defined correctness metrics, whereas open-ended tasks (description, explanation) lack reliable automatic evaluation. A unified matching strategy is therefore inappropriate across task types.

Goal: Naïve baselines—selecting the best-of-K responses for SFT, or using all K samples in SFT+RL—yield only marginal gains, motivating a more principled multi-sample utilization strategy.

Method

Overall Architecture: R-MSD

R-MSD (Reliable Multi-Sample Distillation) comprises three core components: (1) construction of a multi-sample teacher pool; (2) task-adaptive teacher–student matching; and (3) two-stage training (SFT warmup → RL adversarial distillation).

Component 1: Multi-Sample Teacher Pool

  • For each training input (video + question), \(K\) responses (\(K=5\)\(10\)) are sampled from the teacher API.
  • For closed-form tasks, a rule-based verifier (IoU / exact match) automatically scores each response.
  • For open-ended tasks, no quality ranking is performed, avoiding biases introduced by fragile LLM-as-judge approaches.
  • Responses violating format constraints are filtered, retaining only compliant subsets.

Component 2: Task-Adaptive Matching

Different teacher–student pairing strategies are applied according to task type:

  • Closed-form tasks — quality-biased matching: The highest-quality response in the teacher pool is preferentially selected as the SFT target and the RL positive pair, while lower-quality responses serve as RL negative pairs. Quality is measured via IoU (bounding-box tasks) or exact match (MCQ / temporal tasks).
  • Open-ended tasks — uniform matching: In the absence of reliable quality metrics, pairs are sampled uniformly from the pool to avoid introducing systematic bias from fragile proxies such as lexical similarity.

Component 3: Two-Stage Training

Stage 1: SFT Warmup

  • For each input, the best response is selected from the teacher pool (quality score for closed-form tasks; random selection for open-ended tasks).
  • Standard cross-entropy SFT enables the student to acquire the teacher's basic capabilities and output format.
  • This stage provides a well-initialized policy for Stage 2 RL training.

Stage 2: RL + Adversarial Distillation

  • Student rollout: Given an input, the student autoregressively samples a response.
  • Adversarial pairing: Student rollouts are paired with responses from the teacher pool.
  • Closed-form tasks: student rollout vs. teacher best → positive contrast; student rollout vs. teacher worst → negative contrast.
  • Open-ended tasks: student rollout vs. random teacher response → uniform contrast.
  • Online Critic-as-Discriminator: A lightweight critic network is trained to distinguish student from teacher responses; its discrimination probability serves as a distribution-level supervision signal—capturing full-sequence quality rather than per-token KL divergence.
  • Rule Reward: For closed-form tasks, a rule-based reward (exact match score) is added and combined with the critic reward via weighted summation.
  • Policy Optimization: Student parameters are updated via PPO-style policy gradient.

Loss & Training

\[\mathcal{L}_{\text{total}} = \mathcal{L}_{\text{policy}} + \lambda_{\text{critic}} \mathcal{L}_{\text{critic}} + \lambda_{\text{rule}} \mathcal{L}_{\text{rule}}\]
  • \(\mathcal{L}_{\text{policy}}\): PPO policy gradient loss (with clipping)
  • \(\mathcal{L}_{\text{critic}}\): Critic discrimination loss (binary cross-entropy)
  • \(\mathcal{L}_{\text{rule}}\): Rule reward loss (closed-form tasks only)

Key Experimental Results

Setup

  • Teacher: GPT-4o (K=5 responses sampled per query via API)
  • Student backbone: Qwen3-VL-4B
  • Training data: Multi-type video tasks including MCQ, temporal reasoning, bounding-box localization, and open-ended description
  • Evaluation benchmarks: VideoMME, Video-MMMU, WorldSense, MVBench

Main Results

Method Params VideoMME Video-MMMU WorldSense MVBench
Qwen3-VL-4B (baseline) 4B 63.8 55.4 46.7 68.2
SFT (single-sample teacher) 4B 64.1 55.8 47.0 68.5
SFT (best-of-K teacher) 4B 64.5 56.2 47.3 68.9
SFT + RL (same-budget baseline) 4B 64.3 56.0 47.1 68.7
R-MSD (Ours) 4B 65.3 58.6 49.2 70.1
GPT-4o (teacher) 71.9 63.8 55.2 74.5

R-MSD consistently surpasses the Qwen3-VL-4B baseline and same-budget SFT+RL methods across all benchmarks: +1.5 on VideoMME, +3.2 on Video-MMMU, and +2.5 on WorldSense.

Teacher Unreliability Analysis

Unreliability Type Metric Value
Cross-question variance \(\sigma\) of teacher accuracy 0.22
Intra-sampling variance (MCQ) \(\sigma\) of accuracy over K samples 0.07
Intra-sampling variance (open-ended) \(\sigma\) of ROUGE-L over K samples 0.15
Format violation rate (MCQ) Proportion missing option letters ~3%
Format violation rate (bbox) Proportion with malformed bounding boxes ~10%
Format violation rate (description) Proportion with severe truncation / empty output ~1%

Ablation Study

Component VideoMME Video-MMMU WorldSense
Full R-MSD 65.3 58.6 49.2
− Multi-sample (\(K=1\)) 64.3 56.5 47.5
− Task-adaptive matching (unified strategy) 64.8 57.3 48.1
− Stage 2 RL (SFT only) 64.5 56.2 47.3
− Critic discriminator (rule reward only) 64.9 57.8 48.5
− Format filtering 64.7 57.1 48.0

Key Findings

  • Multi-sample sampling is the most critical component: Removing multi-sample collection (\(K=1\)) causes substantial drops across all metrics, corroborating the central claim regarding single-sample unreliability.
  • Task-adaptive matching cannot be replaced by a unified strategy: Applying either quality-biased or uniform matching universally underperforms the adaptive approach.
  • The RL stage contributes substantially: SFT alone fails to fully exploit the diversity information encoded in the teacher pool.
  • The critic discriminator provides complementary supervision: Relying solely on rule rewards leaves open-ended tasks unsupervised; the critic fills this gap.
  • Format filtering is simple yet effective: Removing format-violating responses prevents the student from learning erroneous patterns.
  • Same-budget baselines yield marginal gains: Naïvely applying all K responses to SFT or using random pairing in RL fails to leverage quality differences among responses.

Highlights & Insights

  • This work is the first to systematically quantify three categories of teacher response unreliability in black-box distillation for video LVLMs, grounding methodological design in rigorous statistical analysis.
  • The task-adaptive matching strategy is well-motivated: objective metrics are exploited for closed-form tasks where they exist, and circumvented for open-ended tasks where they are unreliable, preventing the introduction of new biases.
  • The two-stage training pipeline is logically coherent: SFT warmup establishes foundational capabilities, and RL adversarial distillation subsequently extracts deeper information from the teacher pool.
  • The Critic-as-Discriminator provides distribution-level supervision, which is more appropriate than per-token KL divergence in the black-box setting where logits are unavailable.
  • The 4B student model achieves competitive results across multiple benchmarks, demonstrating clear practical value.

Limitations & Future Work

  • Only GPT-4o is evaluated as the teacher; generalizability to other teachers (Claude, Gemini) remains unverified.
  • Sampling \(K=5\) responses incurs API costs five times those of single-sample approaches, significantly increasing expenditure at training scale.
  • The student is validated only on Qwen3-VL-4B; performance with larger or smaller students is unknown.
  • Training stability and hyperparameter sensitivity of the critic network are not discussed in depth.
  • Uniform matching for open-ended tasks avoids bias but also forgoes opportunities to exploit quality variation, leaving potential room for improvement.
  • A substantial performance gap to the teacher remains (VideoMME: 65.3 vs. 71.9), suggesting that information transfer efficiency can be further improved.
  • Black-box knowledge distillation: Alpaca, Vicuna, WizardLM, and related works train smaller models on GPT outputs but uniformly assume teacher response reliability.
  • Video LVLMs: Qwen-VL, InternVL, VideoLLaVA, and others advance video understanding at the cost of high computational overhead.
  • RL for LLMs: RLHF, DPO, and PPO address alignment and optimization; R-MSD introduces RL into the distillation setting.
  • Data quality filtering: Phi-3, LIMA, and related works emphasize the importance of high-quality data; R-MSD's multi-sample filtering reflects an analogous philosophy.
  • Video benchmarks: VideoMME, Video-MMMU, and WorldSense evaluate video understanding from complementary perspectives.

Rating

  • Novelty: ⭐⭐⭐⭐ — The problem insight (systematic quantification of teacher unreliability) is valuable; individual methodological components are relatively standard.
  • Experimental Thoroughness: ⭐⭐⭐⭐ — Ablations are comprehensive and statistical analysis is rigorous, though student/teacher combinations are limited.
  • Writing Quality: ⭐⭐⭐⭐ — Problem analysis is thorough, statistical evidence is well-presented, and method description is clear.
  • Value: ⭐⭐⭐⭐ — The multi-sample distillation paradigm has direct practical relevance to LVLM compression and deployment.