Skip to content

Diffusion-Classifier Synergy: Reward-Aligned Learning via Mutual Boosting Loop for FSCIL

Conference: NEURIPS2025
arXiv: 2510.03608
Code: None
Area: few-shot learning / incremental learning
Keywords: Few-Shot Class-Incremental Learning, diffusion model, Reward-Aligned Generation, Mutual Boosting Loop, data augmentation

TL;DR

This paper proposes the Diffusion-Classifier Synergy (DCS) framework, which establishes a closed-loop mutual boosting cycle between a diffusion model and a classifier. A multi-level reward function (feature-level + logits-level) guides the diffusion model to generate images most beneficial to the classifier, achieving state-of-the-art performance on FSCIL benchmarks.

Background & Motivation

State of the Field

Background: Few-Shot Class-Incremental Learning (FSCIL) requires models to learn new classes incrementally with only a few samples while avoiding forgetting previously acquired knowledge, posing a severe stability-plasticity dilemma.

Limitations of Prior Work

Limitations of Prior Work: Existing FSCIL methods are heavily reliant on knowledge derived from the limited initial dataset, making it difficult to substantially improve intra-task generalization and inter-task discrimination.

Root Cause

Key Challenge: Diffusion models inherently possess data augmentation potential: generating images for old classes enables knowledge replay, while augmenting new classes provides richer training signals.

Solution Direction

Solution Direction: However, directly applying diffusion models introduces two critical issues: (1) images generated conditioned solely on class names suffer from semantic misalignment and insufficient diversity; (2) the generation process lacks classifier feedback, preventing adaptive generation of samples that the classifier genuinely needs.

Paper Goals

Goal: 1. Semantic Misalignment and Diversity Deficiency: When vanilla diffusion models generate images conditioned on class names, a trade-off exists between semantic alignment precision and intra-class diversity, with both metrics falling below the real-data baseline. 2. Absence of Feedback Pathways: Existing data augmentation methods treat diffusion models as "blind teachers," unable to adaptively adjust generation based on the current state of the classifier, and thus incapable of producing hard samples near decision boundaries.

Method

Overall Architecture: Mutual Boosting Loop

The core idea of DCS is to establish a bidirectional mutual boosting loop between the diffusion model \(D\) and the FSCIL classifier \(\sigma\): - Forward direction: Generated images are fed into the classifier; multiple reward signals \(\mathcal{R}_i\) are computed from classifier outputs and used to guide the diffusion model's sampling strategy \(\phi\) via the Diffusion Alignment as Sampling (DAS) algorithm. - Backward direction: High-quality generated images are used to train the classifier, and the improved classifier in turn provides more precise reward signals. - Stable Diffusion 3.5 Medium is adopted as the base generative model, generating approximately 30 images per class—far fewer than conventional million-scale augmentation schemes.

Key Designs

Feature-Level Rewards: Semantic Consistency + Diversity

Prototype-Anchored MMD Reward (\(\mathcal{R}_{\text{PAMMD}}\)): - Employs Maximum Mean Discrepancy to measure distributional alignment between the generated image set and class prototypes. - Comprises a Diversity term (penalizing excessive similarity among generated images) and a Consistency term (encouraging semantic alignment with class prototypes). - Applicable to both new and old classes, and dynamically adjusted as classifier prototypes are updated. - Supports incremental computation to avoid redundant calculations.

Dimension-Wise Variance Matching Reward (\(\mathcal{R}_{\text{VM}}\)): - Inspired by the covariance term of FID, but full covariance matrix estimation is unstable in few-shot settings. - Instead matches the feature variance of generated and real images dimension-by-dimension. - Only participates in reward computation when more than 5 images have been generated, ensuring statistical reliability.

Logits-Level Rewards: Classifier-Aware Generation

Recalibrated Confidence Reward (\(\mathcal{R}_{\text{RC}}\)): - Based on cross-entropy but introduces an adaptive temperature \(T\) that dynamically adjusts according to the classifier's raw confidence on the target class. - High-confidence samples → higher temperature → avoids generating overly easy samples. - Low-confidence samples → temperature remains low → prevents excessive diffusion. - Encourages generation of more exploratory and generalizable intra-class samples.

Cross-Session Confusion-Aware Reward (\(\mathcal{R}_{\text{CSCA}}\)): - Targets the key challenge of feature overlap between new and old classes in FSCIL. - Computes cosine distances between generated images and each class prototype, dynamically assigning class weights accordingly. - Deliberately encourages generation of hard samples located in confusion regions (i.e., samples close to easily confused old classes). - Guides the classifier to learn fine-grained inter-class distinctions via weighted log-probability. - Optionally selects Top-K most similar prototypes to reduce computational cost.

Key Experimental Results

Main Results

Dataset DCS Average Accuracy Prev. SOTA Gain
miniImageNet 68.14% 67.05% (SAVC) +1.09
CUB-200 69.73% 69.35% (SAVC) +0.38
CIFAR-100 See ablation +5.64 (vs. baseline)

Ablation Study (CIFAR-100, improvement on the last session)

Component Combination \(\Delta_{\text{last}}\)
\(\mathcal{R}_{\text{PAMMD}}\) +1.24
+ \(\mathcal{R}_{\text{VM}}\) +1.86
+ \(\mathcal{R}_{\text{RC}}\) +3.50
+ \(\mathcal{R}_{\text{CSCA}}\) (full DCS) +5.64

Key Findings

  • Logits-level rewards (especially \(\mathcal{R}_{\text{CSCA}}\)) contribute the most, demonstrating that classifier feedback is critical to generation quality.
  • Under the constraint of generating fewer than 50 images per class, DCS surpasses vanilla diffusion models operating at much larger generation scales.

Highlights & Insights

  • Elegant closed-loop design: The interaction between the diffusion model and the classifier is elevated from unidirectional knowledge provision to bidirectional co-evolution, forming a mutual boosting loop.
  • Multi-level and complementary reward functions: Feature-level rewards handle semantic anchoring and diversity; logits-level rewards optimize decision boundaries; each component's contribution is verified through incremental ablation.
  • Efficient generation: Approximately 30 generated images per class suffice to achieve state-of-the-art performance, far fewer than the million-scale augmentations required by conventional methods.
  • Plug-and-play: Performance is improved solely through generative data augmentation without modifying the baseline classifier architecture.
  • Incremental computation for PAMMD and dimension-wise variance matching both reflect pragmatic engineering adaptations suited to data-scarce few-shot settings.

Limitations & Future Work

  • Reliance on large pretrained diffusion models such as Stable Diffusion 3.5 incurs non-trivial computational overhead during inference-time image generation.
  • The Sequential Monte Carlo sampling in DAS introduces additional cost, making deployment on edge devices or in real-time systems challenging.
  • Dimension-wise variance matching discards inter-dimensional feature correlations, potentially leading to information loss in high-dimensional feature spaces.
  • Evaluation is confined to standard FSCIL benchmarks (miniImageNet, CUB-200, CIFAR-100); more complex scenarios such as domain-incremental or open-world settings remain unexplored.
  • The classifier still follows the classic paradigm of frozen feature extractor plus prototype updates; combining DCS with more advanced incremental learning strategies may yield further improvements.
  • vs. traditional FSCIL methods (TOPIC, CEC, FACT, SAVC, etc.): These methods rely on network optimization techniques such as self-supervised learning and distribution calibration; DCS surpasses them without modifying the classifier network, relying solely on generative augmentation.
  • vs. diffusion model augmentation (direct use of SD): Vanilla diffusion models perform substantially worse than DCS under small generation budgets (<50 images per class); DCS achieves "fewer but better" generation through reward guidance.
  • vs. generative incremental learning (e.g., SDAFL [1]): Conventional approaches lack classifier feedback; DCS achieves adaptive generation via closed-loop reward signals.
  • vs. original DAS applications: Original DAS uses image quality rewards such as HPSv2/TCE; DCS redesigns the reward function specifically for FSCIL tasks.

Additional Insights

  • The generative-discriminative synergy paradigm merits extension to other data-scarce scenarios such as medical imaging and long-tail recognition.
  • The confusion-aware reward design offers inspiration for hard negative mining in contrastive learning.
  • DCS provides a training-free paradigm for efficiently adapting large generative models to downstream small-scale tasks (via reward guidance rather than fine-tuning).
  • The closed-loop mutual boosting mechanism conceptually parallels RLHF-style human feedback-guided generation; richer reward signal sources warrant further exploration.

Rating

  • Novelty: 8/10 — The closed-loop mutual boosting design between the diffusion model and the classifier is innovative, and the multi-level reward function is well-structured.
  • Experimental Thoroughness: 7/10 — Three standard benchmarks with detailed ablation studies, but lacks validation on more complex scenarios and computational cost analysis.
  • Writing Quality: 8/10 — Problem motivation is clearly articulated, method derivation is rigorous, and figures aid comprehension.
  • Value: 7/10 — Establishes a new paradigm for leveraging generative models in FSCIL, though practical deployment remains constrained by diffusion model overhead.