Skip to content

DiT-IC: Aligned Diffusion Transformer for Efficient Image Compression

Conference: CVPR 2026 arXiv: 2603.13162 Code: Project Page Area: Image Compression / Generative Models Keywords: diffusion transformer, image compression, one-step diffusion, flow matching, latent alignment, variance-guided

TL;DR

This work adapts a pretrained text-to-image DiT (SANA) into an efficient single-step image compression decoder. Three alignment mechanisms are proposed: variance-guided reconstruction flow (pixel-level adaptive denoising intensity), self-distillation alignment (encoder latents as distillation targets), and latent-conditioned guidance (replacing the text encoder). Operating entirely in a deep latent space with 32× downsampling, the method achieves state-of-the-art perceptual quality (BD-rate DISTS −87.88%), decodes 30× faster than prior diffusion-based methods, and can reconstruct 2K images on a 16 GB laptop GPU.

Background & Motivation

Background: Diffusion-based image compression achieves strong perceptual fidelity (PerCo, DiffEIC, ResULIC, StableCodec), but is constrained by multi-step sampling overhead and high memory consumption. Existing methods predominantly use U-Net architectures, whose hierarchical downsampling forces diffusion to operate in shallow latent spaces (8× downsampling), whereas conventional VAE codecs can operate in much deeper latent spaces (16×–64×).

Limitations of Prior Work:

  1. U-Net multi-step diffusion in 8× shallow latent spaces incurs heavy computation and memory overhead (e.g., DiffEIC requires 12.4 s for 50 steps).
  2. Single-step methods (StableCodec, OSCAR) still rely on U-Net and cannot natively perform diffusion in deep latent spaces.
  3. Directly transplanting generative DiTs into compression latent spaces causes severe degradation—the generative objective (denoising from pure noise) fundamentally misaligns with the reconstruction objective (single-step recovery from structured quantized latents).

Key Challenge: The generative prior of diffusion models benefits perceptual reconstruction, yet the paradigm of iterative denoising from pure noise is fundamentally incompatible with the compression requirement of single-step reconstruction from known structured latents.

Goal: Enable efficient diffusion in an extremely compact deep latent space (32×) by collapsing multi-step iteration into a deterministic single-step transformation.

Key Insight: Three "alignment" mechanisms bridge the gap between generation and compression—aligning denoising intensity (variance → timestep), aligning multi-step to single-step (self-distillation), and aligning the conditioning modality (text → latent).

Core Idea: Compressed quantized latents already lie close to the data manifold; their spatial variance naturally encodes local "denoising demand." Mapping variance to pseudo-timesteps collapses iterative denoising into a single-step adaptive reconstruction.

Method

Overall Architecture

The framework builds on pretrained SANA (a text-to-image DiT) with ELIC as an auxiliary encoder. The encoder compresses images into a 64× downsampled coding space and applies entropy coding via a hyperprior combined with an autoregressive context model (using lightweight DepthConvBlocks). The DiT performs single-step diffusion reconstruction in the 32× latent space, and the decoder maps the reconstructed latents back to pixel domain. Parameter-efficient adaptation is achieved via LoRA (VAE decoder rank 32, DiT rank 64). The NoPE (No Positional Encoding) design naturally supports resolution generalization.

Key Designs

  1. Variance-Guided Reconstruction Flow

    • Function: Collapses multi-step denoising into a spatially adaptive single-step transformation.
    • Mechanism: Compression quantization noise is spatially heterogeneous—smooth regions exhibit low noise (small timestep) while textured regions exhibit high noise (large timestep). The encoder-predicted variance \(\boldsymbol{\sigma}\) is mapped through a differentiable function to produce pixel-wise pseudo-timesteps \(t = \mathcal{F}(\text{proj}_\theta(\boldsymbol{\sigma})) \in \mathbb{R}^{H \times W}\). Single-step reconstruction: \(\hat{\mathbf{y}} = \tilde{\mathbf{y}} - \mathbf{v}_\theta(\tilde{\mathbf{y}}, t)\).
    • Design Motivation: A single global timestep cannot accommodate local noise heterogeneity. Ablations confirm that variance is a by-product already available from the encoder at zero additional cost.
  2. Self-Distillation Alignment

    • Function: Stabilizes single-step learning without an external teacher.
    • Mechanism: The encoder is frozen, and its latent output \(\mathbf{y}_0\) serves as the self-supervised target. A margin cosine alignment loss is applied: \(\mathcal{L}_{\text{distil}} = \mathbb{E}[1 - m - \frac{\langle \hat{\mathbf{y}}, \mathbf{y}_0 \rangle}{|\hat{\mathbf{y}}|_2 |\mathbf{y}_0|_2}]\). The encoder is frozen while the DiT and decoder are jointly optimized.
    • Design Motivation: In the compression setting, no multi-step diffusion trajectory exists to distill from; however, encoder outputs already reside close to the data manifold and are naturally suited as single-step distillation targets.
  3. Latent-Conditioned Guidance

    • Function: Replaces the text prompt with compressed latents as the DiT conditioning signal, eliminating the text encoder at inference.
    • Mechanism: A lightweight projection \(c_{\text{lat}} = \text{Proj}_\psi(\hat{y})\) maps latents into the text embedding space. During training, a CLIP-style contrastive loss \(\mathcal{L}_{\text{cond}}\) aligns \(c_{\text{lat}}\) with InternVL-generated \(c_{\text{text}}\); at inference, only the latent condition is used.
    • Design Motivation: Text prompts are inefficient for reconstruction tasks and introduce stochasticity; the latent representation itself already encodes rich semantic structure.

Loss & Training

Training follows a two-stage implicit bitrate pruning (IBP) scheme: Stage 1 uses \(\lambda_{\text{base}} \in \{0.1, 0.5\}\) for 100K iterations with 256² patches (batch size 32); Stage 2 uses \(\lambda_{\text{target}} \in \{0.5\text{–}16.0\}\) for 60K iterations with 512² patches (batch size 16), incorporating adversarial loss. The total loss is \(\lambda\mathcal{R} + \mathcal{D} + \mathcal{L}_{\text{align}} + \lambda_{\text{adv}}\mathcal{L}_{\text{adv}}\), where \(\mathcal{D} = \lambda_1\text{MSE} + \lambda_2\text{LPIPS} + \lambda_3\text{DISTS}\). Optimization uses AdamW (lr=1e-4) with EMA 0.999, trained on two RTX Pro 6000 GPUs.

Key Experimental Results

Main Results

BD-rate comparison (vs. PerCo baseline, ↓ better, averaged over three datasets)

Method Diffusion Steps Latent Space Latency (1024²) LPIPS BD-rate↓ DISTS BD-rate↓
PerCo (ICLR'24) 20 f8 8.8s 0.00% 0.00%
DiffEIC (TCSVT'24) 50 f8→f16 12.4s −36.14% −33.72%
ResULIC (ICML'25) 4 f8→f32 0.83s −62.27% −65.64%
StableCodec (ICCV'25) 1 f8→f64 0.34s −79.19% −83.95%
OSCAR (NeurIPS'25) 1 f8→f64 0.32s −19.04% −58.38%
DiT-IC 1 f32→f64 0.15s −83.65% −87.88%

Ablation Study

Ablation of key designs (BD-rate DISTS, relative to full DiT-IC)

Configuration DISTS BD-rate Note
Full DiT-IC 0.00% Baseline
w/o adversarial loss −1.80% Adversarial loss enhances perceptual sharpness
w/o DISTS loss +5.69% DISTS loss is critical for human perceptual alignment
DiT trained from scratch +32.45% Pretrained weights are essential
LoRA rank 16/16 +13.92% Insufficient rank limits adaptation capacity
Full fine-tuning +8.05% Small batches disrupt the pretrained distribution

Key Findings

  • DiT-IC achieves comprehensive superiority on both LPIPS and DISTS perceptual metrics across all three datasets.
  • At 4096² resolution, diffusion latency is reduced by 95% compared to StableCodec (10.3s → 0.47s).
  • Pretrained weights are critical: training from scratch degrades DISTS BD-rate by 32.45%.
  • LoRA rank 32/64 is optimal; full fine-tuning performs worse due to small-batch distribution distortion.
  • A user study shows 56.8% preference for DiT-IC vs. 27.5% for StableCodec.
  • After INT8 quantization, the model runs on 4 GB of VRAM, enabling deployment on consumer-grade GPUs.

Highlights & Insights

  • This is the first work to apply a DiT to image compression operating entirely within a 32× deep latent space, breaking the architectural bottleneck imposed by U-Net designs.
  • Each of the three alignment mechanisms addresses a concrete practical problem with a clean design—variance-to-timestep reuses existing encoder information, self-distillation requires no external teacher, and latent conditioning eliminates the text encoder.
  • The pixel-level adaptive mapping from variance to pseudo-timestep is particularly intuitive: the spatial heterogeneity of quantization noise naturally encodes local denoising demand.
  • The NoPE design enables stable resolution generalization up to 4096² without modification.

Limitations & Future Work

  • At very low bitrates (<0.01 bpp), latent-only conditioning may be insufficient; auxiliary text priors could be beneficial.
  • Training uses only 150K images; larger-scale data may yield further improvements.
  • Joint fine-tuning of the encoder is unexplored; the current frozen-encoder setup leaves theoretical room for improvement.
  • Adversarial distillation techniques (e.g., ADD) have not been integrated and could further enhance perceptual realism.
  • Semantic consistency at low bitrates remains an area for improvement.
  • vs. StableCodec: Both perform single-step diffusion compression, but StableCodec uses U-Net diffusion in f8 space, whereas DiT-IC uses a DiT in the deep f32 space—yielding a 25× speedup at 4096² resolution.
  • vs. ResULIC: DiT-IC achieves better BD-rate while reducing from 4 steps to 1, validating the sufficiency of single-step reconstruction.
  • vs. OSCAR: OSCAR maps image-level bitrate to a global timestep; DiT-IC extends this to pixel-level variance-to-timestep mapping with finer granularity.
  • Insights: The "alignment" paradigm is worth extending to other low-level vision tasks such as super-resolution and inpainting; the self-distillation approach offers a useful reference for accelerating diffusion inference more broadly.

Rating

  • Novelty: ⭐⭐⭐⭐ First DiT-based image compression framework; all three alignment mechanisms are creative, and the variance-to-timestep mapping is intuitively elegant.
  • Experimental Thoroughness: ⭐⭐⭐⭐⭐ Three datasets, multiple metrics, multiple baselines, ablation studies, user study, latency analysis, and resolution generalization evaluation.
  • Writing Quality: ⭐⭐⭐⭐ Every design choice is supported by ablation; figures are intuitive and the structure is clear.
  • Value: ⭐⭐⭐⭐⭐ Single-step, low-latency, low-memory SOTA perceptual compression with genuine deployment potential.