Nautilus: Locality-aware Autoencoder for Scalable Mesh Generation¶
Conference: ICCV2025 arXiv: 2501.14317 Code: - Area: 3D Vision Keywords: mesh generation, autoregressive, tokenization, locality-aware, point cloud conditioning
TL;DR¶
Nautilus proposes a locality-aware autoencoder for scalable artist-like mesh generation. By introducing a nautilus-shell-structured mesh tokenization algorithm that reduces sequence length to 1/4 of the naive baseline, and combining it with a dual-stream point cloud conditioner to improve local structural fidelity, Nautilus achieves for the first time direct high-quality mesh generation with up to 5,000 faces.
Background & Motivation¶
- Problem Definition: Automatically generating high-quality triangular meshes that exhibit artist-like compactness, structural regularity, and topological correctness.
- Limitations of Prior Work:
- Intermediate representation methods (NeRF, 3DGS, SDF, etc.): Convert to meshes via marching cubes or Poisson reconstruction, yielding overly dense and suboptimal meshes lacking continuous surface quality.
- Direct mesh generation (MeshGPT, MeshAnything): Autoregressively model vertices and faces, but suffer from two fundamental problems:
- Poor local structural fidelity: Manifold defects such as surface holes, overlapping faces, and missing parts, especially under complex topology.
- Limited face count: Excessively long sequences constrain the maximum number of faces, making it difficult to capture fine topological details.
- Core Insight: The locality of manifold meshes—adjacent faces share edges, and each face cluster converges to a central vertex—provides two key inspirations:
- Prioritizing local dependencies among neighbors ensures precise face interconnection and manifold validity.
- Explicitly modeling locally shared edges and vertices significantly compresses sequence length.
Method¶
Overall Architecture¶
- Nautilus-shell mesh tokenization: Serializes artist meshes into compact token sequences.
- Dual-stream point cloud conditioner: Provides geometric guidance for both global consistency and local structural fidelity.
- Autoregressive sequence decoder: Transformer-based decoding supporting point cloud or single-view image conditioning.
Key Design 1: Nautilus-Shell Mesh Tokenization¶
Problem with the naive approach: Flattening the 3 vertices of each face into \(9N\) coordinates fails to preserve spatial adjacency and yields excessively long sequences.
Nautilus-shell representation: Mesh faces are organized into multiple shells, each centered on a vertex \(O\) and containing an ordered sequence of surrounding vertices \(P\). Each face is composed of \(O\) and two adjacent \(P\) vertices. This allows \(N\) faces to be represented with only \(N+2\) vertices instead of \(3N\):
Adding a new face requires extending the sequence by only one vertex, greatly compressing the sequence.
Consecutive shell traversal: Upon completing a shell, the neighbor with the highest degree among the last traversed vertex's neighbors is selected as the center of the next shell. This ensures continuity and spatial proximity between shells.
Coordinate compression: 3D coordinates \((x, y, z)\) are mapped to 2D space \((u, v)\):
With resolution 128 and multiplier 2048, separate codebooks are built for \(u\) (size 1024) and \(v\) (size 2048). The \(u\) coordinate of the center vertex \(O\) uses an independent codebook of size 1024 to mark the start of each shell, eliminating the need for special separator tokens.
Compression efficiency: The resulting sequence length is only 1/4 of the naive approach (compression ratio 0.275, vs. 0.462 for AMT and 0.474 for EdgeRunner).
Key Design 2: Dual-Stream Point Cloud Conditioner¶
Global point cloud encoder: Uses the Michelangelo encoder to extract global features \(f_{glb}\), which serve as keys and values in the decoder's cross-attention layers, providing overall shape information.
Local point cloud encoder: Uses a PointConv module \(f_{loc}(\cdot)\) to capture local geometric information. For each shell's center vertex \(O_k\), KNN sampling selects 100 nearest points to extract local features \(f_{loc}(O_k)\), which are injected into the token features of each vertex \(P_{k,i}\) within the shell (inspired by ControlAR).
Tight integration with the shell structure: Feature injection from the local encoder is synchronized step-by-step with shell generation, enabling progressive local geometric constraints.
Autoregressive Decoding and Training¶
Generation: Follows the next-token prediction paradigm:
Training loss: Cross-entropy loss
Image conditioning: Leveraging Michelangelo's multimodal-aligned feature space, the model is trained with a frozen point cloud encoder and at inference time the encoder is replaced with an image encoder for single-image conditioned generation.
Key Experimental Results¶
Main Results: Quantitative Comparison¶
| Method | C.Dist. ↓ | H.Dist. ↓ | User Preference ↑ |
|---|---|---|---|
| MeshAnything | 0.133 | 0.293 | 10.27% |
| MeshAnythingV2 | 0.106 | 0.248 | 13.17% |
| Nautilus | 0.087 | 0.176 | 88.68% |
Tokenization Algorithm Comparison¶
| Metric | AMT | EdgeRunner | Nautilus |
|---|---|---|---|
| Compression Ratio ↓ | 0.462 | 0.474 | 0.275 |
| Local Ratio ↑ | 0.378 | 0.461 | 0.554 |
Key Findings¶
- Superior compression ratio: Nautilus tokenization achieves a compression ratio of 0.275, significantly outperforming AMT (0.462) and EdgeRunner (0.474), enabling generation of meshes with up to 5,000 faces.
- Strong locality preservation: A local ratio of 0.554 indicates that adjacent tokens in the sequence are also highly likely to be geometrically adjacent in the mesh, facilitating the autoregressive model's learning of local dependencies.
- Overwhelming user preference: An 88.68% user preference rate far surpasses MeshAnything (10.27%) and V2 (13.17%).
- Using coordinate compression or the shell structure alone is insufficient; their combination is necessary to achieve optimal quality and scalability.
- The local point cloud encoder provides critical improvements in regions with complex topology, such as localized holes.
Highlights & Insights¶
- Precise problem identification: The method accurately targets the fundamental bottleneck of current direct mesh generation approaches—neglect of locality inherent in manifold meshes.
- Elegant shell-based tokenization: Mimics the way artists construct meshes (fan-like expansion around a central vertex), simultaneously preserving locality and achieving high compression.
- Dual-stream conditioning mechanism: Global stream preserves overall shape; local stream maintains topological details—perfectly synergistic with the shell-based generation paradigm.
- First-ever direct generation of 5,000-face meshes: A substantial leap compared to MeshGPT (800 faces) and MeshAnything (1,600 faces).
- New evaluation metric—Local Ratio: Quantifies the preservation of local dependencies after serialization, offering broad utility for evaluating mesh tokenization algorithms.
Limitations & Future Work¶
- Inference remains slow: generating a 5,000-face mesh requires approximately 4 minutes.
- Heavy reliance on test samples generated by downstream image-to-3D methods, which may not align with the distribution of real artist meshes.
- The tokenization algorithm's handling of non-manifold meshes is not discussed.
- Training requires 311K high-quality artist meshes, imposing substantial data acquisition costs.
- The greedy traversal strategy for shell construction (selecting the highest-degree neighbor) may not be globally optimal.
Related Work & Insights¶
- Evolution of mesh tokenization: MeshGPT (VQ-VAE) → MeshAnythingV2 (AMT) → EdgeRunner (half-edge) → Nautilus (shell structure)—the trend is a shift from general-purpose tokenization toward specialized tokenization that exploits mesh geometric structure.
- Importance of locality in sequence modeling: Analogous to the role of positional encodings in NLP, preservation of spatial locality in 3D mesh tokenization is critical for generation quality.
- Hierarchical conditioning design: The dual-stream global+local conditioning paradigm is generalizable to other generation tasks requiring multi-scale control.
- Balance between compression and quality: Higher compression ratios allow models to train on more complex samples, serving as a key lever for raising the performance ceiling.
Rating ⭐⭐⭐⭐⭐¶
The paper precisely identifies locality as the key bottleneck, proposes an elegant solution via shell-structured tokenization and dual-stream conditioning, and delivers strong results: 4× compression, 88.68% user preference, and first-ever 5,000-face direct generation. The new Local Ratio metric has broad applicability. The work is highly complete and represents an important advance in direct mesh generation.