Unlike previous methods that use three vertices to represent a face, AMT uses a single vertex whenever possible.
When this is impossible, AMT adds a special token & and restarts.
Our experiments demonstrate that AMT reduces the token sequence length by half on average. Its compact, and well-structured sequence representation enhances sequence learning, thereby significantly improving both the efficiency and performance of mesh generation.
We introduce MeshAnything V2, an autoregressive transformer that generates Artist-Created Meshes (AM) aligned to given shapes. It can be integrated with various 3D asset production pipelines to achieve high-quality, highly controllable AM generation. MeshAnything V2 surpasses previous methods in both efficiency and performance using models of the same size.
These improvements are due to our newly proposed mesh tokenization method: Adjacent Mesh Tokenization (AMT). Different from previous methods that represent each face with three vertices, AMT uses a single vertex whenever possible. Compared to previous methods, AMT requires about half the token sequence length to represent the same mesh in average. Furthermore, the token sequences from AMT are more compact and well-structured, fundamentally benefiting mesh generation. Our extensive experiments show that AMT significantly improves the efficiency and performance of mesh generation.
@misc{chen2024meshanythingv2artistcreatedmesh,
title={MeshAnything V2: Artist-Created Mesh Generation With Adjacent Mesh Tokenization},
author={Yiwen Chen and Yikai Wang and Yihao Luo and Zhengyi Wang and Zilong Chen and Jun Zhu and Chi Zhang and Guosheng Lin},
year={2024},
eprint={2408.02555},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2408.02555},
}