M3ashy: Multi-Modal Material Synthesis via Hyperdiffusion
DOI:
https://doi.org/10.1609/aaai.v40i16.38363Abstract
High-quality material synthesis is essential for replicating complex surface properties to create realistic scenes. Despite advances in the generation of material appearance based on analytic models, the synthesis of real-world measured BRDFs remains largely unexplored. To address this challenge, we propose M^3ashy, a novel multi-modal material synthesis framework based on hyperdiffusion. M^3ashy enables high-quality reconstruction of complex real-world materials by leveraging neural fields as a compact continuous representation of BRDFs. Furthermore, our multi-modal conditional hyperdiffusion model allows for flexible material synthesis conditioned on material type, natural language descriptions, or reference images, providing greater user control over material generation. To support future research, we contribute two new material datasets and introduce two BRDF distributional metrics for more rigorous evaluation. We demonstrate the effectiveness of M^3ashy through extensive experiments, including a novel statistics-based constrained synthesis, which enables the generation of materials of desired categories.Published
2026-03-14
How to Cite
Zhou, C., Hu, Z., Sztrajman, A., Cai, Y., Liu, Y., & Oztireli, C. (2026). M3ashy: Multi-Modal Material Synthesis via Hyperdiffusion. Proceedings of the AAAI Conference on Artificial Intelligence, 40(16), 13575-13583. https://doi.org/10.1609/aaai.v40i16.38363
Issue
Section
AAAI Technical Track on Computer Vision XIII