M3ashy: Multi-Modal Material Synthesis via Hyperdiffusion

Authors

  • Chenliang Zhou University of Cambridge
  • Zheyuan Hu University of Cambridge
  • Alejandro Sztrajman University of Cambridge
  • Yancheng Cai University of Cambridge
  • Yaru Liu University of Cambridge
  • Cengiz Oztireli University of Cambridge

DOI:

https://doi.org/10.1609/aaai.v40i16.38363

Abstract

High-quality material synthesis is essential for replicating complex surface properties to create realistic scenes. Despite advances in the generation of material appearance based on analytic models, the synthesis of real-world measured BRDFs remains largely unexplored. To address this challenge, we propose M^3ashy, a novel multi-modal material synthesis framework based on hyperdiffusion. M^3ashy enables high-quality reconstruction of complex real-world materials by leveraging neural fields as a compact continuous representation of BRDFs. Furthermore, our multi-modal conditional hyperdiffusion model allows for flexible material synthesis conditioned on material type, natural language descriptions, or reference images, providing greater user control over material generation. To support future research, we contribute two new material datasets and introduce two BRDF distributional metrics for more rigorous evaluation. We demonstrate the effectiveness of M^3ashy through extensive experiments, including a novel statistics-based constrained synthesis, which enables the generation of materials of desired categories.

Downloads

Published

2026-03-14

How to Cite

Zhou, C., Hu, Z., Sztrajman, A., Cai, Y., Liu, Y., & Oztireli, C. (2026). M3ashy: Multi-Modal Material Synthesis via Hyperdiffusion. Proceedings of the AAAI Conference on Artificial Intelligence, 40(16), 13575-13583. https://doi.org/10.1609/aaai.v40i16.38363

Issue

Section

AAAI Technical Track on Computer Vision XIII