TANGLED : Generating 3D Hair Strands from Images with Arbitrary Styles and Viewpoints
Anonymous Authors
Anonymous Authors
Abstract
Hairstyles are intricate and culturally significant with various geometries, textures, and structures. Existing text or image-guided generation methods fail to handle the richness and complexity of diverse styles. We present TANGLED, a novel approach for 3D hair strand generation that accommodates diverse image inputs across styles, viewpoints, and quantities of input views. TANGLED employs a three-step pipeline. First, our MultiHair Dataset provides 457 diverse hairstyles annotated with 74 attributes, emphasizing complex and culturally significant styles to improve model generalization. Second, we propose a diffusion framework conditioned on multi-view linearts that can capture topological cues (e.g., strand density and parting lines) while filtering out noise. By leveraging a latent diffusion model with cross-attention on lineart features, our method achieves flexible and robust 3D hair generation across diverse input conditions. Third, a parametric post-processing module enforces braid-specific constraints to maintain coherence in complex structures. This framework not only advances hairstyle realism and diversity but also enables culturally inclusive digital avatars and novel applications like sketch-based 3D strand editing for animation and augmented reality.
Video
Overview
Our model takes hair images with arbitrary styles and viewpoints as conditions, and generate the 3D hair latent through the diffusion process. The conditions are randomly masked and cross-attention with the latent. At inference, we sample hair latent maps and feed the upsampled hair latent map to the strand decoder to extract the 3D hair strands.
Results
TANGLED can generate realistic hairstyles from image conditions with various styles, including photographs, anime, and oil paintings. For more results, please refer to the supplementary video. Note that we manually specified the color for the generated hair during the rendering process.
Row 1 shows the generated hairstyles from hand-drawn sketches. Row 2 illustrates hairstyle modifications(adding pigtails) by altering specific parts in the sketches from Row 1. Row 3-4 depict the process of generating outputs with braid using guidelines (highlighted in red).
With single-view input(Row 1), the asymmetrical shoulder-length hair on the occluded side cannot be accurately reconstructed, resulting in missing details. In contrast, multi-view inputs (Row 2) enable the model to capture the full structure of the hairstyle, including the previously occluded regions.
Citation
Notice: You may not copy, reproduce, distribute, publish, display, perform, modify, create derivative works, transmit, or in any way exploit any such content, nor may you distribute any part of this content over any network, including a local area network, sell or offer it for sale, or use such content to construct any kind of database.