Affiliation:
1. The Chinese University of Hong Kong, Hong Kong, Hong Kong
2. Qianzhi Technology Inc., Shenzhen, China
Abstract
This paper presents a computational pipeline for creating personalized, physical LEGO
®1
figurines from user-input portrait photos. The generated figurine is an assembly of coherently-connected LEGO
®
bricks detailed with uv-printed decals, capturing prominent features such as hairstyle, clothing style, and garment color, and also intricate details such as logos, text, and patterns. This task is non-trivial, due to the substantial domain gap between unconstrained user photos and the stylistically-consistent LEGO
®
figurine models. To ensure assemble-ability by LEGO
®
bricks while capturing prominent features and intricate details, we design a three-stage pipeline: (i) we formulate a CLIP-guided retrieval approach to connect the domains of user photos and LEGO
®
figurines, then output physically-assemble-able LEGO
®
figurines with decals excluded; (ii) we then synthesize decals on the figurines via a symmetric U-Nets architecture conditioned on appearance features extracted from user photos; and (iii) we next reproject and uv-print the decals on associated LEGO
®
bricks for physical model production. We evaluate the effectiveness of our method against eight hundred expert-designed figurines, using a comprehensive set of metrics, which include a novel GPT-4V-based evaluation metric, demonstrating superior performance of our method in visual quality and resemblance to input photos. Also, we show our method's robustness by generating LEGO
®
figurines from diverse inputs and physically fabricating and assembling several of them.
Funder
The Research Grants Council of the Hong Kong Special Administrative Region
Publisher
Association for Computing Machinery (ACM)
Reference55 articles.
1. Personal Fabrication
2. C-Shells: Deployable Gridshells with Curved Beams
3. BrickLink. 2024. Bricklink Color Guide. https://www.bricklink.com/catalogColors.asp
4. Kaidi Cao, Jing Liao, and Lu Yuan. 2018. Carigans: Unpaired photo-to-caricature translation. arXiv preprint arXiv:1811.00222 (2018).
5. Wenhu Chen, Hexiang Hu, Yandong Li, Nataniel Rui, Xuhui Jia, Ming-Wei Chang, and William W Cohen. 2023a. Subject-driven text-to-image generation via apprenticeship learning. arXiv preprint arXiv:2304.00186 (2023).