Roy Shilkrot1, Daniel Cohen-Or2, Ariel Shamir3, Ligang Liu4
1MIT Media Laboratory, 2Tel-Aviv University, 3The Interdisciplinary Center, Herzliya, 4University of Science and Technology of China, Hefei
We present a method for transferring the identity of a given subject to a target image for try-on experience of clothes. The method involves cloning the user's identity into a catalogue of model images wearing the desired garments.
We present an accurate segmentation procedure for human heads that separates three semantic parts: face, hair, and background. We use a tri-kernel statistical model based on Textons and segment using graph cut.
Using an offline simple training phase the extracted head can be cloned automatically into photos of catalogue models.
The skin color is adjusted according to a statistical model, and the head is relighted using Spherical Harmonics. Lastly, the body dimensions are warped to fit the user's dimensions using a parametric model.
This creates high quality compositions imitating the identity of the user in the desired garment. We show some realistic results, and present a study that supports their quality.