Designing and simulating realistic clothing is challenging and, while several methods have addressed the capture of clothing from 3D scans, previous methods have been limited to single garments and simple motions, lack detail, or require specialized texture pat- terns. Here we address the problem of capturing regular clothing on fully dressed people in motion. People typically wear multiple pieces of clothing at a time. To estimate the shape of such clothing, track it over time, and render it believably, each garment must be segmented from the others and the body. Our ClothCap approach uses a new multi-part 3D model of clothed bodies, automatically segments each piece of clothing, estimates the naked body shape and pose under the clothing, and tracks the 3D deformations of the clothing over time. We estimate the garments and their mo- tion from 4D scans; that is, high-resolution 3D scans of the subject in motion at 60 fps. The model allows us to capture a clothed per- son in motion, extract their clothing, and retarget the clothing to new body shapes. ClothCap provides a step towards virtual try-on with a technology for capturing, modeling, and analyzing clothing in motion.



  • 15.06.2020: ClothCap registrations and scans are now publicly available, in addition to more captured data for CAPE. Downloads available at: https://cape.is.tue.mpg.de

More Information

Referencing the Dataset

Here are the Bibtex snippets for citing MPI Clothcap in your work.

        title = {ClothCap: Seamless 4D Clothing Capture and Retargeting},
        author = {Pons-Moll, Gerard and Pujades, Sergi and Hu, Sonny and Black, Michael},
        journal = {ACM Transactions on Graphics, (Proc. SIGGRAPH)},
        volume = {36},
        number = {4},
        year = {2017},
        note = {Two first authors contributed equally},
        crossref = {},
        url = {http://dx.doi.org/10.1145/3072959.3073711}