The task of this challenge is to accurately reconstruct a full 3D textured mesh from a partial 3D human scan acquisition acquired with the Shapify booth of Artec3D, similar in its quality to the 3DBodyTex dataset presented at 3DV’18 but collected following a different protocol. We refer to this new dataset 3DBodyTex.v2. 3DBodyTex.v2 is a new original dataset that will be shared with the research community after signing a data license agreement. It consists of about 2500 clothed scans with a large diversity in clothing and in poses. This covers 500 different subjects in up to 6 poses each, as well as an extended version of the 3DBodyTex data consisting of over 800 scans of 230 people in tight-fitting clothing, with up to 3 poses per subject.
The training data consists of pairs, (X, Y), of partial and complete scans, respectively. The goal is to recover Y from X. As part of the challenge, we share routines to generate partial scans X from the given complete scans Y.
However, the participants are free to devise their own way of generating partial data, as an augmentation or a complete alternative.
Any custom procedure should be reported with description and implementation among the deliverable.
A quality-check is performed to guarantee a reasonable level of defects.
The partial scans are generated synthetically.
For privacy reasons, all meshes are anonymized by blurring the shape and texture of the faces, similarly to the 3DBodyTex data.
During evaluation, the face and hands are ignored because the shape from raw scans is less reliable.
More information can be found here: https://gitlab.uni.lu/cvi2/eccv2020-sharp-workshop/