Notice:
① As explained by the authors of EG3D (please see the tri-plane issue), ffhq512-128.pkl and ffhqrebalanced512-128.pkl were achieved using buggy [XY, XZ, ZX] planes. As they suggested in the issue, we trained all of our EG3D-based models using fixed tri-plane [XY, XZ, ZY]. So be careful if you use set reload_modules=True
in EG3D generation, please make sure to use the [XY, XZ, ZY] triplane when you reload modules.
② var3-64.pkl and var3-128.pkl are finetuned by fixing the camera parameters that are inputted into the generator as
var1-64.pkl : The model
$E^{Ours}_{var1}$ in our paper, with neural rendering resolution of$64^2$ . Trained on FFHQ+LPFF.var1-128.pkl : The model
$E^{Ours}_{var1}$ in our paper, with neural rendering resolution of$128^2$ . Trained on FFHQ+LPFF.var2-64.pkl : he model
$E^{Ours}_{var2}$ in our paper, with neural rendering resolution of$64^2$ . Trained on FFHQ+LPFF-rebal.var2-128.pkl : The model
$E^{Ours}_{var2}$ in our paper, with neural rendering resolution of$128^2$ . Trained on FFHQ+LPFF-rebal.var3-64.pkl : he model
$E^{Ours}_{var3}$ in our paper, with neural rendering resolution of$64^2$ . Trained on FFHQ+LPFF.var3-128.pkl : The model
$E^{Ours}_{var3}$ in our paper, with neural rendering resolution of$128^2$ . Trained on FFHQ+LPFF.
To provide a fairer comparison, we also retrained EG3D using [XY, XZ, ZY] plane on the FFHQ dataset:
ffhq-fixed-triplane512-64.pkl : EG3D model trained with FFHQ dataset and has tri-plane fixed (using [XY, XZ, ZY]), with neural rendering resolution of
$64^2$ . Trained on FFHQ.ffhq-fixed-triplane512-128.pkl : EG3D model trained with FFHQ dataset and has tri-plane fixed (using [XY, XZ, ZY]), with neural rendering resolution of
$128^2$ . Trained on FFHQ.
model | ||||||||
---|---|---|---|---|---|---|---|---|
ffhq512-128.pkl |
6.523 | 23.598 | 4.273 | 22.318 | 23.698 | 36.641 | 4.025 | 23.301 |
ffhq-fixed-triplane512-128.pkl |
7.689 | 23.962 | 6.572 | 22.537 | 22.567 | 33.063 | 6.102 | 25.115 |
var1-128.pkl |
7.997 | 20.896 | 6.623 | 19.738 | 21.300 | 22.074 | 6.093 | 16.026 |
ffhqrebalanced512-128.pkl |
6.589 | 20.081 | 4.456 | 19.983 | 19.469 | 30.181 | 4.262 | 23.717 |
var2-128.pkl |
9.829 | 16.775 | 6.672 | 15.047 | 13.022 | 14.836 | 6.571 | 12.221 |
var3-128.pkl |
6.536 | 15.852 | / | / | / | / | / | / |
FFHQ_LPFF.pkl : The model
$S^{Ours}_{var1}$ in our paper. Trained on FFHQ+LPFF.FFHQ_LPFF_rebalanced_maxsize89590.pkl : The model
$S^{Ours}_{var2}$ in our paper. We mistakenly achieved this model usingmax_size = 89590
. Notice that the dataset was shuffled before being clipped tosize=89590
, so the pose distribution are not affected much in this pretrained model. We additionally achieved a model using the entire dataset, please see below. Trained on FFHQ+LPFF-rebal (max_size = 89590).FFHQ_LPFF_rebalanced.pkl (recommended): This model was achieved using the same training strategy and dataset as
FFHQ_LPFF_rebalanced_maxsize89590.pkl
, but without a max size limit. Trained on FFHQ+LPFF-rebal.