StyleAvatar: StyleAvatar takes too long to train
Thanks for the great repo.
I’m training the Full StyleAvatar, specifically with the command python train.py --batch 3 path-to-dataset. Training from scratch as the checkpoints have not been shared yet.
On the A10 GPU, it takes about a week to run for the default training parameters. Is that normal? I ask because the paper mentioned
The proposed network can converge within two hours while ensuring high image quality and a forward rendering time of only 20 milliseconds.
So maybe I’m missing something, can you help? 😃
About this issue
- Original URL
- State: closed
- Created a year ago
- Comments: 29 (13 by maintainers)
Yes, I also think its quite hard to understand, so I have updated this part into the faceverse code. But I’m too lazy, so I didn’t change the readme here, I hope users can read the code and find this part by themselves lol.
Yes, just re-run it from the checkpoint when you see the discriminator loss is always lower than 0.1.
Thank you for the message. This bug has been fixed now.
Fixed now. It was mainly caused by the flickering tracking of FaceVerse. Smooth term in preprocessing has been updated. The pretrained model uploaded can perform like this:
https://github.com/LizhenWangT/StyleAvatar/assets/26791093/a4ac5f63-6b7b-47da-baca-97092a0d2025
Just stop when the generated image meets your requirements (several hours). Don’t care about the default steps.