Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What's the difference between vox-full.yaml and vox.yaml? #16

Open
subin6 opened this issue Sep 16, 2019 · 5 comments
Open

What's the difference between vox-full.yaml and vox.yaml? #16

subin6 opened this issue Sep 16, 2019 · 5 comments

Comments

@subin6
Copy link

subin6 commented Sep 16, 2019

I want to train your model with VoxCeleb dataset.
Which configuration do I use between them?

Also, could you provide VoxCeleb pretrain model?

Thank you

@AliaksandrSiarohin
Copy link
Owner

AliaksandrSiarohin commented Sep 16, 2019

Hi,
The difference between the vox-full.yaml and vox.yaml is the fact that vox.yaml run keypoint detector and dense motion on downscaled 64x64 images. On the other hand vox-full train all networks on 256x256. I suggest to use vox.yaml because it is faster, also you can download VoxCeleb dataset with my preprocessing (difference with original is that face aspect ratio is preserved, backround is not moving, and all the small resolution videos is removed) (https://github.com/AliaksandrSiarohin/video-preprocessing).

Here there is a https://yadi.sk/d/A0Jq_01xiXky3g which I use to compare with my newer model. The corresponding config is vox.yaml. The movement there is quite limited compared to newer model, here is the example:
id10289#Rn0Z_lIiL1w#00001.txt#000.mp4-id10286#9K2YB1d8BqY#00008.txt#000.mp4.png.mp4.zip

By columns: source,driving,new model,monkey-net,x2face

@kashi211
Copy link

i'm sorry but is the "new" model you're reffereing to is 'vox-full.yaml' and could you provide me with the cpkt file?

@AliaksandrSiarohin
Copy link
Owner

No. It is https://github.com/AliaksandrSiarohin/first-order-model
But I can not provide a checkpoints yet, because of some privacy related issues.

@JialeTao
Copy link

JialeTao commented Mar 2, 2021

Hi,
The difference between the vox-full.yaml and vox.yaml is the fact that vox.yaml run keypoint detector and dense motion on downscaled 64x64 images. On the other hand vox-full train all networks on 256x256. I suggest to use vox.yaml because it is faster, also you can download VoxCeleb dataset with my preprocessing (difference with original is that face aspect ratio is preserved, backround is not moving, and all the small resolution videos is removed) (https://github.com/AliaksandrSiarohin/video-preprocessing).

Here there is a https://yadi.sk/d/A0Jq_01xiXky3g which I use to compare with my newer model. The corresponding config is vox.yaml. The movement there is quite limited compared to newer model, here is the example:
id10289#Rn0Z_lIiL1w#00001.txt#000.mp4-id10286#9K2YB1d8BqY#00008.txt#000.mp4.png.mp4.zip

By columns: source,driving,new model,monkey-net,x2face

Hi, noted that the datasets of monkey-net and your new model used are diffrent, inculding different vesions of taichi and vox. I'm making a comparison with monkey-net and FOMM on the dataset provided in FOMM, could you provide the monkey-net checkpoints of vox, taichi and bair, that you used to compare with your new model ? Thanks a lot.

@AliaksandrSiarohin
Copy link
Owner

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants