-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How much time do you need to lip sync a 10 sec or 1 minute video? #584
Comments
well I have tried this repository with colab only and it seems fine if you are trying to merge a video of a certain length basically up to 25-30 secs, then anything after that is gonna take a lot of time. Rather if you want to increase the performance of the model, These are some of the things you can try:
|
I have been running this model on 1080p input videos between 10-30 seconds long on my machine (rtx 3060 12gb vram) and have had to set the --rescale argument for inference.py to 3 to not run out of memory. To generate a lipsync'd clip it takes a little over a minute. I also had to modify the code in order to run this locally on my machine for the preprocessing and discriminator training scripts. If you want to get this working on your machine I would suggest using environment setup described here: https://github.com/natlamir/Wav2Lip-WebUI |
Thank you , will check it out. |
Ok thanks will cehck it, might contact you again if needs be. |
Sure |
Hello again @sahreen-haider |
Hello, The model can be changed with the pertained model for face recognition which is another library, |
Is it possible to get help on that? (maybe send me the modified version by PM if you want it to stay not too much spread, I will only use it myself) |
Hey @AIhasArrived, But I would require some significant time to do this grunt work, unfortunately I might not be able to do this at this time. But rather you have asked for any alternatives for this, You might want to check it out. |
@AIhasArrived, Connect with me over this email: sahreenhaider@gmail.com |
Already did: sent you an email few days ago titled "Contact from github :)" |
I can sync 8 sec video in like 15s and time could improve with better parameters. But ,when started i had really 4x slower time and i realized something was just wrong , starting chunks were loading really slow. After doing some research i realized problem is new Torch and GPU not working properly. By following other topics i did try with older versions ex " torch==2.0.1+cu118 and my chunk loading speed increased drastically. Hope it helps, and i hope they fix this shit with a new version. |
Hello @davidkundrats I just tried this repo, it looks nice but when I run it I got into a problme (nothing happening while GPU is being used) did you get that problem yourself? and if yes what did you do to solve it? thanks |
I have been trying the last days with both wav2lip HD (not in auto) and retalker, and found that both are slow and very GPU consuming.
I would like to know everyone of you HOW MUCH GPU do you use (what card) and HOW MUCH time does it take for you to do it? What kind of videos/animations are you lip syncing and for how long? (How much time to train X seconds/minutes?)
Please contribute. Because I am about to drop this technology and give up on it, maybe others peoples experiences will give me hope. Maybe this repo is faster? (could not try it yet)
The text was updated successfully, but these errors were encountered: