Last week’s hottest news other than the drama unfolding at Open.AI with Sam, this was a very positive news when Stability.AI announced that they have released Stable Video Diffusion<\/a> (21st November 2023). A few days later the first two models are available for experimental\/evaluation use (ie. Non-commercial), so of course then we wait for the community to take this to next level and make it usable on your own PC.<\/p>\n\n\n\n
We have two models available:<\/p>\n\n\n\n
Here are some samples available and shared at the above links. These are smooth GIF images.<\/p>\n\n\n\n
Now that you have the links to the models you can download the model file (.safetensors) which is about ~10 GB, depending upon what you prefer to use. I will be using Stable Video Diffusion XT version to push my RTX4080 16GB card<\/a>.<\/p>\n\n\n\n
Next we need the ComfyUI workflow which is available for download<\/a>. There is one workflow for Text-to-Image-to-Video and another for Image-to-Video.<\/p>\n\n\n\n
I am going to experiment with Image-to-Video which I am further modifying to produce MP4 videos or GIF images using the Video Combine node included in ComfyUI-VideoHelperSuite<\/a><\/p>\n\n\n\n
The first experiment is using a frame that I have a frame from another project I’ve been working on. The result was pretty cool using the default settings of the workflow. The camera movement is pretty cool as the subject moves the camera is following her.<\/p>\n\n\n\n
This result was produced in 52 seconds in my setup.<\/p>\n\n\n\n