AI Video 2 Video Animation Beginners Guide , in Stable Diffusion and A1111

How to
21 Sept 202324:02

TLDRThis video tutorial is aimed at beginners interested in AI-generated videos, specifically using Stable Diffusion and A1111. The guide covers the process of creating video animations from a reference video, emphasizing that while flickering and inconsistency can't be entirely avoided, they can be minimized with the right understanding and settings. The video explains that there's no one-size-fits-all setting and that customization is key. It also highlights the use of image-to-image methods for simplicity and control, and the importance of selecting suitable videos with simple backgrounds for easier subject separation. The tutorial provides detailed steps for using DaVinci Resolve to refine the generated frames, including changing frame rates, applying noise multipliers, and using control maps to improve results. It also demonstrates how to remove backgrounds and use various tools within DaVinci Resolve to reduce flickering and enhance video quality. The guide concludes with tips on generating consistent characters, using special names and prompts, and the potential for creating deep fakes or face matches with the right settings and techniques.


  • 🎬 Use AI video generation tools like Stable Diffusion and A1111 to create videos from reference videos.
  • 🔍 Understand that flickering or inconsistency can't be completely avoided but can be reduced by adjusting settings based on the video content.
  • 🚀 Train your own checkpoint for the best results, but this is a time-consuming process, so using existing checkpoints is recommended for beginners.
  • 🖼️ Choose videos with simple backgrounds or green screens to facilitate easier subject-background separation for more complex animations.
  • 👕 Opt for consistent and simple clothing in the video to ensure smoother animations.
  • 📏 Set the resolution and frame rate according to the desired video style, like 720x720 or 16 frames per second for a 'loopy' style.
  • 🔄 Use the 'Image to Image' method for its simplicity, control, and effectiveness.
  • 🛠️ Experiment with different control Nets like 'Open Pose' and 'Normal Map' to see which works best for your specific video.
  • 🌟 Enable 'After Detailer' with a strength of 0.3 for better face and detail enhancement.
  • ✂️ Remove the background using tools in DaVinci Resolve or A1111 for a cleaner subject focus.
  • 🔄 Use 'D Flicker' in DaVinci Resolve to reduce flickering in the generated video.
  • 📉 Adjust the speed and re-time controls to match the desired pace of the final video.

Q & A

  • What is the main topic of the video?

    -The video is a beginner's guide on generating AI videos based on a reference video using Stable Diffusion and A1111.

  • What is the importance of having a simple background in the reference video?

    -A simple background or green screen makes it easier to separate the subject from the background, which is crucial for producing more complex animations without flickering.

  • What is the recommended method for beginners to generate AI videos in this process?

    -The recommended method is image-to-image, as it is effective, easy to use, and provides more control over the final result.

  • How can flickering or inconsistency in the generated video be reduced?

    -Flickering can be reduced by understanding the limitations of Stable Diffusion, selecting appropriate settings based on the video, and using techniques like D Flicker in DaVinci Resolve.

  • What is the role of training your own checkpoint in creating AI videos?

    -Training your own checkpoint can produce better and more impressive results, but it is a time-consuming process, so the video focuses on using existing checkpoints.

  • Why is it suggested to use a video with consistent clothing?

    -Consistent clothing, such as jeans or simple clothing, helps make the animation smoother and reduces the complexity of the video, which can lead to better results.

  • What is the significance of the frame rate in the video generation process?

    -The frame rate affects the smoothness and speed of the generated video. A higher frame rate can make the video smoother, but it also increases the processing time.

  • How can the generated images be made more consistent?

    -Consistency can be improved by using control nets like Open Pose for hand and body location, Normal Map for general scene mapping, and enabling After Detailer for better face and feature details.

  • What is the purpose of using the 'D Flicker' effect in DaVinci Resolve?

    -The 'D Flicker' effect is used to reduce flickering in the background of the generated video, resulting in a cleaner and more professional look.

  • How can the background be removed from the generated images?

    -The background can be removed using an extension in A1111 called 'Background Removal' or using the magic mask feature in DaVinci Resolve.

  • What is the impact of using multiple control nets on the video generation process?

    -Using multiple control nets can improve the quality of the generated video by providing more detailed control over the image, but it also increases the processing time.



🎨 Introduction to AI Video Generation

This paragraph introduces the video's focus on AI video generation for beginners. It discusses the use of stable diffusion and automatic 1111 to create videos based on a reference video. The narrator emphasizes that while flickering or inconsistency can't be entirely avoided, understanding the process and adjusting settings can help achieve better results. The video will demonstrate how to compose the generated videos in DaVinci Resolve, and the importance of selecting the right video and background for smoother animations is highlighted.


🖼️ Enhancing Video Quality with Post-Processing

The second paragraph delves into post-processing techniques to improve the quality of AI-generated videos. It covers the use of control Nets like normal and open pose for better mapping and detail enhancement. The paragraph also discusses the process of background removal using extensions in automatic 1111 and adjusting frame rates and speeds in DaVinci Resolve for a smoother video experience. Techniques such as Optical flow are mentioned for potentially smoother results.


🛠️ Reducing Flickering in AI Videos

This section focuses on strategies to reduce flickering in AI-generated videos using DaVinci Resolve. The use of the D flicker effect is explained, along with the caution that adding too many instances can degrade the image quality and increase rendering time. Alternative tools like floor light and dirt removal are also discussed, with a note on their varying effectiveness. The importance of testing different settings to find the best outcome is emphasized.


🌱 Advanced Techniques for Video Generation

The fourth paragraph explores advanced techniques for generating AI videos. It talks about the use of sampling steps and seed fixing for consistent results. The effectiveness of after detailer and in-painting for enhancing faces and other details is highlighted. The paragraph also covers the use of multiple control Nets and their impact on generation speed. It advises on the experimental nature of control Net usage and the importance of maintaining consistency between source and destination images.


📈 Batch Processing and Deepfake Creation

The final paragraph covers batch processing for generating a series of images with consistent characters and settings. It also touches on the use of image-to-image generation to create deepfakes or face matches by merging faces with the help of after detailer and in-painting. The narrator provides a tip on using a lower denosing level for a deepfake-like effect and concludes with a reminder to check each frame for quality before converting to video.



💡AI Video Generation

AI Video Generation refers to the process of creating videos using artificial intelligence. In the context of the video, it involves generating videos based on a reference video using AI tools like Stable Diffusion and A1111. The process can include image to image transformations, which are essential for creating animations or modifying existing video content.

💡Stable Diffusion

Stable Diffusion is an AI model used for generating images and videos from textual descriptions. It is mentioned in the video as a tool for creating AI videos. The script discusses how to use Stable Diffusion to generate consistent and flicker-free results, which is crucial for producing high-quality AI-generated videos.

💡Automatic 1111

Automatic 1111, also referred to as A1111 in the script, seems to be a software or tool used alongside Stable Diffusion for video generation. It is used for tasks such as background removal and adjusting settings to improve the quality of the generated videos, which is a significant part of the video's tutorial.

💡DaVinci Resolve

DaVinci Resolve is a professional video editing software that is used in the video for compositing and editing the AI-generated frames into a cohesive video. The script mentions using DaVinci Resolve to manage frame rates, apply effects like D Flicker for reducing inconsistencies, and to ultimately produce the final video.

💡Green Screen

A green screen is a technology used in video production where a green background is replaced with other images or videos during post-production. In the context of the video, it is suggested to choose videos with a green screen or simple background to facilitate the separation of the subject from the background, which is a common technique in AI video generation.

💡Frame Rate

The frame rate refers to the number of frames that appear on a video screen in one second. It is an important aspect of video quality and smoothness. The script discusses adjusting the frame rate in both the AI generation process and during editing in DaVinci Resolve to achieve desired video playback speeds.


Denoising is the process of reducing noise or unwanted visual artifacts in a video or image. In the video, denoising is a critical step when generating AI videos to reduce flickering and inconsistencies. The script mentions adjusting denoising levels and using tools like D Flicker to improve the final output.

💡Control Nets

Control Nets are tools or algorithms used in AI video generation to guide the process and achieve specific visual outcomes. The script discusses using different types of control nets, such as normal maps and open pose, to control aspects like facial features, body posture, and other details in the generated videos.

💡Batch Processing

Batch Processing refers to the execution of a program or set of programs in a predetermined sequence without manual intervention. In the context of the video, batch processing is used to automate the generation of multiple images or videos using AI, which saves time and ensures consistency across the generated content.


Inpainting is a technique used in image processing to fill in missing or damaged parts of an image. In the video, it is used to enhance the quality of generated images, particularly for details like faces. The script mentions using 'enable after detailer' with a strength of 0.3 for inpainting to improve the final video's appearance.


Keyframes are frames in an animation that define the starting and ending state of a transition. The video script discusses the importance of checking each keyframe for quality and consistency, as bad frames can be manually removed or replaced to ensure the final video's integrity.


This video serves as a beginner's guide to generating AI videos based on a reference video using Stable Diffusion and A1111.

Flickering or inconsistency in AI-generated videos can be reduced but not completely stopped.

Different settings in Stable Diffusion must be selected and adjusted based on the video and understanding of the process.

Training your own checkpoint can achieve the best results but it is a time-consuming process.

The simplest and most effective method demonstrated is image-to-image, avoiding complex extensions.

Free websites like Freebick offer green screen videos that can be used for easier subject-background separation.

A simple background and consistent clothing can make the animation smoother and reduce flickering.

DaVinci Resolve can be used to create new projects and adjust frame rates and resolutions for the AI video.

The use of a noise multiplier in A1111 can help reduce flickering in the generated images.

Experimentation with different control Nets is necessary to determine which settings yield the best results.

The removal of the background using A1111's background removal extension can improve accuracy and usefulness.

Adjusting the speed of the video frames in DaVinci Resolve can help synchronize frame rates and improve the video.

The D Flicker effect in DaVinci Resolve can significantly reduce flickering in the video.

Using multiple D Flicker effects can sometimes degrade the image quality and increase rendering time.

The use of After Detailer can enhance facial features and improve the consistency of generated images.

In-painting with a lower strength can create a style transfer effect without extreme resemblance to the original subject.

Batch processing can be used to generate a full set of images with consistent settings for efficiency.

Manually inspecting each frame for quality and removing bad frames is a crucial step in video generation.

Consistent characters can be achieved by using a special name and a complete prompt for batch processing.

Clear key effects in DaVinci Resolve can be used to remove backgrounds and prepare images for stylized videos.

Increasing the denoising level can help in generating videos with dramatic changes, such as turning a person into a robot.

Consistency in source and destination images is key to maintaining a coherent style throughout the video.

Image-to-image generation can also be used to produce deep fakes or face matches by merging faces with careful settings.