Prompts For Ultra Realistic AI Images: Stable Diffusion

All Your Tech AI
7 Mar 202311:39

TLDRThis tutorial demonstrates how to create ultra-realistic AI images using Stable Diffusion on a local PC. The key to success is crafting the right prompts and selecting the appropriate model trained on specific datasets. The video introduces for downloading various checkpoint models that enhance the aesthetic output. It guides viewers through the process of integrating these models into Invoke AI, generating images, and adjusting prompts for desired results. The host, Brian Lovett, also shows how minor prompt changes can significantly alter the image output, offering tips for refining AI-generated images for personal projects.


  • 🖼️ Generating photorealistic images with AI can be challenging, but it's possible with the right setup and prompts.
  • 💡 The quality of AI-generated images depends heavily on the choice of prompts and negative prompts, which guide the AI in what to include and exclude.
  • 🔍 There are free tools available for creating AI images, such as the Stable Diffusion setup on a Windows PC.
  • 🌐 Different versions of Stable Diffusion (e.g., 1.4, 1.5, 2.1) have been trained on different datasets, affecting the output aesthetics.
  • 📚 Additional image layers can be added to base datasets to influence the AI's output towards a specific aesthetic.
  • 🌐 Checkpoint models with various aesthetics can be found and downloaded for free from websites like
  • 🔧 Once downloaded, checkpoint models can be integrated into the AI setup through the model manager in the user interface.
  • 🔄 The process of generating images involves using prompts to create images and then making minor adjustments to the prompts for variations.
  • 🔑 Syntax for prompts may vary between different AI systems, and understanding these differences is crucial for achieving desired results.
  • 🎨 By altering keywords in the prompts, users can significantly change the style and content of the AI-generated images.
  • 📈 The resolution of generated images can be increased by using the 'send to image to image' feature, resulting in higher quality outputs.
  • 🌈 The techniques discussed are not limited to human images; they can also be applied to landscapes, cars, and other subjects.

Q & A

  • What are the two key tricks to achieving photorealistic images in Stable Diffusion?

    -The two key tricks are: 1) Using the right prompts, including both positive and negative prompts to guide the AI, and 2) Choosing the appropriate model that the AI was trained on, which can be further customized with additional images to achieve a specific aesthetic.

  • Where can users find different checkpoint models for Stable Diffusion?

    -Users can find different checkpoint models on the website, which offers various models trained on different image sets to achieve specific aesthetics.

  • How can users add a new checkpoint model to their Stable Diffusion setup?

    -Users can add a new checkpoint model by downloading the model file, going to the model manager in their AI setup, clicking 'add new', selecting 'add checkpoint safe tensor model', and providing the path to the downloaded file.

  • What should users do if the syntax of a prompt does not yield the expected results?

    -Users should check and adjust the syntax of the prompt according to the specific AI tool they are using (e.g., Invoke AI, Automatic 1111, Midjourney) as different tools recognize different delimiters and formats.

  • What effect does changing keywords in the prompt have on the generated images?

    -Changing keywords in the prompt can result in vastly different images, as it alters the parameters the AI uses to generate the images, thus modifying the aesthetic and details of the output.

  • How can users upscale an image generated by Stable Diffusion?

    -Users can upscale an image by using the 'send to image to image' option in the menu, selecting the upscaling factor (e.g., 4X), and invoking the upscale function to generate a higher resolution version of the image.

  • Can Stable Diffusion generate images of objects other than people, such as animals or cars?

    -Yes, Stable Diffusion can generate images of a wide range of subjects including animals, cars, landscapes, and more, depending on the prompts and models used.

  • How can users modify the aesthetic of an image generated by Stable Diffusion?

    -Users can modify the aesthetic by changing or removing specific keywords in the prompt, using trigger words from the model's documentation, and selecting different checkpoint models that have been trained on desired aesthetic styles.

  • What should users do if the initial prompt does not produce the desired result?

    -Users should experiment with removing or changing specific keywords in the prompt, checking for proper syntax according to the AI tool being used, and trying different models or additional training images to refine the results.

  • How can users share their prompt ideas and get more prompt suggestions?

    -Users can share their prompt ideas and get more suggestions by joining communities such as the Discord channel mentioned by the creator, where members share and discuss various prompt ideas and techniques.



🖼️ Generating Photorealistic AI Images

This paragraph introduces a tutorial on creating photorealistic images using stable diffusion software on a local PC. It emphasizes the importance of the right prompts and negative prompts to guide the AI in generating desired images. The video also mentions the significance of the model's training data and how layering additional images can enhance the output. The speaker introduces a resource,, where various checkpoint models with different aesthetics can be downloaded for free to improve the AI's image generation capabilities.


🔍 Customizing AI Image Generation

The second paragraph delves into the customization of AI-generated images through the use of specific prompts and the adaptation of existing prompts to fit the syntax of the invoke AI system. It discusses the process of selecting and downloading checkpoint files from, adding them to the model manager in invoke AI, and using them to generate images. The paragraph provides examples of photorealistic images produced with specific prompts, and it shows how altering keywords within those prompts can significantly change the resulting images, from the age of a person to the style of a car or cityscape.


🌐 Exploring AI Image Variations and Community Resources

The final paragraph discusses the exploration of AI image variations by manipulating prompts and trigger words to achieve different aesthetics. It explains how removing certain words can lead to more subdued or realistic images, as opposed to stylized or alien landscapes. The speaker encourages viewers to use online prompts as a starting point and refine them to match their desired aesthetic. The paragraph concludes with a call to action for viewers to like, subscribe, and join the speaker's Discord community for more prompt ideas and engagement.



💡Stable Diffusion

Stable Diffusion is a type of artificial intelligence model used for generating images from text prompts. In the video, it is the core technology that the host demonstrates to create photorealistic images. The script mentions different versions like Stable Diffusion 1.4, 1.5, and 2.1, each trained on different datasets, affecting the output's aesthetic.


Photorealism in the context of this video refers to the quality of AI-generated images appearing as if they were taken by a camera, with a high level of detail and realism. The host aims to achieve this by using specific prompts and models in Stable Diffusion to produce images that closely mimic real photographs.


Prompts are the textual descriptions or commands given to the AI to guide the generation of images. They are crucial in defining the content and style of the images produced. The script emphasizes the importance of both positive prompts, which specify what to include, and negative prompts, which specify what to exclude.

💡Negative Prompt

A negative prompt is a directive given to the AI to avoid including certain elements in the generated image. It serves as a constraint to guide the AI towards the desired outcome, as illustrated in the script where terms like 'deformed' and 'poorly drawn face' are used as negative prompts to ensure high-quality results.

💡Checkpoint Models

Checkpoint models in the video refer to the different versions of the Stable Diffusion model that have been trained on specific datasets. These models can be downloaded and used to influence the style and quality of the AI-generated images, as demonstrated by the host with the 'stably diffused wild mix' model.


Civitai is a website mentioned in the script where various checkpoint models with different aesthetics are available for free download. These models have been trained on distinct image sets, allowing users to select one that aligns with their desired visual outcome for image generation.

💡Invoke AI

Invoke AI is the software interface used in the video to interact with the Stable Diffusion model. It is through this interface that the host loads the checkpoint models, enters prompts, and generates images. The script describes how to use Invoke AI to set up and utilize the AI model.


Aesthetic in this video pertains to the visual style or appearance that the user wants to achieve in the AI-generated images. The host discusses how different checkpoint models and additional layered images can influence the aesthetic, such as 'bright colors' and 'high contrast' associated with RPG or cyberpunk styles.


Syntax in the context of the video refers to the specific format and structure of the prompts that the AI understands. The script mentions that syntax can vary between different AI systems, and it is important to adjust the prompts accordingly to achieve the desired results.


Upscaling in the video is the process of increasing the resolution of an AI-generated image to create a higher quality version. The host demonstrates using the 'send to image to image' feature in Invoke AI to upscale an image by 4X, maintaining the same visual content but with greater detail.

💡Trigger Words

Trigger words are specific terms used in prompts that evoke particular styles or elements in the AI-generated images. The script explains how using trigger words like 'SW Punk', 'synthwave', and 'paint splatters' can create a stylized, neon cyberpunk aesthetic in the images.


Demonstrates how to generate photorealistic images using Stable Diffusion on a local PC.

Discusses the importance of prompts and negative prompts in AI image generation.

Introduces the concept of using different versions of Stable Diffusion trained on various datasets.

Explains how to layer additional images on top of base datasets to influence the model's output.

Recommends as a source for free checkpoint models with different aesthetics.

Guides on downloading and integrating checkpoint models into Invoke AI.

Shows how to select and load a specific checkpoint model for image generation.

Presents examples of photorealistic images generated with specific prompts.

Illustrates the use of positive and negative prompts to refine image generation.

Details the process of adjusting prompts to achieve different image outcomes.

Explains the impact of prompt syntax on the AI's interpretation and image results.

Demonstrates how minor changes in prompts can lead to significantly different images.

Shows how to upscale images to a higher resolution for improved quality.

Explores the use of trigger words in prompts to achieve specific aesthetics.

Discusses the flexibility of AI image generation across various subjects like cars, landscapes, and animals.

Provides tips for refining prompts found online to suit personal project needs.

Encourages viewers to subscribe for more content and join the community for shared ideas.