StableDiff
StableDiff
1. Upload Reference Image: You've already done this. This is the image that will guide the style,
composition, and/or content of your generated image.
3. Enter a Prompt: This is crucial. Your prompt describes the image you want to generate. Be as
specific as possible. Consider including:
o Style: What artistic style do you want (e.g., photorealistic, anime, painting, digital
art)?
o Atmosphere/Mood: What feeling should the image evoke (e.g., serene, dramatic,
vibrant)?
o Details: Specific elements you want to include (e.g., colors, textures, lighting).
4. Adjust Parameters: These settings control how the AI interprets your reference image and
prompt. Key parameters include:
Low values (e.g., 0.2-0.4): The generated image will closely resemble the
structure and composition of the reference image, with subtle changes in
style and details influenced by the prompt.
High values (e.g., 0.7-0.9): The generated image will be significantly different
from the reference image, mainly using it as a loose guide for composition or
subject matter. Be cautious with very high values as they can lead to
incoherent results.
o Sampling Steps: The number of iterations the sampler performs. Higher steps
generally lead to more refined images but take longer to generate. 20-50 steps are
often a good starting point.
o CFG Scale (Classifier-Free Guidance Scale): Controls how closely the generated image
adheres to your prompt. Higher values (e.g., 7-12) generally lead to more prompt-
accurate results but can sometimes introduce artifacts or over-saturation. Lower
values might give more artistic freedom but could deviate more from the prompt.
o Seed: A random number that determines the initial noise. Using the same seed with
the same parameters will produce the same result. Use -1 for a random seed.
o Batch Count and Batch Size: If you want to generate multiple images at once.
5. Generate: Click the "Generate" button to start the image generation process.
Start with a clear and concise prompt. Even with a reference image, the prompt is essential
for guiding the AI.
Experiment with the Denoising Strength. This is the key parameter for controlling how much
the output resembles the input. Start with a moderate value (around 0.5) and adjust based
on your desired level of change.
Consider the resolution of your reference image. While Stable Diffusion can work with
various resolutions, very low-resolution images might lead to blurry or less detailed outputs.
Pay attention to aspect ratios. If your reference image has a different aspect ratio than your
desired output, you might get stretching or cropping. You can use inpainting or pre-
processing to address this.
Use negative prompts. Specify things you don't want in your generated image (e.g., "blurry,
deformed, bad anatomy, unrealistic").
Experiment with different samplers and CFG scales. These can significantly impact the style
and quality of the generated image.
Use image editing tools (like inpainting) for further refinement. If certain parts of the
generated image aren't quite right, you can use inpainting to regenerate specific areas while
keeping the rest.
Check your VRAM usage. Generating high-resolution images with many steps can consume a
lot of VRAM. If you encounter issues, try reducing the resolution, steps, or batch size.
Example Scenario:
Let's say your reference image is a photo of a cat sitting on a windowsill. You want to transform it into
a watercolor painting.
Prompt: "A cute cat sitting on a windowsill, watercolor painting, soft brushstrokes, vibrant
colors"
Steps: 30-40.