Yesterday, I spent hours experimenting with Animagine XL V3.1 and realized most tutorials online only “give you fish rather than teach you how to fish.” Understandably, why share the secrets when you can make money from them? On Xianyu (a Chinese second-hand platform), some people even charge 200 yuan just to deploy Deepseek for others. Still, I want to write this article—it may not be comprehensive, but I’ll try to show you how to find resources or craft your own prompts instead of just copying and pasting.
First, here’s what the Stable Diffusion WebUI interface looks like:
Below, I’ll only introduce key or commonly used parameters (with brief explanations)—detailed demos follow later:
stable-diffusion-webui/models/Stable-diffusion.-1 = random seed.We’ll start with demos of non-prompt parameters (Prompt and Negative Prompt are more complex and covered later).
1girl, looking at viewer, wool coat, fur collar, smile, standing on a hillside, snowfall, mountains in the background, masterpiece, best quality, very aesthetic, 1990s style, retro artstyle,
lowres, bad anatomy, displeasing, ugly, fewer digit, extra digit, missing fingers, bad hands, blurry, (low quality, worst quality:1.3)
As mentioned, dimensions directly impact detail quality. Compare 512x512 vs. 1024x1024 (Seed: 3503979678, Sampling steps: 20, CFG Scale: 7):
The detail gap is striking.
Dimensions also affect composition. Compare 512x1024, 1024x1024, and 1024x2048:
Since SD is trained on 1024x1024 images, results are best when (Width + Height) is a multiple of 1024. Recommended aspect ratios:
Sampling steps enhance detail. 30–40 steps are sufficient for most cases:
Demo with Sampling steps increased to 35 (same base parameters):
Notice more facial/background details and richer color gradients—zoom in on the hair for a clearer comparison.
Important Note: Even with a fixed seed, Batch count and Batch size will increment the seed. For example, the 4 images below use seeds 3503979678~3503979681.
Generated images:
Images generate simultaneously:
Generated images:
The seed acts as the “foundation” for image generation. As shown in the batch demos, seed variations drastically change the output.
stable-diffusion-webui/output—easy to retrieve later:
Images with larger seed gaps show more significant differences:
These are the most critical components. Let’s break down the base prompts and explore advanced usage.
1girl, looking at viewer, wool coat, fur collar, smile, standing on a hillside, snowfall, mountains in the background, masterpiece, best quality, very aesthetic, 1990s style, retro artstyle,
Covers: Character (1girl), pose (looking at viewer), clothing (wool coat, fur collar), expression (smile), setting (standing on a hillside, snowfall, mountains), quality (masterpiece, best quality), and style (1990s style, retro artstyle).
This is my go-to Negative Prompt for most scenarios:
lowres, bad anatomy, displeasing, ugly, fewer digit, extra digit, missing fingers, bad hands, blurry, (low quality, worst quality:1.3)
Targets: Low resolution (lowres), anatomical errors (bad anatomy), unpleasant/ugly details, digit abnormalities (fewer/extra/missing fingers), bad hand rendering (bad hands), blurriness, and low quality (amplified by 1.3x weight).
Animagine XL V3.1 supports style specification in two ways:
1990s style, retro artstyle).(Artist Name:1.3) to the prompt (1.3x weight amplifies the style).A comprehensive list of compatible artists is available here: Animagine XL v3.1 - Artists’ Style Sheet. Example snippet:
1girl, looking at viewer, wool coat, fur collar, smile, standing on a hillside, snowfall, mountains in the background, masterpiece, best quality, very aesthetic, (tinnies:1.3)
Generated image:
1girl, looking at viewer, wool coat, fur collar, smile, standing on a hillside, snowfall, mountains in the background, masterpiece, best quality, very aesthetic, (sekina:1.3)
Generated image:
Notice how prompts use short phrases (not full sentences)? Danbooru tags is a community-driven database of tags covering styles, poses, expressions, objects, and more (note: contains NSFW content).
For example, the tag flustered (described as “embarrassed/nervous”) includes visual examples:
Let’s demo with expressions and backgrounds:
Change smile to flustered (I think this translates better to “shy” than “flustered” based on examples):
The expression clearly shifts to shy/embarrassed.
Replace standing on a hillside, mountains in the background with city street in the background:
1girl, looking at viewer, wool coat, fur collar, smile, snowfall, city street in the background, masterpiece, best quality, very aesthetic, 1990s style, retro artstyle,
Generated image (correct background):
Generated image (forgot to remove standing on a hillside—inconsistent composition):
Explore more prompts by browsing Danbooru tags and experimenting on your own!
I hope these will help someone in need~