Prompts Guide by CDcruz : r/WaifuDiffusion Skip to main content

Get the Reddit app

Scan this QR code to download the app now
Or check it out in the app stores
r/WaifuDiffusion icon
r/WaifuDiffusion icon
Go to WaifuDiffusion
r/WaifuDiffusion

A haven to find comfort and solace in posting your AI generated Anime Waifus!


Members Online

Prompts Guide by CDcruz

Guide

EDIT: For more up to date and comprehensive information, visit my newly made webpage: https://stablediffusion.cdcruz.com/ There are no ads, just information.

Hello, I've been playing around with Stable Diffusion and Waifu Diffusion almost non-stop for the last few days. It is super fun and facinating to see what this ai can create and what its limits currently are.

With all of my generating, I've learnt a lot about how the prompts work and what works best for desired outcomes. I've also found tips from other people online. I'd like to make this post a summary of everything I've learnt so far to help others create cool art.

For reference, I am using the WebGui that was downloaded following this guide https://rentry.org/voldy

This GUI includes many useful features including batch processing, face restoration, upscaling, inpating, outpainting and more.

This github link provides more detailed information on special features of the webgui https://github.com/AUTOMATIC1111/stable-diffusion-webui-feature-showcase

Sampling Methods

There are quite a few different models to choose from. However, from my testing there isn’t a huge difference in what they produce, except the a versions which look similar to each other.

For me, the Euler A sampling method is the best for both anime and realistic renders. The only down side is that the output changes much more than the other methods depending on sampling steps.

The Euler (not A) method does stay consistant no matter how many steps you render for, but in my limited testing, I feel it creates a image that isn't as refined as Euler A. I'm not sure why this is, or I haven't rendered enough Euler images, but that's how I feel from testing so far. And cranking the steps up on Euler does not improve the image much after about 200 steps. I have done a quick test of the same seed from 10 - 5000 samples and the results were consistant, but not improved by the higher sample steps.

If anyone has better results with a method other than Euler A please let me know. For now, my advice is to stick with Euler A and all of my current advice is mostly based on that method.

Prompt Order

I found this useful video https://www.youtube.com/watch?v=c5dHIz0RyMU about how different things can affect the ai output. One notable affect is the position of prompts. Prompts that are at the front of the list can sway the ai more than ones further down the list.

However there are ways to manipulate the strength of prompts even more.

Using (curly brackets) around words can emphasis it more, more brackets the stronger it is. Opposingly, [square brackets] will make the prompt less impactful.

Another way to do this is by doing prompt:1. I am unsure if the number refers to position or strength, but from testing I believe it means the position of the prompt regardless of actual position in the list. So if I did gold, shoes:1 shoes would be more influential than gold even though it is second in the list.

Styles

There are many ways to create different artistic styles using prompts. Most common is things like Artstation, Dali, photo, painting, etc. As this is a Waifu sub, I assume you are wanting an anime style, It will require more testing, but currently the best working prompt for an anime style I have found is "by Toei Animation", I'm sure other animation companies will work too, but this one has worked best for my needs.

Here's a more detailed list of art/photo styles that I've found to work well:

These were all tested with Waifu Diffusion, Euler A, with each prompt at the beginning of the prompt list, so results will vary a lot if you use Stable Diffusion and different settings.

  • Monochome or black and white - massive influence - will make everything black and white

  • Sepia - mild influence - will give a sepia colour palette.

  • Color/ colour - weak influence - Not really required for more images as they are usually in colour.

  • Colourful/ colorful - mild influence - Will add more variety of colour in the image.

  • Traditional Art - mild influence - a general art prompt with varying styles and detail.

  • Painting - strong influence - will add paintery brush strokes to image.

  • Ink art - mild influence - Basically makes black and white drawings.

  • Chalk art - mild influence - Sometimes adds a chalk effect to image, sometimes adds random blackboard scribbles to background.

  • Oil painting - weak influence - seems to be less powerful than simply using "painting".

  • Watercolour/watercolor - strong influence - adds a similar effect as "painting" but does distinctly look more watercolour-like.

  • Drawing - mild influence - Will add a more pencil-like style to the image. But seems more digital pencil than real pencil.

  • Photo/photograph - strong influence - will make the image much more realistic, but in Waifu Diffusion it will maintain anime characteristics and skin will be smooth.

  • Canon 50D (or other camera types) - mild/ strong influence - will make the image more realistic.

  • Cinematic lighting - mild/strong influence - will add more dramatic lighting and usually a point light in the scene. Also darker scenes and backlit scenes.

  • God Rays - mild/strong influence - will add a strong sun light that will usually shine on one side of a persons head/ hair. Can also add rays of light but less likely.

  • Cell Shading - strong influence - a detailed anime style, most similar to visual novel characters.

  • Anime - strong influence - Anime

  • Waifu - Despite the name of the ai, doesn't seem to change much style wise.

  • Key Visual - strong influence - A detailed anime style while keeping the anime characteristics or big eyes, etc.

  • Hentai - mild/strong influence - This prompt varies between realistic and anime style depending on other prompts and will also obviously make the image more lewd.

  • Artstation - strong influence - A variety of artstyles, usually more professional looking.

  • By Toei Animation - strong influence - A better anime prompt for more detailed anime.

  • By Studio Ghibli - strong influence - A better anime prompt for simple anime characters and classic Ghibli backgrounds.

  • In the style of Mappa - mild influence - Another anime style.

  • By Kyoto Animation - mild influence - Anime

  • In the style of Cloverworks - mild influence - Anime

  • Model - mild influence - Helps a person look more "perfect", they will also usually look directly at the camera.

  • Instagram - mild influence - Characters have a much wider range of poses and camera angles

  • vtuber - strong influence - will generate anime characters mainly in the style of vtubers.

  • Snapchat - weak influence - Similar to "Instagram" but also produces more selfie-like images.

  • Professional - weak influence - looks more professional shot and has more background variety butthe background is usually out of focus.

  • Ponytail/pigtails/braids/(Various hairstyles) - strong influence - will change hairstyle a lot.

Negative Prompts

Negative prompts are extremely useful. Whatever negative prompt you enter, will be excluded, or reduced in the generating of an image. For example, if I had the prompt "Vehicle" it could show cars, bikes, buses, etc. If I were to add "Buses" to the Negative prompts, the ai will be much less likely to generate a bus image and would instead only generate bikes or cars.

This can be extremely useful for fine-tuning a image, removing disfigured body parts, or reducing the strength of prompts. The prompt order system works for negative prompts as well, but instead, the stronger the prompt, the less likely it will appear and vice versa.

My current default negative prompts list is below:

deformed, blurry, bad anatomy, disfigured, poorly drawn face, mutation, mutated, extra_limb, ugly, poorly drawn hands, two heads,child, kid, gross, mutilated, disgusting, horrible, scary, evil, old, conjoined

A more controlled image

As stated previously, Euler A can be very unpredictable in its output. On the otherhand, Euler (not A) is incredibly consistant. If you are looking to really finetune an image, Euler (not A) is what you need. I recommend putting in all of your basic settings that you want until you start to see images that represent what you're looking for. Once you've found a good image, lock down the seed and start changing prompts & even try changing the CFG scale to see how it changes the image. Unlike Euler A, it will keep fairly consistent even with different prompts added/ removed. On top of this, you can use the "Extra" section next to "Seed" that adds see variation. This will mix in a different seed to your main seed to create subtle differences while keeping your main seeds image characteristics intact.

Following these steps, you'll be able to create the exact image you're looking for, and if it's still got some problems, move over the inpainting to try and fix those. Or even photoshop could work for finishing touches.

Fixing Issue Areas

There are some body parts that the ai is completely horrible at drawing. Most notably hands, eyes, ears, arms and giantic boobs.

A few ways to combat this is to firstly use negative prompts to remove disfigured body parts. But you can also prefix "perfect" infront of prompts to help the ai.

For eyes, "deep eyes" seems to help it create less errors, although I'm not entirely sure why.

Hands seem to be extremely hard to get right and I am yet to figure out how to render them properly. If anyone has tips for hands, It would be really helpful.

Gigantic boobs seem to appear even when you aren't specifically stating them in your prompt. Even writing "small boobs" or similar may sometimes still create massive honkers. One way to combat this, is to add boobs to your negative prompts, but this can have unintended consequences so I will usually compensate and add similar prompts to the normal prompts that sounds like boobs but isn't, like "curvy".

Sampling Steps

Your natural instinct when generating images might be to max out the sampling steps for the highest quality image possible. However theres a few reasons this might not be the best thing to do.

  1. Takes a long time to render.

  2. The end result art may be very different compared to if you used a lower sampling step, and the lower step result may be a better result.

  3. Especially for anime style art, real life anime art is not very detailed, so making the ai run for a low amount of steps will help immitate that simplistic artstyle better than over-doing it.

  4. If you're trying to render longer to remove glitches an errors, (from my testing) I don't believe it will fix much or may even change the image drastically as point 2 mentions. You are better off trying inpainting to fix mistakes.

I have tried extremely high sampling steps (5000) and while the results can be impressive, you can acheive roughly the same results in the 100-500 steps range, and for anime using the Euler A model, you can even get decent results from 20-100 steps.

I hope these tips are helpful, I will continue to update this post as I learn new information. Also please leave comments with your tips and tricks for greating cool art.

r/WaifuDiffusion - Prompts Guide by CDcruz

Share
Sort by:
Best
Open comment sort options
u/yayuuuhhh avatar

Wow, this truly helped! Thank you so much OP!

u/masaishi2001 avatar

Thank you for a really helpful post.

I tried some strong influence styles by same seed to check what type of effects these prompts make.

The below image shows results. The right top one is original picture, and I add one style to each 15 pictures.

https://ibb.co/GJHCK0H

Thanks for doing a but of extra research on the prompts. I’m currently working on creating a webpage where I can share all the information I’ve learnt and continually update it. It’ll have a larger list of prompts.

More replies
u/Jiten avatar

The reason Euler a (as well any other sampler with a in the name) gives different results from others, as well as for every number of steps, is that it adds more random noise to the image every step it does.

You can see this happening if you turn on displaying the in-progress image. With the ancestral samplers, the image will be quite unstable for quite a long time. While with non-ancestral samplers it tends to just slowly condense into the result without changing over time.

[deleted]
[deleted]

Why my images are looking so bad despite giving the same prompts.

any important settings I'm missing?

Also where do I change the CLIP thing.

You’ll need to give me more info on what image you’re creating. Like what model, prompt, settings, etc.

The clip skip setting is in the settings tab

More replies

We run an AI art platform for anime fans where you get 100 free credits to make lewds from 3 different models. Would love to invite you and ask for feedback for Yodayo.com   

u/coilovercat avatar

I've done som experimenting with step count and I've found that DPM++ SDE karras produces the best results (on my end at least)

with this sampler I've noticed it takes steps differently, so turning them up to 5k won't give insane results and might actually make it worse. It basically knows how many steps you've given it and then has each step do smaller and smaller changes based on what step cound you have.

At the roughly 30-40 step count area, differences become negligible, and I've found a lot of perfection with 35 steps specifically.

You actually don't have to have that many steps in the first place. I get coherent images from 5 steps using DPM++ SDE Karras, and a step count as low as 4 gives you a serviceable face.

How do I add more then 150 steps. I can not exceed this block.

u/coilovercat avatar

set the x-y plot thingo to more than 150

More replies
More replies
u/Adunaiii avatar

I have no idea what I'm looking at (I have ~60 IQ). But it seems to be about installing it on one's own machine (thus requiring >8 GB RAM)? If so, I'll leave these two links for my fellow 60-IQ folks.

  1. Google Colab;

  2. BATbot.

Considering how incomprehensible this whole ordeal seems to me, I'm quite close to paying for NovelAI. Although considering how difficult that website is, too, and how tech-savvy their users are as well, I probably will leave this to stew for another year until it becomes sufficiently user-friendly.

someone can help me how to use lora on google colab a1111?

How do I get a consistent shading/style?

Hi, it can take a lot of trial an error and depends on what style you're going for. It's best to find an artist/ animation studio that draws the style you're after and add them to the prompt. For anime, the prompts "anime" and "kawaii" work well to get started. Shading is a bit harder if you're going for something unique, but again if you know an artist that shades like you want, add them to the prompt. Alternatively, you can try training a textual inversion, hypernetwork or dreambooth for the style you want, but that is a lot of time and effort.

More replies
u/Profanion avatar

Question: Why adding negative prompts sometimes makes it so that it would act like I've typed positive prompt? Especially with colors.

I’m not sure what you mean sorry. I don’t think I have come across that problem. Do you have an example? Feel free to DM me.

u/Profanion avatar

Let's say I type "blue shirt" and "black shirt" as a negative prompt. This causes characters to wear blue shirts and black shirts.

I'm trying to replicate what you're saying, but I do not seem to have the problem you're having. All I can suggest is that you could put the colour you do want in the positive prompts, or increase the influence of the negative prompts with brackets or moving the most important negative prompts to the front of the list.

Same thing happens to me too, lol. I just am like, why. Dont have a solution, I just keep trying in different ways.

More replies
More replies
More replies

So say the Danbooru tag you want already has a curly bracket in it (like, for example, character names that are attributed to certain artists), how does that affect us surrounding it with even more curly brackets? Can it tell that it's a tag? Should we use something else to denote priority? I really can't tell whether it's working or not (probably because I keep being suggested notebooks and don't know what model I'm using).

If you need to use brackets in your actual prompt, you need to escape those characters with a backslash. Like \(artist\) and then you can use normal brackets around that and it should increase influence as normal.

You should use the newer way of increasing influence which is (prompt:1.3) with the number being the influence. for your artist example, it would be (manga\(artist\):1.4) which should work correctly.

Gonna have to get down with the placement of those backslashes, but I had no clue that was even a thing. So is it basically backslashes before each internal bracket?

The format I was working with was | game_name_(character_name) |, so would that be |game_name_\(character_name\) \? I am doing trial-and-error with the commas and underscores, so that format isn't set-in-stone, but is the placement at least accurate?

Thats correct

More replies
u/JoshxDarnxIt avatar

I don't really understand the new influence system. I'm assuming that larger numbers result in greater influence, but why are they decimals, and why does everyone seem to just use the numbers 1.2 to 1.4 for all of their tags?

Its a multiplier. So if i wrote (blue:1.2) blue will have 1.2 times more influence. If i wrote 2.0 it would have double influence. Since it is a multiplier you don’t need to increase the value too much to see a difference.

u/JoshxDarnxIt avatar

AHH, that makes sense. Thank you!

More replies
More replies
More replies
u/geekgame876 avatar

I agree with the subject, I have tried to have a character with the style of a specific artist (sakimichan, nyatcha, cslucaris) but it does not work, also if the character example: kali belladonna that does not have many references does not appear correctly and of course have more than one character, misty, dawn. we have a lot to learn

More replies
Edited

Thanks, very useful information! But I can't solve the problem with the inpaint function in any way, it just doesn't work. I watched a lot of guides and tried various settings, but the mask just blurs on the final image and.. that's it. I can't change images. I tried different models to work with anime art, but the mask just blurs and does not change parts of the image in any way. It's like I'm getting censored instead of the desired result. Has anyone encountered this problem?

UPD: --no-half --precision full --no-half-vae --opt-sub-quad-attention --opt-split-attention-v1

This line must be included in set COMMANDLINE_ARGS= in the webui-user.bat file

That is, in the end, the line should look like this:

set COMMANDLINE_ARGS=--no-half --precision full --no-half-vae --opt-sub-quad-attention --opt-split-attention-v1

After the done manipulations, inpaint works correctly on AMD GPU

Without being able to see what you're doing, the only problem I can think of is that you haven't set the Masked Content option to use "original" if you use "fill" or other options, it will generate noise. Setting Masked Content to "original" will keep the original and just add things to it.
You also need to set the denoising strength low, otherwise, it will just change the image completely.

Doesn't work, unfortunately. There may be a problem with my particular AUTOMATIC1111 installation. I used the instructions from the wiki for AMD graphics cards as I am using the RX 6600. May be compatibility issues

Ah yeah, Im not too sure of AMD cards. I use Nvidia and I know it works best with them.

More replies
More replies
More replies
u/AkkoLotteSucy avatar

i have a question about someone whose just started using stuff like yodayo recently; something i've been trying to generate stuff of is "costume swaps"; like an anime girl from one show dressed up as one from another show.

by any chance do you know the best way to make the AI understand what I am trying to say? it seems to struggle a bit with "- dressed as -" or "- cosplaying -", so i was wondering if anyone knew a good way to make the AI understand this

That might be a bit difficult to achieve with prompts alone unless they are both really popular characters, the AI has no context about what it is actually drawing. You may need to use a Control Net, textual embeddings or even train your own Dreambooth model on one or both of characters.
The only other option would be to try inpainting, but that would probably take a long time to get a good result.

More replies
u/Tynixo avatar

Most upvoted post that isnt a nude anime girl

u/GamerMan762 avatar

So i downloaded Stable Diffusion using Rentry.org i followed its guide and i launched the webui-user.bat and it gave me a local url which i typed into my browser and my browser just says that the url didnt send any data? How do i fix this?

(this is my first time using anything like this)

I’m not too sure sorry. It sounds like everything went right. All i can suggest is to make sure you typed in the url correctly. It should just be an ip address.

u/GamerMan762 avatar

i figured it out apparently i didn’t set the python location in the webui-user.bat i had to watch a youtube video

More replies
More replies
u/Fanficwriter21 avatar

Good god I was looking for a little hep…..little, but hot damn I’ll take this anyday. Thanks a lot.

u/Low_Engineering_5628 avatar

I really like the control you get from ControlNet (no pun intended).

ControlNet1.1 has introduced some new models lineart models such as lineart_anime. You can draw out a scene (very crudely it seems) or even take lineart and move around bits in something like Photopea. this way you can get very accurate poses without it affecting the style too much.

See exhibits A and B:

A: https://www.deviantart.com/fulforget85/art/Are-you-off-your-diet-again-958786943

B: https://www.deviantart.com/fulforget85/art/Are-you-off-your-diet-again-Earth-616-958951419

I've often found that switching from an Anime based model to a realistic on tends to change the output a lot - with these new ControlNet models it really helps making general purpose checkpoints fit a specific idea.

u/tayzzerlordling avatar

anyone have tips on how to get dicks to stop sprouting out of random places? thanks

[deleted]
[deleted]

Other prompts that seems to have a strong effect:

Sketch

Character Design Sheet

[deleted]
[deleted]

do I need a custom lora if I wanna do stuff like fangs?

u/urkikichan avatar

take my thanks

u/Imaginary_Cod2326 avatar

Damn thats a lot of knowledge… thx bro it will be put to a good use 😘

Awesome

u/JustALittleSexy avatar

Which method is recommended to try out, AUTOMATIC1111 webGUI sounds like the "best" option but looks a little daunting to get started. The NMKD GUI and Artroom.ai seem simpler, just maybe not a good. Is there maybe a video that explains how to install any of them?

u/JustALittleSexy avatar

I downloaded the NMKD GUI and need a little help getting it right. I tried a few prompts, just copying some for the ones posted here, and they look super bad, which I kind of expected at first. I think I need to use models, but not sure how to download them. Do I just put them in the models folder of SDGUI or do they go in a specific one; Checkpoints, Embeddings, LoRAs, or VAEs. After that, is there anything else I need to do to make images as good as the ones here?

Cake icon Edited

Hi, I need help with the installation. When I try to run webui-user.bat, it gives me an error. The error is called Error 1 and then it says "press any key to exit". Can someone help me in figuring this out. I followed the instructions as per the guide but I can't seem to figure this out. I installed Python correctly too.

https://imgur.com/a/dtgIcEx