Tutorial Tool Others Stable Diffusion Setup Guide [Questianon]

applecider212

New Member
Jun 20, 2020
6
4
AgAABe92gVeaNEBcNfhMn5hQiXJXOzgN - Copy.png

Hi guys, it's been a while since my last post. I finally finished my Stable Diffusion guide for Windows + NVIDIA setups. I hope you find everything you need in there to set up and use Stable Diffusion to its fullest:

What is Stable Diffusion and why should you care (copied from my poll post)?
Stable Diffusion is a new piece of software that can allow many more people to easily make artwork their games and/or improve their artworks through AI. All you have to do is write a prompt of what you want and you'll have your desired image as your output.

You can generate things from scenery to character art, and in the right hands, there are limitless capabilities. Here's an example of someone who used Stable Diffusion to generate a webcomic:

As of right now, the process is confusing and it's hard for a newcomer to find the right resources. Some guides feel too convoluted and everything is everywhere. However, I've been compiling a compendium of resources that I think will let anyone set up and use Stable Diffusion. This is especially directed toward game developers that lack the technical expertise or funding to get good art for their games and people that just want good art (and smut) but can't find any online.

Current goals:
Linux + AMD setups
More training information
Guides on integrating SD into your workflow

Socials:
Trip: questianon !!YbTGdICxQOw
Discord: malt#6065
Reddit: u/questianon
Github:
Twitter:

Donation Links:
All donations will be put back into creating resources for the community. Do not donate if you cannot afford it
Patreon:
Ko-fi:

Note: If there's anyone wrong with this post, comment it or send a DM so I can edit it. Thanks
 
Last edited:
Sep 10, 2022
227
267
Sorry to ask, but on your tutorial you didnt stated how much vram is needed to use Stable Diffusion, from what I research 12 vram is the minimum correct?
 

x.utilitarian.x

New Member
Feb 13, 2018
2
1
Sorry to ask, but on your tutorial you didnt stated how much vram is needed to use Stable Diffusion, from what I research 12 vram is the minimum correct?
if you use the --medvram or --lowvram switches you can get away with less. I'm doing 4 minutes for a 20 pass 512x512 image on 3gb. If I try going much bigger at all I run out of memory and it has to bet he only thing running.

I use --medvram and use the core automatic1111 repo with some scripts added from other ones, and a myriad of porny checkpoints. it is worth doing if you have the time and inclination


edit: I want to mention, I think this could be a great tool for folks wanting to make games.

TRAINING a model, which is what you want to do for actual customization takes at least 6gb based on reports, but is unstable with less than 8. per the reddit communit.
 

Qwert99

Member
Mar 2, 2018
461
1,326
I hope the 5 minutes to install thing was just a joke. Been at this for 3 hours trying to troubleshoot shit and it still isn't working. Now the webui.bat can't seem to find the specific version of torch online when trying to install it and I've no idea what to get or where to put it or what to do with it so I'm completely stuck.
 

applecider212

New Member
Jun 20, 2020
6
4
I hope the 5 minutes to install thing was just a joke.
No, the 5 minutes thing is legitimate. The last two times I installed fresh copies of the webui took ~5-10 minutes since I already had the models installed on my computer. So far, my guide only covers Windows + NVIDIA, though I'll be adding Linux, Apple, and AMD eventually.

Now the webui.bat can't seem to find the specific version of torch online when trying to install it and I've no idea what to get or where to put it or what to do with it so I'm completely stuck.
For troubleshooting:
- Are you running webui-user.bat (not webui.bat)? It should install everything on the first run.
- The torch thing might be a bug, you might have to download the correct version of torch yourself and put it in your stable-diffusion-webui/venv/Lib/site-packages folder. Otherwise, webui-user.bat should install it for you

from what I research 12 vram is the minimum correct?
That was the case in the past. Now you can go as low as 2GB vram (I believe) or use CPU to run webui, though the generations might take minutes instead of seconds.

if you use the --medvram or --lowvram switches you can get away with less.
You can also use `--opt-split-attention` and/or `--always-batch-cond-uncond` (in order of which you should use first) for even less vram usage.
 
Sep 10, 2022
227
267
Hey, dunno if you can help but I have two questions, I have been using Invoke AI recently and tried to make some nsfw art with characters like Uraraka but only managed to make her naked (no actual sex) and I can't find a way on the internet on how to train my own model, basically I have my own folder with nsfw art and want to make a model based on that folder, any ideas?
 

Kryptozoid

Well-Known Member
Sep 3, 2016
1,210
2,541
Hey, dunno if you can help but I have two questions, I have been using Invoke AI recently and tried to make some nsfw art with characters like Uraraka but only managed to make her naked (no actual sex) and I can't find a way on the internet on how to train my own model, basically I have my own folder with nsfw art and want to make a model based on that folder, any ideas?
Can't explain anything related to InvokeAI cause I use Auto1111, I also can't explain how to train a full model (but I know it involves days/weeks of work and a Gigachad GPU). But if your main goal is just to make porn of a specific character, then just download a model specialised in nsfw stuff like the variations of NovelAI for anime stuff or URPM, or Hassan Blend for realistic style.
You can find plenty on Civitai.

Lastly if you want to use a specific person/character, you just need 50 to 120 pics of it, all cropped 512x512, and then find a tutorial on how to do embeddings. If you have a good GPU, once you have your dataset, it takes around 35mn to train, it results in small file around 10Ko that you put in the "embedding" folder (in Auto1111 at least, but must be similar in Invoke) and let's say you made a Zangief.pt file, when you'll type Zangief in your prompt it will know who is Zangief.

There's also the Dreambooth method, or the new Loras method which give better results, but are more complex to do. And honestly embeddings are pretty good once you know how to make a good dataset.
 
Sep 10, 2022
227
267
Can't explain anything related to InvokeAI cause I use Auto1111, I also can't explain how to train a full model (but I know it involves days/weeks of work and a Gigachad GPU). But if your main goal is just to make porn of a specific character, then just download a model specialised in nsfw stuff like the variations of NovelAI for anime stuff or URPM, or Hassan Blend for realistic style.
You can find plenty on Civitai.

Lastly if you want to use a specific person/character, you just need 50 to 120 pics of it, all cropped 512x512, and then find a tutorial on how to do embeddings. If you have a good GPU, once you have your dataset, it takes around 35mn to train, it results in small file around 10Ko that you put in the "embedding" folder (in Auto1111 at least, but must be similar in Invoke) and let's say you made a Zangief.pt file, when you'll type Zangief in your prompt it will know who is Zangief.

There's also the Dreambooth method, or the new Loras method which give better results, but are more complex to do. And honestly embeddings are pretty good once you know how to make a good dataset.
Yeah, tried the dreambooth but my gpu is not strong enough, Im gonna take a look at the embeddings method, thanks
 

Kryptozoid

Well-Known Member
Sep 3, 2016
1,210
2,541
Good luck, but probably abandon all hope if you only have 6GB VRAM (or maybe it works with a lot of configuration but it will be slow as hell and limited in various aspects).
 

Illhoon

Member
Jul 7, 2019
425
505
Hey guys i need some help i dont seem to get over this level of quality (really blatantly obvious bad AI art) what am i doing wrong can someone point me into a direction what i could be doing wrong ? and how to get better pictures like these

Here are some of mine ( btw im using Stable diffusion 1.5 pruned safentesors and abyssorangemix2_Hard.safetensors)
(this below was done in sd 1.5)
1729191710370.png
1729191826318.png
those are made in abyssorangemix2_Hard.safetensors
1729190505634.png 1729191667188.png
1729191991442.png
 

Illhoon

Member
Jul 7, 2019
425
505
ok Some are getting better now it seems the prompts aswell as the sampling steps helps ? i am wondering tho i wrote another person in the prompt aswell but the person is not appearing
1729193393229.png
 

Kryptozoid

Well-Known Member
Sep 3, 2016
1,210
2,541
ok Some are getting better now it seems the prompts aswell as the sampling steps helps ? i am wondering tho i wrote another person in the prompt aswell but the person is not appearing
View attachment 4142737
Hi, from this and your previous post, it seems you are new to this so I will give you some general advice : don't use base model, just go on Civitai and download one of the latest models with a style you like. Since you seem into anime, try a Pony 1.5 or SDXL model, there is some things to read about Pony models (some difference in the way you prompt, some negative and positive embeddings specific to Pony models who can help etc...), but they are pretty good for anime style. Some models need a VAE file, which is usually available too on Civitai, some have the VAE included in them.

If you prompt something like : "woman posing with man" and you don't see the male character, several solutions :
1/ change aspect ratio, sometimes if you expect to see 2 characters, but your picture is oreiented vertically, it will help to make it wider. So for example, you may not see 2 characters on a 512x960 picture, but you may see it on a 768x960 picture.

2/ change respective weight of characters. (btw this is true for 1.5 models, not so much for SDXL) :
if you write : "woman posing with (man:1.1)" you'll have more chance to see the man on your image. In theory you can go up to 2, but in practice you practically never go beyond 1.5, otherwise you'll get unusable results. You can weight every word or phrase in your prompt like that, then it's a matter of balancing it. Just take into account that if you don't weight anything, the first word in your prompt will be more important and weight will decrease until the last word in your prompt.

So for example,you write "big breasted woman posing with man in a supermarket, surrounded by ninjas doing acrobatic moves" then you check what your image looks like. You can find you don't see the ninjas enough and the woman takes too much space, and too much space is occupied by the supermarket floor. so you can try to change the order of the prompt, or more simply just this : "(big breasted woman:0.7) posing with man (in a supermarket:0.65), (surrounded by ninjas:1.125) doing acrobatic moves"
And then you generate again, see the changes, adapt your prompt etc.

Of course there is plenty of other things to know. but that's a start.
 
  • Like
Reactions: Illhoon

Illhoon

Member
Jul 7, 2019
425
505
Hi, from this and your previous post, it seems you are new to this so I will give you some general advice : don't use base model, just go on Civitai and download one of the latest models with a style you like. Since you seem into anime, try a Pony 1.5 or SDXL model, there is some things to read about Pony models (some difference in the way you prompt, some negative and positive embeddings specific to Pony models who can help etc...), but they are pretty good for anime style. Some models need a VAE file, which is usually available too on Civitai, some have the VAE included in them.

If you prompt something like : "woman posing with man" and you don't see the male character, several solutions :
1/ change aspect ratio, sometimes if you expect to see 2 characters, but your picture is oreiented vertically, it will help to make it wider. So for example, you may not see 2 characters on a 512x960 picture, but you may see it on a 768x960 picture.

2/ change respective weight of characters. (btw this is true for 1.5 models, not so much for SDXL) :
if you write : "woman posing with (man:1.1)" you'll have more chance to see the man on your image. In theory you can go up to 2, but in practice you practically never go beyond 1.5, otherwise you'll get unusable results. You can weight every word or phrase in your prompt like that, then it's a matter of balancing it. Just take into account that if you don't weight anything, the first word in your prompt will be more important and weight will decrease until the last word in your prompt.

So for example,you write "big breasted woman posing with man in a supermarket, surrounded by ninjas doing acrobatic moves" then you check what your image looks like. You can find you don't see the ninjas enough and the woman takes too much space, and too much space is occupied by the supermarket floor. so you can try to change the order of the prompt, or more simply just this : "(big breasted woman:0.7) posing with man (in a supermarket:0.65), (surrounded by ninjas:1.125) doing acrobatic moves"
And then you generate again, see the changes, adapt your prompt etc.

Of course there is plenty of other things to know. but that's a start.
thank you so much for the answer!

Prompting seems to work alot different than other AI applications do you know if there is something like a README or Biggoogle doc where you have a great selection of prompts ? (I used to have a website bookmarked where prompts where sorted for example for certain type of clothing or hairstyles but it seems its down now)like a selection of ressources i can use to learn more about what prompts to use what models etc ? or is there maybe a self hosteable llm that is trained to write Generativ ai Prompts ?

also i also realised the base models seem pretty bad haha what can they be used to then ? to creat loras ?
I also got myself SD1.5 based and Pony checkpoint they defenitly seem to work better then the 1.5 base model but im still SUUUUUUPER far off of usable ai art