[Stable Diffusion] Prompt Sharing and Learning Thread

Mr-Fox

Well-Known Member
Jan 24, 2020
1,401
3,802
From what I saw with my testing the benefit to upscaling twice is that the first time is a small increase and it increases the resolution while keeping the detail ( mostly the eyes ) looking the way they should without distortion.

After its upscaled slightly once, when you do the second upscale the detail ( such as round pupils ) is already there and it refines it further. When I tried just upscaling by x4 the eyes looked all distored in the pupils and required photoshop for me to fix.
Ok but did you use SD Upscale script or just the normal img2img upscale? It's a big difference
 
  • Like
Reactions: Synalon

Mr-Fox

Well-Known Member
Jan 24, 2020
1,401
3,802
As I said I just doubled the res size in img to img. I haven't even started learning to use the scripts yet :p
Well this clears things up a lot. Also I forgot to ask what denoising strength you used. This is also very important. If you get distorsion, something is obviously wrong.. The normal upscale in img2img and extra tab, doesn't create any new pixels to the image it only enlarge what is already there. It can't invent new pixels. hiresfix and SD Upscale is part of the generative process and creates new pixels. This actually increases the image quality, while normal upscaling only make the image larger without any real increase in quality. If you look carefully most of the time you actually lose details. The normal upscaling smoothes the details out and you actually lose sharpness. I guess you will just have to try it out for yourself to see what I'm talking about.
 
Last edited:

Synalon

Member
Jan 31, 2022
225
663
Well this says a lot. Also I forgot to ask what denoising strength you used. This is also very important. If you get distorsion, something is obviously wrong.. The normal upscale in img2img and extra tab, doesn't create any new pixels to the image it only enlarge what is already there. It can't invent new pixels. hiresfix and SD Upscale is part of the generative process and creates new pixels. This actually increases the image quality, while normal upscaling only make the image larger without any real increase in quality. If you look carefully most of the time you actually lose details. The normal upscaling smoothes the details out and you actually lose sharpness. I guess you will just have to try it out for yourself to see what I'm talking about.
0.25 Denoising strength in the picture I posted, I tested up until 0.50. I mentioned it there because I wasn't sure it was correct, so its best to give everybody all the information I have so they can experiment as well.
 
  • Like
Reactions: Mr-Fox

felldude

Active Member
Aug 26, 2017
572
1,694
Has anyone had success with IMG to 3D or text to 3D (ALT title AI is commin fer yer jerbs...soon)
Img-3d.jpg

Also is Comfy UI the best solution to use SDXL 1.0
 

humanno21

Member
Dec 4, 2019
283
750
Sorry for a newbie question but I want to check if this is a route I should take or even explore. What is the purpose of this? I see only portraits posted. Can you do full scenes with sex with this?
 

Dagg0th

Member
Jan 20, 2022
279
2,746

me3

Member
Dec 31, 2016
316
708
Nice to see highres fix still has it's extremely interesting memory usage...
1.9x uses much less than 2.0x and 1.8x uses more than 1.9...and 1.5x manages to OOM with the initial image creation for some reason...
Worryingly it seems that some of this poor memory management is spreading to the upscaling as well. At one point i had it OOM with trying to allocate an addition 6gb on a simple upscaling to just ˝1300x1300...

00000-3422812635.0.jpg
Still got a long way to go but it's getting there...

And the AI really doesn't like creating different "people", despite everything it keeps gradually moving things closer and closer until it at a bare minimum fits with the instructions...
00030-3359903690.jpg
 

me3

Member
Dec 31, 2016
316
708
Has anyone done a good Adam Warren Lora? I checked CIVITAI and I couldn't find one
If you mean the comic book writer/illustrator, it might already be there by default as a lot of "styles" are.

(rant not reflecting on quoted poster but meant as a general "comment")
That being said, imo, the more closely ppl try to "rip off" other ppls work or what they are "known" for, the more "misused" the AIs ability is, specially if they are trying to profit from it in any way as well. There's more than enough examples of ppl trying to claim "fan art" in very poor faith and "taste".
Same goes for the "misuse" of public personalities. If we think about it in a more old fashioned way, caricatures have been generally seen as a humorous and acceptable way to "misuse" likenesses so it's likely that the same kind of thing will be considered the same with AI generated scenes too.
It's pretty obvious that there's gonna start being lawsuit in the relatively near future over coping/stealing "art" or deformation for having used peoples likeness in situation they aren't known for. It's probably not gonna be a good idea to be on the receiving of those since there's limited existing laws.

So mix styles, create something new, that's kinda what creativity is about. If you put the likeness of a pornstar in those kinds of images you are far less likely to get in massive lawsuits than using someone not known for it, but it's so easy to create unique people by simply combining existing faces, nationalities, features etc that you shouldn't really have to.

Be creative, create, not steal ;)
 

Sharinel

Active Member
Dec 23, 2018
598
2,509
If you mean the comic book writer/illustrator, it might already be there by default as a lot of "styles" are.

(rant not reflecting on quoted poster but meant as a general "comment")
That being said, imo, the more closely ppl try to "rip off" other ppls work or what they are "known" for, the more "misused" the AIs ability is, specially if they are trying to profit from it in any way as well. There's more than enough examples of ppl trying to claim "fan art" in very poor faith and "taste".
Same goes for the "misuse" of public personalities. If we think about it in a more old fashioned way, caricatures have been generally seen as a humorous and acceptable way to "misuse" likenesses so it's likely that the same kind of thing will be considered the same with AI generated scenes too.
It's pretty obvious that there's gonna start being lawsuit in the relatively near future over coping/stealing "art" or deformation for having used peoples likeness in situation they aren't known for. It's probably not gonna be a good idea to be on the receiving of those since there's limited existing laws.

So mix styles, create something new, that's kinda what creativity is about. If you put the likeness of a pornstar in those kinds of images you are far less likely to get in massive lawsuits than using someone not known for it, but it's so easy to create unique people by simply combining existing faces, nationalities, features etc that you shouldn't really have to.

Be creative, create, not steal ;)
But if you do want to steal



:p
 

felldude

Active Member
Aug 26, 2017
572
1,694
HI, can I ask you if SDXL is better of SD? If answer is yes, I can use LORA with XL?
From a math technology standpoint it is superior. But it is 6GB and if you want to use the refiner that is another 5.9GB.

You can use lora's but if they where not trained on SDXL then they likely will not have a good effect.
If it is easy to train on it will probably surpass 1.5 (Unlike those 2.0 models)

Without out refiners in comfy UI I can do a native 1600x1200 image in 63 seconds (2.12 sec/it) with the SDXL model (With a 8GB RTX 3050)
ComfyUI_00108_.png
 
Last edited:

Rovente95

New Member
Jan 5, 2018
14
1
From a math technology standpoint it is superior. But it is 6GB and if you want to use the refiner that is another 5.9GB.

You can use lora's but if they where not trained on SDXL then they likely will not have a good effect.
If it is easy to train on it will probably surpass 1.5 (Unlike those 2.0 models)

Without out refiners in comfy UI I can do a native 1600x1200 image in 63 seconds (2.12 sec/it) with the SDXL model (With a 8GB RTX 3050)
View attachment 2812497
what do you mean by train them and how do you do it? Furthermore, ultimately it weighs more but on the other hand the results are better if your system has the requirements?
 

felldude

Active Member
Aug 26, 2017
572
1,694
what do you mean by train them and how do you do it? Furthermore, ultimately it weighs more but on the other hand the results are better if your system has the requirements?
Yes, for example their is a face set of data trained from Meta that is 100GB, (They tell everyone not to use it for facial recognition wink wink)
Most people can't do anything with it.

Training from scratch involves a 100k rig or more and usually takes days even with 80GB of video ram and 32 processors feeding that.

When us normal folk train a lora or checkpoint we are training with an existing set so in this case it would be SDXL, how it responds to training with 8bit optimizer or FP16, will determine the number of modified checkpoints or loras that support it
 

Rovente95

New Member
Jan 5, 2018
14
1
Yes, for example their is a face set of data trained from Meta that is 100GB, (They tell everyone not to use it for facial recognition wink wink)
Most people can't do anything with it.

Training from scratch involves a 100k rig or more and usually takes days even with 80GB of video ram and 32 processors feeding that.

When us normal folk train a lora or checkpoint we are training with an existing set so in this case it would be SDXL, how it responds to training with 8bit optimizer or FP16, will determine the number of modified checkpoints or loras that support it
forgive me I still don't understand what you mean by training... I only understood that Lora uses pre-existing models and XL uses Meta Human models
 

felldude

Active Member
Aug 26, 2017
572
1,694
So the realistic vision checkpoint was trained against 1.5 SD for the purpose of drawing realistic humans. If the 1.5 SD checkpoint was only trained on landscapes to begin with, I could feed it 1,000,000 images of humans and it would still draw landscapes with human features.

For XLSD they claim the math is better

But it doesn't mean an exisiting model won't fit your needs.

Compare:

SDXL

ComfyUI_00119_.png

in painting the SDXL image at .6 with realistic vision

ComfyUI_00126_.png

Realistic vison need to in paint the image as it would have made a composite of two images at that high of a native resolution. (SD 1.5 was intended for 512x512 rather then 1024x1024)

ComfyUI_00127_.png
 
  • Like
Reactions: Sepheyer and Mr-Fox