[Stable Diffusion] Prompt Sharing and Learning Thread

me3

Member
Dec 31, 2016
316
708
Indeed, I have the same experience when attempting i2i frames using controlnets, the end results are nothing but acid trips.
I put them in spoiler just incase it might be annoying/"trippy" to look at for someone, so view at your own risk.
You don't have permission to view the spoiler content. Log in or register now.
 
  • Haha
Reactions: sharlotte

me3

Member
Dec 31, 2016
316
708
So i did some tests on seeing how it would work with weighting prompts to deal with the controlnet restrictions, basically trying to brute force it do allow for "change". I had to get right the limit of causing very bad distortions for some of it to even work, this is slightly lowered weights.
In the first clip you can see how the "long dress" is glitching in to some frames, it seems fairly stable as a dress but that's it.
Second clip i tried to adjust the body type/size...and it has some issues...
So purely prompting out of this seems unlike, the hunt continues...
View attachment w1.mp4
View attachment w2.mp4
 
  • Like
Reactions: Mr-Fox

Sepheyer

Well-Known Member
Dec 21, 2020
1,570
3,766
So i did some tests on seeing how it would work with weighting prompts to deal with the controlnet restrictions, basically trying to brute force it do allow for "change". I had to get right the limit of causing very bad distortions for some of it to even work, this is slightly lowered weights.
In the first clip you can see how the "long dress" is glitching in to some frames, it seems fairly stable as a dress but that's it.
Second clip i tried to adjust the body type/size...and it has some issues...
So purely prompting out of this seems unlike, the hunt continues...
View attachment 3163295
View attachment 3163294
In this case the IPAdapter is your tool. You show it what dress you want, then inject it's output into the sampler. It is a "more better" approach then using prompt weight's. In fact it becomes a micromodel that must improve consistency.

If you need a snippet of the workflow let me know.
 

hkennereth

Member
Mar 3, 2019
237
775
So i did some tests on seeing how it would work with weighting prompts to deal with the controlnet restrictions, basically trying to brute force it do allow for "change". I had to get right the limit of causing very bad distortions for some of it to even work, this is slightly lowered weights.
In the first clip you can see how the "long dress" is glitching in to some frames, it seems fairly stable as a dress but that's it.
Second clip i tried to adjust the body type/size...and it has some issues...
So purely prompting out of this seems unlike, the hunt continues...
You should check out this video here for an alternative method for video generation that is perhaps better at temporal coherence:
 

Mr-Fox

Well-Known Member
Jan 24, 2020
1,401
3,802
You should check out this video here for an alternative method for video generation that is perhaps better at temporal coherence:
This is Ai enhanced animation rather than Ai created animation. I think that most would prefer to not create an animation in a 3d editor such as blender but rather rely mostly on SD instead. SD is getting better and better so with time it will probably happen that it's actually possible and more straight forward than it currently is. As far as I have seen, animatediff is the only tool or method that is only using Ai to create the animation. All other options are some form of Ai enhancement or hybrid videos. Live caption to SD video is also something that is getting more and more possible with such extensions as LCM Live.
Though with all that waffle being waffled, there is nothing wrong with creative ways of using Ai to enhance animations or videos obliviously. There are new tools and extensions and models etc coming very soon that will expand on what is currently possible with Ai created animations and videos. such as "Loose Control" and "Motion Control".
Watch this video to learn about it: .
 
Last edited:

rogue_69

Newbie
Nov 9, 2021
87
298
This is Ai enhanced animation rather than Ai created animation. I think that most would prefer to not create an animation in a 3d editor such as blender but rather rely mostly on SD instead.
I think there will be a big demand for both. Personally, I like animating in Daz, Blender, and Unity; but I'm not happy with the "polish" of those renderers. That's why I like the advancements in Video to Video. SD can really make a plastic looking animation look cinematic. It just needs to get more consistent.
 

DreamingAway

Member
Aug 24, 2022
248
640
I think there will be a big demand for both. Personally, I like animating in Daz, Blender, and Unity; but I'm not happy with the "polish" of those renderers. That's why I like the advancements in Video to Video. SD can really make a plastic looking animation look cinematic. It just needs to get more consistent.
Before stable diffusion - I always wanted my 3D Models to have a painted / artistic aesthetic to them. As if hand drawn.
I would of expected a solution to this be available through custom shaders but I was always disappointed with the results.

AI + Rotoscoping feels like the more likely technology to get there. Imagine extremely detailed hand draw art, animated and rendered at the speed of 3D - if that can be achieved it's almost the best of both worlds.

--

Before SD this was impossible.

Animation in old Disney movies always had extremely detailed backgrounds then simple flat shaded characters / animation because they had to draw them frame-by-frame. If a single frame takes an artist 3 weeks, you can't possibly achieve 24 frames per second and the likelihood of consistency falls dramatically as well.

This would be something AI could do (hopefully) that is essentially impossible today.
 

rogue_69

Newbie
Nov 9, 2021
87
298
Before stable diffusion - I always wanted my 3D Models to have a painted / artistic aesthetic to them. As if hand drawn.
I would of expected a solution to this be available through custom shaders but I was always disappointed with the results.

AI + Rotoscoping feels like the more likely technology to get there. Imagine extremely detailed hand draw art, animated and rendered at the speed of 3D - if that can be achieved it's almost the best of both worlds.
The beauty is that you'll be able to apply whatever style you want. I go back and forth between wanting to do realistic and cartoonish animation.
 

DreamingAway

Member
Aug 24, 2022
248
640
The beauty is that you'll be able to apply whatever style you want. I go back and forth between wanting to do realistic and cartoonish animation.
I just wish more work had been done in the industry to support "hand drawn styles" in 3D. Genshin did a pretty good job emulating "anime" and Arcane seems to be pushing the boundary on that "hand drawn" style but it's still not there for me.

When I see concept arts for certain video games or MMO's I'm usually disappointed in how they translated into the 3D world. I hope AI can make those a reality.
 
  • Like
Reactions: Mr-Fox
Dec 12, 2021
108
58
I was just wondering if someone could explain something to me. I'm starting out with AI generation using stable diffusion and I'm not sure how I would generate a consistent character ( planning to make a game with it).

How exactly did the developer of this game , create a consistent character with different clothes, poses etc. ?
 

Mr-Fox

Well-Known Member
Jan 24, 2020
1,401
3,802
I was just wondering if someone could explain something to me. I'm starting out with AI generation using stable diffusion and I'm not sure how I would generate a consistent character ( planning to make a game with it).

How exactly did the developer of this game , create a consistent character with different clothes, poses etc. ?
controlnet and/or img2img
 
  • Like
Reactions: Sepheyer

me3

Member
Dec 31, 2016
316
708
In this case the IPAdapter is your tool. You show it what dress you want, then inject it's output into the sampler. It is a "more better" approach then using prompt weight's. In fact it becomes a micromodel that must improve consistency.

If you need a snippet of the workflow let me know.
I did some tests with IPadapter, but i've tested so many different methods that i'm mixing up what i tried with what. I don't think i did it specifically with just the clothing though, that might work better.

Here's an upscale of an image i got for something else. It's using IPadap, with a closeup of the person and "far off" shot of a town/settlement with with foggy/cloudy surrounding as "seed images". SD15 without any detailer fix.
Upscale_0001.png
You should check out this video here for an alternative method for video generation that is perhaps better at temporal coherence:
Unfortunately the controlnet seems to be made for animatediff which works rather poorly with low VRAM and it's very limited in frame count and rate. Really wish someone would figure out a "non batching" method for it. If batching makes it possible to keep this "alike", it has to somehow be referencing each other or the same image/noise/whatever. So in theory you should be able to do much the same with looping. We already have ways to use images or parts of images as reference points, in a sense we have 100% recreateability for a single image too, so there has to be some way to combine the various "parts"...anyway
I'll take a look at the controlnet model, it might have some impact even without animatediff, thanks for pointing it out.
 

Mr-Fox

Well-Known Member
Jan 24, 2020
1,401
3,802
how exactly should I use img2img?

Can I use an image of a character as a reference then add prompts and LORA's to change clothes, poses etc. ?
I haven't experimented with this scenario so I can only give general tips. I suspect that you would either create a character in txt2img or use a ready made reference image. Then in img2img use controlnet with openpose for the poses and something like ip adapter for the clothes. The thing is you need to give more information in order for anyone to help you better because the workflow will depend what is your starting point and what is your goal in more specifics and details.
- Do you have a character already?
- Are you going to work in A1111 or ComfyUi?
Most importantly. Are you going to make something like the link you gave? Then count me out. I don't do lolly type stuff or anything other questionable content that looks even vaguely like CP
 
Last edited:
Dec 12, 2021
108
58
I haven't experimented with this scenario so I can only give general tips. I suspect that you would either create a character in txt2img or use a ready made reference image. Then in img2img use controlnet with openpose for the poses and something like ip adapter for the clothes. The thing is you need to give more information in order for anyone to help you better because the workflow will depend what is your starting point and what is your goal in more specifics and details.
- Do you have a character already?
- Are you going to work in A1111 or ComfyUi?
Most importantly. Are you going to make something like the link you gave? Then count me out. I don't do lolly type stuff or anything other questionable content that looks even vaguely like CP
Thank you for replying, to answer your last question, not at all, I share your disgust for people who like that, it was just the first thing I saw on dlsite that was similar to what I want to make.

I'm working on A1111, as for the character, I currently have nothing particular in mind regarding the character, I just started researching AI art yesterday and have no concrete plan yet.

My plan is to make a ntr game in RPGmaker that I'd sell on the japanese market, I've been seeing alot of these ai art games sell pretty quickly making a nice profit so I thought I'd try it out as a side hustle.

I haven't heard of controlnet and openpose so I'll have to read up on that. Thank you for answering :)
 
  • Like
Reactions: Mr-Fox

Sepheyer

Well-Known Member
Dec 21, 2020
1,570
3,766
Thank you for replying, to answer your last question, not at all, I share your disgust for people who like that, it was just the first thing I saw on dlsite that was similar to what I want to make.

I'm working on A1111, as for the character, I currently have nothing particular in mind regarding the character, I just started researching AI art yesterday and have no concrete plan yet.

My plan is to make a ntr game in RPGmaker that I'd sell on the japanese market, I've been seeing alot of these ai art games sell pretty quickly making a nice profit so I thought I'd try it out as a side hustle.

I haven't heard of controlnet and openpose so I'll have to read up on that. Thank you for answering :)
There is a hope to make consistent images where you first model them in DAZ / HS2 / Other and then run image-2-image with AI to add gloss as well as good old photoshop. I think the consensus so far is that no AI-only method can give you true consistency, although some combinations of methods can get you reasonably close.

Here is a start of one such discussion about "converting" DAZ girls into any style you want: https://f95zone.to/threads/stable-diffusion-prompt-sharing-and-learning-thread.146036/post-12343859
 
Dec 12, 2021
108
58
There is a hope to make consistent images where you first model them in DAZ / HS2 / Other and then run image-2-image with AI to add gloss as well as good old photoshop. I think the consensus so far is that no AI-only method can give you true consistency, although some combinations of methods can get you reasonably close.

Here is a start of one such discussion about "converting" DAZ girls into any style you want: https://f95zone.to/threads/stable-diffusion-prompt-sharing-and-learning-thread.146036/post-12343859
I've never used those two so I'll try to find some alternative solution for now, thank you anyways :)
 

me3

Member
Dec 31, 2016
316
708
Having been posting a lot of prompt/workflow-less images/clips lately i thought i'd do a cleaned up one to maybe give ppl ideas on how to do something. While this is from comfyui the image itself should make it possible to see what would be needed to set it up in a1111, i hope.
This is basically the same setup as i used to create the image in this post, and it can be used to layer or create "depth" in images. It obviously depends on what image you put to each layer and how they are handled/applied. I used a bunch of math and composite nodes to create the colored masking image, this can obviously be done other ways and in other shapes etc. IE if you put each "color/mask" next to each left to right, you could use it for panoramic views or multiple characters, a bit like regional prompts but with "image inputs".

This is not a very tidy/nice flow, i tried to keep things compact for the "screenshot" while still readable. Ipadapter nodes are colored by mask channel for visual aid. I've never installed nodes through anything else than the manager so any missing ones should be found there. I've included the mask image if anyone should want that.

workflow.png mask_0001.png

Edit: Adding an image using the workflow, just dropping 3 image i had into it.
_gmtmt_00022_.jpg
 
Last edited:

me3

Member
Dec 31, 2016
316
708
Before stable diffusion - I always wanted my 3D Models to have a painted / artistic aesthetic to them. As if hand drawn.
I would of expected a solution to this be available through custom shaders but I was always disappointed with the results.

AI + Rotoscoping feels like the more likely technology to get there. Imagine extremely detailed hand draw art, animated and rendered at the speed of 3D - if that can be achieved it's almost the best of both worlds.

--

Before SD this was impossible.

Animation in old Disney movies always had extremely detailed backgrounds then simple flat shaded characters / animation because they had to draw them frame-by-frame. If a single frame takes an artist 3 weeks, you can't possibly achieve 24 frames per second and the likelihood of consistency falls dramatically as well.

This would be something AI could do (hopefully) that is essentially impossible today.
You mentioning hand drawn stuff reminded me of some images i got quite a while ago. I was running prompts of descriptions of fictional places, and for some reason this is the "style" and look the AI decided to give when using various descriptions of Rivendell. While maybe not hand drawn it reminds me of old book illustrations and that kind of thing.

_00031_.jpg _00025_.jpg