- Mar 3, 2019
- 237
- 775
Yeah, I tried a few different methods, my current one actually involves making an initial image without the lora, but using IP Adapter for a basic likeness with SDXL models to get a first pass with better colors and composition (but lower facial fidelity), and then using img2img with ControlNet to recreate that visual with SD1.5 + Lora, to add the likeness back to the image. I don't usually like using faceswap methods after because I feel they tend to make facial expressions look too bland for my taste, though I can't say I have really tried ReActor yet. But at least in my personal experience, and for my needs I feel that loras give me higher fidelity than any other method I have tried while giving me full control over character pose and composition on the prompt, and I usually don't feel the need to add another step to increase that fidelity.Don't get me wrong I love using loras too. They are just very inconsistent. I think it's all about the use case scenario though. If you are happy with the result you get, that's all that matters. One thing doesn't exclude the other though. You can use both, or neither.
It's very difficult to get perfect likeness with a lora in my experience no matter how good it is. With a faceswap you get closer though. There is a different option that has not been mentioned and that is outpainting. Essentially only re generating the face from an image and outpainting the rest. You can just mask the face and select inpaint not masked. Have you tried to use a good lora and then use a faceswap over it? This would make use of the bone structure from the lora and get the better likeness from the faceswap. Essentially using the best from both. You can use real images with openpose etc for the different scenarios.
There are always new things to try. I think the fact that SD is not perfect is one of the things that keeps it interesting. If it were easy there wouldn't be any "sport"..
While I won't say that I never have issues getting a good resemblance when using loras, I feel that it's usually the result of a lora that didn't get trained with the best source images or just the random nature of Stable Diffusion, but I have some loras that to me are basically as good as it gets, so the inconsistency problems you describe, again in my experience are less due to an issue with limitations of the technology, and more a case of bad implementation (i.e. poorly trained loras).
For example, here are some generations using my custom models compared to photos of the real women trained in the loras:
So yeah, that's why I don't think loras are going anywhere. If you want to be able to create images from the prompt without being restricted to get a likeness from pre-existing photo, loras are a lot more flexible and lead to better quality than the alternatives... to me at least. But my quality bar for likeness is perhaps higher than the average user, so don't take my words as gospel.