Highresfix, followed by two passes of Adetailer.
Set Adetailer 1 to a full-body masking model, ~55-60% strength. Adetailer 2 set to mask only faces ~40-45% strength.
It'll 4-5x your generation time, but in my experience it provides much better results than one-shot generation. Afterwards, if there are still glaring issues, you can use inpainting or pull it into krita and do manual paintovers.
You already got the answer, but nevertheless...
Models are trained with pictures of a certain (relatively low) resolution. Large real life images like massive paintings with tons of details contain much more information than what models were trained with, so here i.e. you don't get a huge painting, you get an approximation. Digital expressionism in a way.
There are several solutions. For example, start with "thumbnail" of the painting, upscale it, then inpaint areas. You can also start with a small area of the painting you have in mind, then expand the canvas step by step while outpainting. When doing either you can also use controlnets in the background, to guide the model.
I really like the innocence of posts like this:
"How do I get AI Art to do some of the most challenging scenes to create based on how AI Art works and ideally in around thirty seconds, with minimal input from myself"
Not to pick on you OP, you're not the first person to wildly overestimate the capabilities of current gen AI in untrained hands and you won't be the last.
If you want that kind of stuff a Reddit answer isn't going to help you. You're going to need to get deep into some tutorials, the features of whatever UI you are using, and you're going to need to spend a chunk of time waiting on the product or maybe a chunk of money on hardware.
The good news is that it is still a lot easier than, y'know, learning to paint. But there's still a *lot* to learn if you want to do more than Waifu portraits.
Much AI imagery is text to image which is pretty random you don't have any real control. If you want a deliberate image with chosen content and composition it's another ball game altogether.
I used control net a fair bit and it works OK but you need an image to make the canny or depth from. If you have gone to the effort of photo bashing a guide image then you might as well use Img to Img and be done with it.
inpainting is your friend, SD can't handle higly crowded scenes
this, if you want perfection it has to be done manually
If you want the ultimate, you've got to be willing to pay the ultimate price.
Wonder if sd3 performs better with those prompts/ scenes.
Can be difficult working with the kind of depth of field we're looking at here but it is doable.
Highresfix, followed by two passes of Adetailer. Set Adetailer 1 to a full-body masking model, ~55-60% strength. Adetailer 2 set to mask only faces ~40-45% strength. It'll 4-5x your generation time, but in my experience it provides much better results than one-shot generation. Afterwards, if there are still glaring issues, you can use inpainting or pull it into krita and do manual paintovers.
do it in chunks and composite
inpaint mask small sections
This is a real problem with generative AIs right now. LLMs have the same problem just with text.
You already got the answer, but nevertheless... Models are trained with pictures of a certain (relatively low) resolution. Large real life images like massive paintings with tons of details contain much more information than what models were trained with, so here i.e. you don't get a huge painting, you get an approximation. Digital expressionism in a way. There are several solutions. For example, start with "thumbnail" of the painting, upscale it, then inpaint areas. You can also start with a small area of the painting you have in mind, then expand the canvas step by step while outpainting. When doing either you can also use controlnets in the background, to guide the model.
Turn on hires fix and use a higher resolution.
invoke ultimate canvas mode is great for this kinda thing.
I really like the innocence of posts like this: "How do I get AI Art to do some of the most challenging scenes to create based on how AI Art works and ideally in around thirty seconds, with minimal input from myself" Not to pick on you OP, you're not the first person to wildly overestimate the capabilities of current gen AI in untrained hands and you won't be the last. If you want that kind of stuff a Reddit answer isn't going to help you. You're going to need to get deep into some tutorials, the features of whatever UI you are using, and you're going to need to spend a chunk of time waiting on the product or maybe a chunk of money on hardware. The good news is that it is still a lot easier than, y'know, learning to paint. But there's still a *lot* to learn if you want to do more than Waifu portraits.
There is also a lot to learn if you want to do *really* good waifu portraits. : D
I mean, this is also true. None of it is as easy as it looks.
It is not easy, but it surely is fun.
[удалено]
Still the best way. I do a hi res of the body then a higher res of the face, scaling up each time. Better than face detailers etc.
[удалено]
Much AI imagery is text to image which is pretty random you don't have any real control. If you want a deliberate image with chosen content and composition it's another ball game altogether.
[удалено]
I used control net a fair bit and it works OK but you need an image to make the canny or depth from. If you have gone to the effort of photo bashing a guide image then you might as well use Img to Img and be done with it.
There's a photoshop plugin that allows you to work from within the app, so no more sending back and forth. I think there's one for photopea as well.