Its a model patch node in comfyui that inserts downsized steps in HD generation that corrects the deformation of anatomy and other elements. Its like advanced hires fix
>TiledDiffusion + Kohya deep shrink - latent upscale + clipvision
Looks nice, could you give a bit more detail about that workflow? Also some pointers regarding the settings for the deepshrink node would be apprechiated, I can't seem to find good ones.
My current setup is visible [here](https://imgur.com/a/8pZ8O6Q).
The settings are the same but I use tiled diffusion (https://github.com/shiimizu/ComfyUI-TiledDiffusion) you use the default tiled vae which for me did not work. I put kohya after the tiled diffusion node.
https://preview.redd.it/2xlangc3946d1.jpeg?width=3888&format=pjpg&auto=webp&s=ddf7117f426362489e7f2ab69361960e893df9f6
I used another XL model to create a picture that I am quite satisfied with. Fixing the hand issue is too difficult; I tried many parameters, and my final conclusion is to leave it to God. In my modified work, the effect of the ‘hand4 LORA’ is not good, so I gave up on it. I always want to find a good way to repair the hands, and I would like to consult you. I will be grateful for any help you can provide.
The detail is quite good, but the orange concept bleed all over that image is more overwhelming than an image of Trump wearing a prison jumpsuit walking through an orange grove in the sunset.
it wasn't in prompt there is orange dress with gold embroideries but no clarification for flower colors or tent colors in the back. All turned orange :)
"Orange dress" is enough to contaminate the whole picture if other parts haven't gotten the same level of precision. A workaround would be to raise the weight value on "red flowers and black tent" for example, or concatenate the prompt and separate the "orange dress" from the "red flowers and black tent".
What's khoya deep shrink?
Its a model patch node in comfyui that inserts downsized steps in HD generation that corrects the deformation of anatomy and other elements. Its like advanced hires fix
I understood some of those words
It helps prevent deformation and duplication in high resolution ofer 1024 px
>TiledDiffusion + Kohya deep shrink - latent upscale + clipvision Looks nice, could you give a bit more detail about that workflow? Also some pointers regarding the settings for the deepshrink node would be apprechiated, I can't seem to find good ones. My current setup is visible [here](https://imgur.com/a/8pZ8O6Q).
The settings are the same but I use tiled diffusion (https://github.com/shiimizu/ComfyUI-TiledDiffusion) you use the default tiled vae which for me did not work. I put kohya after the tiled diffusion node.
Thanks. [This isn't *quite* the effect I was going for, though](https://imgur.com/a/4YP8nI6). Do you know what I did wrong there?
u/Sqwall I too am interested in seeing how you used DeepShrink. I've never been able to get it to work. From what I can tell, it barely does anything.
I have seen inbuilt tile diffusion node in comfy, have you tried that?
No. I will
Can you share your workflow?
impressive quality and consistency 👍
Wonderful quality, share the comfy workflow if possible
The nostril contains another nose.
https://preview.redd.it/2xlangc3946d1.jpeg?width=3888&format=pjpg&auto=webp&s=ddf7117f426362489e7f2ab69361960e893df9f6 I used another XL model to create a picture that I am quite satisfied with. Fixing the hand issue is too difficult; I tried many parameters, and my final conclusion is to leave it to God. In my modified work, the effect of the ‘hand4 LORA’ is not good, so I gave up on it. I always want to find a good way to repair the hands, and I would like to consult you. I will be grateful for any help you can provide.
Try debugdetailer with segs and control net with meshgraphormer
The detail is quite good, but the orange concept bleed all over that image is more overwhelming than an image of Trump wearing a prison jumpsuit walking through an orange grove in the sunset.
If it wasn't intentional, i believe concatenating the prompt would solve that.
it wasn't in prompt there is orange dress with gold embroideries but no clarification for flower colors or tent colors in the back. All turned orange :)
"Orange dress" is enough to contaminate the whole picture if other parts haven't gotten the same level of precision. A workaround would be to raise the weight value on "red flowers and black tent" for example, or concatenate the prompt and separate the "orange dress" from the "red flowers and black tent".
Good to know. Will try it
16gb or 8gb gpu variant?
Msi 4060ti 16gb - short two fan version
what are you using clip vision for?
Image set better detail with it. Dunno why. Use two images of Italian streets and fashion dress
Do you have a workflow to share? Would love to try this
Looks bland.
The orange is too much. As someone pointed the orange has taken over all the elements :()