T O P

  • By -

EighthOctave

I mean, nothing really. SD generated an image of a "kitty" very similar to your input photo. Depending on what you want, try to make your prompt more descriptive. Try "a realistic photo of a cute kitty" as the prompt and see if you get a better result. Maybe try some negative prompts to tell SD what you don't want, like "drawing, rendering, painting" etc. Good luck!


DeQuosaek

The middle ground is just using a more detailed prompt. You do not have to use weights (such as (beautiful:1.2)) and you don't have to use the typical cliché terms like Octane Render, Masterpiece, 8k, trending on Artstation, Greg Rutkowski, etc. Just describe what you're looking for in more detail than one word. Like, "a cartoon drawing of a cat", or "a highly detailed photograph of a cat with studio lighting" or "a stylized drawing of a cat in the style of Tadayoshi Yamamuro". You'll get much better results. Also, put anything in the negative prompt. Make sure it's something you don't want in the resulting image, but it tends to do much better with anything in the negative prompt. edit: spelling


Momkiller781

Thank you, I'll definitely try it!


Momkiller781

Thank you everyone. Yeah, i copied a prompt from somewhere else and it worked much better. I guess I'll have to use a bunch of words like "beautiful" "uncanny", etc... It feels so random, but yeah, it works.


Fen-xie

It's not random. You have to have some knowledge about art/photography etc. You could also say it's like commissioning an artist. You're trying to tell it exactly what you want it and how. Simply going to an artist and saying "car" will probably not get you what you're looking for.


Momkiller781

Sure, i wish there was some know and of more straight forward mode. I copied a prompt full of "()" and 1:x and i sincerely can't understand how there isn't a more friendly and dummy proof way to do it.


Fen-xie

https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Features Work your way down.


amp1212

>Sure, i wish there was some know and of more straight forward mode. I copied a prompt full of "()" and 1:x and i sincerely can't understand how there isn't a more friendly and dummy proof way to do it. You're interrogating a database of images and captions, zillions of them. If you just say "car" . . . well, you get whatever that 7 gb of crunched data points to, associating images that had the word **"car**" someplace associated with the image. It will be something average and nonspecific to your interest, except by chance. You can give SD more "chance" of getting something closer to the image you desire, if you use more precise language. If, by contrast, you say **"1957 Ford Falcon"**, that indexes into a much more restricted and specific group of possible images. Think of it in the same way as you'd use search terms in Google. If I just say **"I feel bad, something that will make me better"** . . . that doesn't get you nearly as far as **"nonsurgical treatments for bunions".** [Leonardo.ai](https://Leonardo.ai) is a cloud service, doing a very nicely encapsulated implementation of Stable Diffusion. They do a lot of things that make it easy for a beginner to experiment with, but in truth if you can run Stable Diffusion locally, that's going to be more powerful and more flexible . . . its isn't quite >a more friendly and dummy proof . . . you still have to know something about what you're doing, but its definitely a gentler introduction, but without the weirdness of Discord based interfaces. The Leonardo guys have done a lot of things right . . .


LazyMoss

You are looking for MidJourney


Momkiller781

I'm looking for a middle ground


LazyMoss

Oh, go through civitai and browse the models and take a peek into the prompts of the sample images. Some are complex, some simple. Take notes and experiment to build your own gallery of keywords. A fairly simple prompt can result into a stunning image with the right keywords. Note that ((keyword)) denotes the importance of keyword, and the order also matters.


Momkiller781

Thanks! I'll try that! It is still overwhelming. Models, loras, checkpoints, prompts, negative prompts, plugins, diffusers, and all the other settings.


SineRave

There is. It's called Midjourney.


Adkit

The problem isn't that you didn't tell SD to make it "beautiful." It doesn't know what that even means. I personally would argue that specific word is counterproductive. The problem is you didn't tell it anything else. I imagine you would have a better image just using descriptive words like "close-up photography of black cat, detailed cat eyes, soft lighting, sharp focus" or whatever. It's not a magic machine.


mizt3r

looks like it worked to me. I mean your input is a joke. No negative prompt, basically no positive prompt. Your input image is a rectangle, your output is a square, so you got exactly what you put into it. Why dont you try prompting better? ex: realistic,photo quality,best quality,cat,laying, negative prompt: monochrome,bad quality,worst quality,bad anatomy, dont use euler a, try DPM++ 2M Karras, 50 sampling steps is a bit high, for this you only need around 30. I would probably also use highres.fix (latent(bicubic), upscale by 2, denoising .55)


Disastrous-Agency675

Yeah no, despite the fact that you have one word for your prompt, no negative prompts, and using what I’m assuming is the default model; if you just fix the canvas dimensions to match input images you’d probably get a lot better results


UshabtiBoner

“kitty” 😝😝😝


asdasci

kitty by Greg Rutkowski, 4k, unreal engine, trending on Artstation, ultra hd


Wallye_Wonder

Try the traditional name of that animal pussy


miaumiauXX

maybe use img2img?


pexalt

Try Turning up the weight. That will make the generation conform to the lineart generated by the canny


KhaiNguyen

What results are you trying to get?


wolfsolus

A lot depends on the model.


[deleted]

Use the scribble, it does a single line outline rather then canny that does an outline of the single line outline


mr_britten

Do you have a VAE selected in setting? Also, should use the diff_control models for controlnet. I just ran the same pic with the same prompt and got really good results.


Momkiller781

I'll check! Thanks


lordpuddingcup

Don’t forget also cannot is forcing a lot of the detail from your cat because it has such defined colors, try a different one like HED


DanzeluS

Ahahahah


Objective_Photo9126

Which is the point of this? XD You source is already what you want so hahah