“A cozy biophilic seaport village. In the distance there are tall building and plants. There are spaceships flying above. In the foreground there is a cute dog sitting on a bench.” - ArtBot https://tinybots.net/artbot

        • tal@lemmy.today
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          2 hours ago

          I’m not familiar with Artbot.

          investigates

          Yes, it looks like it supports inpainting:

          https://tinybots.net/artbot/create

          Look down in the bottom section, next to “Image-to-image”.

          That being said, my experience is that inpainting is kind of time-consuming. I could see fine-tuning the specific look of a feature – like, maybe an image is fine except for a hand that’s mangled, and you want to just tweak that bit. But I don’t know if it’d be the best way to do this.

          • I don’t know if this is actually true, but I recall reading that prompt term order matters for Stable Diffusion (assuming that that is the model you are using; it looks like ArtBot lets you select from a variety of models). Earlier prompt terms tend to define the scene. While I’ve tended to do this, I haven’t actually tried to experiment enough to convince myself that this is the case. You might try sticking the “dog” bit earlier in the prompt.

          • If this is Stable Diffusion or an SD-derived model and not, say, Flux, prompt weighting is supported (or at least it is when running locally on Automatic1111, and I think that that’s a property of the model, not the frontend). So if you want more weight to be placed on a prompt term, you can indicate that. Adding additional parentheses will increase weight of a term, and you can provide a numeric weight: A cozy biophilic seaport village. In the distance there are tall building and plants. There are spaceships flying above. In the foreground there is a cute ((dog)) sitting on a bench. or A cozy biophilic seaport village. In the distance there are tall building and plants. There are spaceships flying above. In the foreground there is a cute (dog:3) sitting on a bench.

          • In general, my experience with Stable Diffusion XL is that it’s not nearly as good as Flux at taking in English-language descriptions of relationships between objects in a scene. That is “dog on a bench” may result in a dog and a bench, but maybe not a dog on a bench. The images I tend to create with Stable Diffusion XL tend to be a list of keywords, rather than English-language sentences. The drawback with Flux is that it’s heavily weighted towards creating photographic images, and I’m guessing, from what you submitted, that you’re looking more for a “created by a graphic artist” look.

          EDIT: Here’s the same prompt you used fed into stoiquoNewrealityFLUXSD35f1DAlphaTwo, which is derived from Flux, in ComfyUI:

          Here it is fed into realmixXL, which is not derived from Flux, but just from SDXL:

          The dog isn’t on the bench in the second image.