Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question: Inpainting model for SDXL #45

Open
Markus-Pobitzer opened this issue Jul 17, 2023 · 21 comments
Open

Question: Inpainting model for SDXL #45

Markus-Pobitzer opened this issue Jul 17, 2023 · 21 comments

Comments

@Markus-Pobitzer
Copy link

Dear Stability AI Team

Thanks for an other great release!

Do you plan on releasing an inpainting model for SDXL too?

@aleksmirosh
Copy link

I will much appreciate any news about inpainting too!

@mlconnor
Copy link

+1

1 similar comment
@jiqizaisikao
Copy link

+1

@wangqyqq
Copy link

Waiting for the inpainting model for SDXL, too!

@bghira
Copy link

bghira commented Jul 26, 2023

it will be v-prediction and zero terminal SNR, most likely. i think that will come with 1.1.

@yang-fei
Copy link

when to release the inpainting model? there will be much use of the inpainting model

@jaidevshriram
Copy link

Thanks @bghira! Is there a timeline for this? Also, is the SDXL-base model v-prediction too or is that also in the 1.1 release?

@Markus-Pobitzer
Copy link
Author

SDXL can already be used for inpainting, see: https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion/stable_diffusion_xl#inpainting

To me it seems like they are using the base model for inpainting.

Will there be a fine tuned version like stable-diffusion-2-inpainting?

@landmann
Copy link

Anyone notice inpainting with a mask alters the image the mask says it should not be altering? Is this expected?

@Markus-Pobitzer
Copy link
Author

Markus-Pobitzer commented Jul 27, 2023

@landmann If you are referring to small changes, than it is most likely due to the encoding/decoding step of the pipeline. We bring the image in a latent space (containing less information than the original image) and after the inpainting we decode it back to an actual image but in this process we are losing some information (the encoder is lossy as mentioned by the authors). This is expected behavior.

Workaround can be found here: huggingface/diffusers#3514 (comment)

@landmann
Copy link

landmann commented Jul 27, 2023

Haha good ol ctrl+c ctrl+v - love it.

@shreyassks
Copy link

Is it possible to finetune SDXL 1.0 base model for inpainting?

I tried to do that but I get a error from UNET that it expects the input to have 4 channels but it got 9 channels (4 - input image, 4 - masked image, 1 - mask (B/W))

Am I doing anything wrong or should I wait for inpainting model release?

@bghira
Copy link

bghira commented Aug 13, 2023

if you're feeling adventurous you can modify the pix2pix finetuner from Diffusers project.

it already uses 8 channel input.

@nasy
Copy link

nasy commented Aug 17, 2023

+1

@JohanHuynh0130
Copy link

+1 im waiting for inpainting model of SDXL

@JohanHuynh0130
Copy link

SDXL INPAINTING MODEL has been out!

@yjhong89
Copy link

yjhong89 commented Nov 6, 2023

Is there any in-paint configs for SDXL 1.0 (base)?

@gel-crabs
Copy link

A yaml file for SDXL-inpainting is desperately needed.

@kalle07
Copy link

kalle07 commented Dec 19, 2023

yes pls !

@kalle07
Copy link

kalle07 commented Dec 21, 2023

@wangqyqq
Copy link

as i mentioned in webui post, this sdxl-inpaint model can work in 16-bit float and you don't need to appy --no-half-vae or --no-half.
you can download the model, merge the code in pull request AUTOMATIC1111/stable-diffusion-webui#14390 and try it!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

16 participants