-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ControlNet 1.1] The updating track. #736
Comments
Update: ControlNet 1.1 is released here. |
I think we can ignore cnet11 Tile model right now. We are not very sure how to make use of it. |
I’m the author of sd-webui-segment-anything and I am planning to connect my extension to your inpainting model. So at this moment, the inpainting ControlNet cannot target at the mask only while not changing other parts, right? Edit on 2023/04/18 already connected. Checkout my extension readme for how to use. |
I have been long working on tiles. Have you tried cooperating with noise inversion tricks? I think this can be very good, with a better trained model it may be comparable to the quality of GigaGAN. My extension is here -> https://github.com/pkuliyi2015/multidiffusion-upscaler-for-automatic1111 I will adapt your tile model to see the result and update it here. |
Yes the tile model can be a saviour for upscaling and no doubles |
This thread is already amazing. ^ 3 amazing devs collaborating |
my gradio demo does not have masked diffusion in it. what is displayed now is just original results from standard non-masked diffusion. but masked diffusion will be better. |
The model works as expected in automatic1111 txt2img; it does generate the guided content. However, as I directly download the model and use it in this extension, it produces severe artifacts. I read the source code for a while but still not clear what should be done to make it work. Some initial observations:
See here for one result: https://imgsli.com/MTY5ODQw |
what preprocessor we should use with tile controlnet model ? Using it without preprocessor gets "some" results but the resolution is kinda lower than if i would inpaint with 0.55 denoise, have to use cfg 2-3 |
Do you think there is a need to wait for an update of this extension? Is the current extension compatible with the new models, especially the inpainting model? |
which one is cn11tile? left or right? |
The right one. I must have done something wrong. But until now I cannot fix it. |
Is there a PR in this repo yet for implementing ControlNet v1.1? |
from the result it looks like your input image is bigger than h/8 w/8. for example, if you diffuse at 512 512, your tile need to be 64 64 and then use 3 cv2. pyrup to interpolate to 512. or you can add a gaussion blur to the inputs to make it smoother |
Hi I have a recommended list of updates: Control Model: Depth: Normal: Canny/MLSD: Scribble: SoftEdge: Segmentation: Openpose: Lineart: Shuffle: What do you think? |
That list looks good to me. Are the instructpix2pix and inpainting models already working out of the box? The former seemed to work but I also felt like it gave me mixed results, but I wasn't going to judge the quality yet, not knowing if it's missing something. Inpainting model I haven't tried yet. Tile model I assume would come a bit later since the model itself is in an unfinished state currently. |
PR WIP at #742. |
Recently renaming of annotators caused some downstream developers unhappy. We can implement renamings as display name change instead of ID change which causes API breaks. Also on naming, the annotator name should imply which cnet model should be used and vice versa. |
i have an idea. what about adding some descriptions to yaml file of each cnet like "xxx_canny.yaml" has a "desc: this model needs canny preprocessor" and show it to gradio ui? |
The gradio part seems less than ideal. List items cannot show hover infos, at least, I tried the the DDIM sampler item in WebUI, it doesn't, however if you select it and hover on the selection box it shows. |
i mean like adding a gradio.label or something and show some desc text from model yaml after a model is loaded. |
if u think it is ok i will begin to work on all 14 yaml files |
What about the old cnets (prior to 1.1)? They have no isolated yamls. I think it's better to implement it at code level, which is also localization friendly. I will wait response from repo owner. |
old cnets can just use blank text. we can only show texts when desc is avaliable |
@Mikubill why invert always binarize images? |
That is a known issue, will be fixed. |
Would be awesome to auto select the most likely model after preproc is
selected & vice versa. Won't prevent people from changing, but will save a
needed step 90% of the time.
…On Thu, Apr 13, 2023 at 3:30 PM lllyasviel ***@***.***> wrote:
i have an idea. what about adding some descriptions to yaml file of each
cnet like "xxx_canny.yaml" has a "desc: this model needs canny
preprocessor" and show it to gradio ui?
—
Reply to this email directly, view it on GitHub
<#736 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAAEQMXRITOKXTD7KL6AY53XBBH4FANCNFSM6AAAAAAW4RNXZE>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Although it is not official, temporalnet also uses none as preprocessor. |
What is temporalnet, and is there any documentation for using it with a1111 + controlnet? Sounds interesting... |
Here is their hugging face: https://huggingface.co/CiaraRowles/TemporalNet. You can see an example of output on the linked twitter post: https://twitter.com/CiaraRowles1/status/1637486561917906944 It is not perfect, but it helps a lot with getting consistent results. It worked best with a lower weight in my limited testing. To use it with this extension, you must put controlnet in batch mode and enable the checkbox "loopback" at the bottom of the unit that has temporalnet as the selected model. What this does is, it will automatically feed the latest generated image of the batch back into this unit for each iteration. The first image of the batch for this unit is either the first image of the selected directory or the uploaded image. |
hello everyone, the finished tile model is released #1033 |
Can hires fix be used with ControlNet on just the second pass of hires fix? I don't have an input image for the first pass. I want ControlNet to use the first low-res generated image as the input image to help compose the hires second pass image. If I don't put in an input image at the start, it says there is no input image, and doesn't use ControlNet at all. |
What prevents you from simply generating and transferring the photo to the controlnet area? |
I don't want ControlNet to use the low-res image on the first pass, which I want SD to generate purely from the text prompt, just on the second pass when it makes the image hi-res (to help control that second pass). Typically if you use a high denoising in hires fix, then the composition is lost in the second pass. ControlNet could be used to help maintain composition on that second pass, even at higher denoising strengths (the new Tile model could be very helpful here to refine details on the second hires pass). |
You can use ControlNet in img2img then. Hires fix is simply an img2img shortcut. |
Yes. But then it makes it a two-step process to make a hires image from a text prompt. The purpose of hires fix is to automate that and make it a one-step process to generate a hires image in txt2img. It seems like if you don't have an input image to begin with, ControlNet should just use the first pass low-res image generated as the control image. |
I have met this problem in mac os, and I solved if by installing the "Certificates.command" in the python application |
if you are in China and has SSL: CERTIFICATE_VERIFY_FAILED, please turn off the proxy. You can directly connect huggingface without proxy in China |
But what if the annotator creates artifacts during preprocessing?
Edit:
@lllyasviel Can this be fixed? |
hi everyone we improved preview window in |
@theboylzh seems fixed in 1.1.128, try update and restart completely |
@theboylzh Upgrading python version to 3.10+ may help you! |
See PEP 604. The pipe operator is introduced in Python 3.10. Please confirm that you are using the correct python version. |
@huchenlei |
hi everyone, a bit late but we added depth leres++ |
It seems |
I cannot reproduce, using 1 unit as leres++ depth. Can you share your settings and other extensions? Also, you may want to create an issue instead of reporting an error here to help us keep track of the resolution. |
we will close this issue because it is too long. please use Discussion or Issue to report problems or discuss features |
hi everyone we released a discussion about tile model's color offset problem with anime images/models |
hi everyone we released a [Major Update] Reference-only Control |
We will use this repo to track some discussions for updating to ControlNet 1.1.
The text was updated successfully, but these errors were encountered: