Fooocus 2.1.0 Image Prompts (Midjourney Image Prompts) #557
Replies: 94 comments 134 replies
-
Awesome! |
Beta Was this translation helpful? Give feedback.
-
Using more images leads to worse result quality |
Beta Was this translation helpful? Give feedback.
-
wow .... fooocus is getting better and better |
Beta Was this translation helpful? Give feedback.
-
This looks like an incredible update! Could someone help me use it please? When I try to put an image in the image prompt and render I just get an error message:
EDIT: It works fine if I select "PyraCanny" or "CPDS" but if I try to use "Image Prompt" is gives me that ValueError. |
Beta Was this translation helpful? Give feedback.
-
hard to say ...for me works ok |
Beta Was this translation helpful? Give feedback.
-
2.1.19: PyraCanny improved a bit |
Beta Was this translation helpful? Give feedback.
-
@lllyasviel the new release is awesome. Is there anyways to do style blending like this? Or like what is done here: https://www.tensorflow.org/tutorials/generative/style_transfer I have tried the latest image prompt but couldn't get similar results. |
Beta Was this translation helpful? Give feedback.
-
fixed some errors in CPDS in 2.1.24 |
Beta Was this translation helpful? Give feedback.
-
The new features are so cool. The "Image Prompt" one seams a bit like "Revision". I'm enjoying the pose control we get from "PyraCanny" too! Very nice! Thank you for the sweet new features! |
Beta Was this translation helpful? Give feedback.
-
2.1.24: |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Hint: you do not need to turn off "Fooocus V2" in most cases. "Fooocus V2" is handled in a different way than text prompts. You do not need to worry about unwanted texts are added to your prompts. |
Beta Was this translation helpful? Give feedback.
-
can anyone please help me figure out how can i put my/anyone's face into this |
Beta Was this translation helpful? Give feedback.
-
Guys, how to make my AI influencer stop looking at the camera ? |
Beta Was this translation helpful? Give feedback.
-
How do prompt weights work?
Ciao, .mau.
…On Tue, Jun 25, 2024, 16:54 Manuel Schmid ***@***.***> wrote:
@theaccofai <https://github.com/theaccofai> You can define your prompts
accordingly, using keywords like side view or head towards XYZ may work.
Also consider using prompt weights for specific keywords.
—
Reply to this email directly, view it on GitHub
<#557 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAMV6KQNO2KN665VOTXY5C3ZJGACVAVCNFSM6AAAAAA5XHHTN2VHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM4TQNZSG4YTG>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
Hi Guys, |
Beta Was this translation helpful? Give feedback.
-
Hello guys launched Fooocus using Google Colab, used Pyracanny, Image Prompt, and other features. However, there is still a problem that I can not fix. I generated like thousands of pics and still can't figure out how to make my ai influencer not look at the camera + how to make face expressions like models from insta. For example, taking out the tongue with eyes closed, or looking left, and etc. Please help me figure this out, if you can answer here or text me on telegram my nickname @theaccofg |
Beta Was this translation helpful? Give feedback.
-
Dear friends, can someone help me create real people photos in exactly this style using Fooocus, thank you very much. |
Beta Was this translation helpful? Give feedback.
-
How can i download this on my Chromebook i've been having some problems with it because im new to this and i really wanted to download this |
Beta Was this translation helpful? Give feedback.
-
What is the best model or method to make faces similar but not identical, like siblings (brother/sister)? Also, how can I maintain the same style and posture? |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
please can you explain me the different use of "stop at" and "weight"? whic is the result on the final immage created? it's not clear, tnx |
Beta Was this translation helpful? Give feedback.
-
sorry for the dumb question, but is there a way to run Fooocus 2.5 from colab? I noticed the warning "You are using gradio version 3.41.2, however version 4.44.1 is available, please upgrade." and I think that maybe I am using an old notebook. |
Beta Was this translation helpful? Give feedback.
-
How I can configure (via Docker) fooocus to use the CPU? |
Beta Was this translation helpful? Give feedback.
-
Broken image links in this guide. Guide still useful, but though you might want to know. |
Beta Was this translation helpful? Give feedback.
-
Hello. I would like to thank each and every one of them for their work. He does great work. I have a request from you, please provide the possibility of outputting as vector or SVG. (If you have, can you tell me how to do it? I searched but I could not find it). Thank you very much again |
Beta Was this translation helpful? Give feedback.
-
Hello, Great work, team! I would like to integrate this using an API—is this possible and permitted? I reviewed the Gradio API documentation included in the UI, but it seems a bit confusing. Could you please assist me? I also want to contribute in terms of development if possible. Thank you in advance! |
Beta Was this translation helpful? Give feedback.
-
Fooocus 2.1.0 has completed the implementation of image prompts. Because after this version, almost all features of Midjourney are included, the version directly jump to 2.1.0.
Image Prompt is one of the most important feature of Midjourney. Below is the banner from Midjourney:
In Fooocus, it looks like this:
Technically, this feature is based on a mixture of IP-Adapter, and a pre-computed negative embedding from Fooocus team, an attention hacking algorithm from Fooocus team, and an adaptive balancing/weighting algorithm from Fooocus team.
The motivation of these efforts is to achieve a best match to the Midjourney Image Prompt. In other software like A1111/ComfyUI/InvokeAI, the IP-Adapter still has some open problems like ignoring text prompts, or over-burned results when multiple images are used. These problems are solved in Fooocus and users can enjoy Midjourney-like experience of Image Prompt.
The detailed differences are in the below table:
Using this method will download 2.5GB files at the first time!
Example: Single Image Prompt without Text Prompts
(Non-cherrypicked random batch, default parameters, real results should be better if tuned)
(seed 1234, here is the image)
(this example uses default style and Fooocus V2 style)
Example: Single Image Prompt with Text Prompts
Note that mixing text and IP-Adapter is extremely difficult in ComfyUI/A1111. Fooocus does not have this problem.
(Non-cherrypicked random batch, default parameters, real results should be better if tuned)
(this example uses default style and Fooocus V2 style)
Example: Multiple Images without Text Prompts
Note that mixing multiple IP-Adapters is likely to cause lower result quality in ComfyUI/A1111. Using Fooocus can resolve this to some extents.
(Non-cherrypicked random batch, default parameters, real results should be better if tuned)
(this example uses default style and Fooocus V2 style)
Example: Multiple Images with Text Prompts and Even Multiple Styles
This is almost impossible in A1111/ComfyUI since mixing text and IP-Adapter is extremely difficult in ComfyUI/A1111, and mixing multiple IP-Adapters is likely to cause lower result quality in ComfyUI/A1111.
(Non-cherrypicked random batch, default parameters, real results should be better if tuned)
This image is too complicated to understand so I annotated here:
So mixing too many things make it hard to recognize but everything is there and it does not fail or causing quality decerase, unlike ComfyUI/A1111/InvokeAI.
Fooocus Image Prompt (Advanced)
If you check “advanced”, you will be able to use two structure controls:
PyraCanny: A pyramid-based Canny edge control. The reason is that SDXL uses 1024px images and standard Canny tends to miss some image details from time to time, at such a high resolution. This method uses multiple resolutions to detect canny edges and then combine them softly, so that more structures are captured (than canny). The pyramid part is from “Edge Drawing: A combined real-time edge and segment detector”. You will download 350MB control models when using it.
CPDS: A structure extraction algorithm from “Contrast Preserving Decolorization (CPD)”. The “CPDS” means CPD Structure. The control model is modified by Fooocus team – it starts from SAI’s depth control-lora. The reason for using this method is for the fast speed and download-free preprocessor. Note that we only use the structure part of images, and it is not really “decolorization”. You will download 350MB control models when using it.
(Non-cherrypicked random batch, default parameters, real results should be better if tuned)
(this example uses default style and Fooocus V2 style)
(this example uses default style and Fooocus V2 style)
(this example uses default style and Fooocus V2 style)
For developers:
In Developer Debug Mode, you can mix the upscale/vary/inpaint with all above features if you know what you are doing and REALLY need it (the denoising strength can also be set in Developer mode). You can also get the preprocessor result by checking the “debug preprocessor”.
But keep in mind:
If you accidentally get satisfying results in Fooocus by tuning a lot of advanced parameters, you should try to copy your positive prompt, reopen Fooocus, do not change anything, and paste the prompt. You will find that results are even better, and all those tunings are unnecessary. (The only exception is probably changing base model in “Advanced”.)
Beta Was this translation helpful? Give feedback.
All reactions