Giter VIP home page Giter VIP logo

fooocus-mre's Introduction

Fooocus-MRE

image

Fooocus-MRE is an image generating software (based on Gradio), an enhanced variant of the original Fooocus dedicated for a bit more advanced users.

Fooocus-MRE is a rethinking of Stable Diffusion and Midjourney’s designs:

  • Learned from Stable Diffusion - the software is offline, open source, and free.

  • Learned from Midjourney - it provides high quality output with default settings, allowing users to focus on the prompts and images.

  • Learned from SD web UI and ComfyUI - more advanced users would like to have some control over image generation process.

Fooocus has included and automated lots of inner optimizations and quality improvements. Users can forget all those difficult technical parameters, and just enjoy the interaction between human and computer to "explore new mediums of thought and expanding the imaginative powers of the human species" [1].

Fooocus has simplified the installation. Between pressing "download" and generating the first image, the number of needed mouse clicks is strictly limited to less than 3. Minimal GPU memory requirement is 4GB (Nvidia).

Fooocus also developed many "fooocus-only" features for advanced users to get perfect results. Click here to browse the advanced features.

[1] David Holz, 2019.

Download

Windows

You can directly download Fooocus with:

>>> Click here to download <<<

After you download the file, please uncompress it, and then run the "run.bat".

image

In the first time you launch the software, it will automatically download models:

  1. It will download sd_xl_base_1.0_0.9vae.safetensors from here as the file "Fooocus\models\checkpoints\sd_xl_base_1.0_0.9vae.safetensors".
  2. It will download sd_xl_refiner_1.0_0.9vae.safetensors from here as the file "Fooocus\models\checkpoints\sd_xl_refiner_1.0_0.9vae.safetensors".
  3. Note that if you use inpaint, at the first time you inpaint an image, it will download Fooocus's own inpaint control model from here as the file "Fooocus\models\inpaint\inpaint.fooocus.patch" (the size of this file is 1.28GB).

image

If you already have these files, you can copy them to the above locations to speed up installation.

Note that if you see "MetadataIncompleteBuffer", then your model files are corrupted. Please download models again.

Below is a test on a relatively low-end laptop with 16GB System RAM and 6GB VRAM (Nvidia 3060 laptop). The speed on this machine is about 1.35 seconds per iteration. Pretty impressive – nowadays laptops with 3060 are usually at very acceptable price.

image

Besides, recently many other software report that Nvidia driver above 532 is sometimes 10x slower than Nvidia driver 531. If your generation time is very long, consider download Nvidia Driver 531 Laptop or Nvidia Driver 531 Desktop.

Note that the minimal requirement is 4GB Nvidia GPU memory (4GB VRAM) and 8GB system memory (8GB RAM). This requires using Microsoft’s Virtual Swap technique, which is automatically enabled by your Windows installation in most cases, so you often do not need to do anything about it. However, if you are not sure, or if you manually turned it off (would anyone really do that?), or if you see any "RuntimeError: CPUAllocator", you can enable it here:

Click here to the see the image instruction.

image

And make sure that you have at least 40GB free space on each drive if you still see "RuntimeError: CPUAllocator" !

Please open an issue if you use similar devices but still cannot achieve acceptable performances.

Colab

(Last tested - 2023 Sep 13)

Colab Info
Open In Colab Fooocus Colab (Official Version)
Open In Colab Fooocus-MRE Colab (MoonRide Edition)

Thanks to camenduru!

Linux (Using Anaconda)

If you want to use Anaconda/Miniconda, you can

git clone https://github.com/MoonRide303/Fooocus-MRE.git
cd Fooocus-MRE
conda env create -f environment.yaml
conda activate fooocus
pip install pygit2==1.12.2

Then download the models: download sd_xl_base_1.0_0.9vae.safetensors from here as the file "Fooocus\models\checkpoints\sd_xl_base_1.0_0.9vae.safetensors", and download sd_xl_refiner_1.0_0.9vae.safetensors from here as the file "Fooocus\models\checkpoints\sd_xl_refiner_1.0_0.9vae.safetensors". Or let Fooocus automatically download the models using the launcher:

conda activate fooocus
python entry_with_update.py

Or if you want to open a remote port, use

conda activate fooocus
python entry_with_update.py --listen

Linux (Using Python Venv)

Your Linux needs to have Python 3.10 installed, and lets say your Python can be called with command python3 with your venv system working, you can

git clone https://github.com/MoonRide303/Fooocus-MRE.git
cd Fooocus-MRE
python3 -m venv fooocus_env
source fooocus_env/bin/activate
pip install pygit2==1.12.2

See the above sections for model downloads. You can launch the software with:

source fooocus_env/bin/activate
python entry_with_update.py

Or if you want to open a remote port, use

source fooocus_env/bin/activate
python entry_with_update.py --listen

Linux (Using native system Python)

If you know what you are doing, and your Linux already has Python 3.10 installed, and your Python can be called with command python3 (and Pip with pip3), you can

git clone https://github.com/MoonRide303/Fooocus-MRE.git
cd Fooocus-MRE
pip3 install pygit2==1.12.2

See the above sections for model downloads. You can launch the software with:

python3 entry_with_update.py

Or if you want to open a remote port, use

python3 entry_with_update.py --listen

Linux (AMD GPUs)

Installation is the same as Linux part. It has been tested for 6700XT. Works for both Pytorch 1.13 and Pytorch 2.

Mac/Windows(AMD GPUs)

Coming soon ...

List of "Hidden" Tricks

Below things are already inside the software, and users do not need to do anything about these.

  1. GPT2-based prompt expansion as a dynamic style "Fooocus V2". (similar to Midjourney's hidden pre-processsing and "raw" mode, or the LeonardoAI's Prompt Magic).
  2. Native refiner swap inside one single k-sampler. The advantage is that now the refiner model can reuse the base model's momentum (or ODE's history parameters) collected from k-sampling to achieve more coherent sampling. In Automatic1111's high-res fix and ComfyUI's node system, the base model and refiner use two independent k-samplers, which means the momentum is largely wasted, and the sampling continuity is broken. Fooocus uses its own advanced k-diffusion sampling that ensures seamless, native, and continuous swap in a refiner setup. (Update Aug 13: Actually I discussed this with Automatic1111 several days ago and it seems that the “native refiner swap inside one single k-sampler” is merged into the dev branch of webui. Great!)
  3. Negative ADM guidance. Because the highest resolution level of XL Base does not have cross attentions, the positive and negative signals for XL's highest resolution level cannot receive enough contrasts during the CFG sampling, causing the results look a bit plastic or overly smooth in certain cases. Fortunately, since the XL's highest resolution level is still conditioned on image aspect ratios (ADM), we can modify the adm on the positive/negative side to compensate for the lack of CFG contrast in the highest resolution level. (Update Aug 16, the IOS App Drawing Things will support Negative ADM Guidance. Great!)
  4. We implemented a carefully tuned variation of the Section 5.1 of "Improving Sample Quality of Diffusion Models Using Self-Attention Guidance". The weight is set to very low, but this is Fooocus's final guarantee to make sure that the XL will never yield overly smooth or plastic appearance (examples here). This can almostly eliminate all cases that XL still occasionally produce overly smooth results even with negative ADM guidance. (Update 2023 Aug 18, the Gaussian kernel of SAG is changed to an anisotropic kernel for better structure preservation and fewer artifacts.)
  5. We modified the style templates a bit and added the "cinematic-default".
  6. We tested the "sd_xl_offset_example-lora_1.0.safetensors" and it seems that when the lora weight is below 0.5, the results are always better than XL without lora.
  7. The parameters of samplers are carefully tuned.
  8. Because XL uses positional encoding for generation resolution, images generated by several fixed resolutions look a bit better than that from arbitrary resolutions (because the positional encoding is not very good at handling int numbers that are unseen during training). This suggests that the resolutions in UI may be hard coded for best results.
  9. Separated prompts for two different text encoders seem unnecessary. Separated prompts for base model and refiner may work but the effects are random, and we refrain from implement this.
  10. DPM family seems well-suited for XL, since XL sometimes generates overly smooth texture but DPM family sometimes generate overly dense detail in texture. Their joint effect looks neutral and appealing to human perception.
  11. A carefully designed system for balancing multiple styles as well as prompt expansion.
  12. Using automatic1111's method to normalize prompt emphasizing. This significantly improve results when users directly copy prompts from civitai.
  13. The joint swap system of refiner now also support img2img and upscale in a seamless way.

Advanced Features

Click here to browse the advanced features.

MoonRide Edition Features

  1. Support for Image-2-Image mode.
  2. Support for Control-LoRA: Canny Edge (guiding diffusion using edge detection on input, see Canny Edge description from SAI).
  3. Support for Control-LoRA: Depth (guiding diffusion using depth information from input, see Depth description from SAI).
  4. Support for Control-LoRA: Revision (prompting with images, see Revision description from SAI).
  5. Adjustable text prompt strengths (useful in Revision mode).
  6. Support for embeddings (use "embedding:embedding_name" syntax, ComfyUI style).
  7. Customizable sampling parameters (sampler, scheduler, steps, base / refiner switch point, CFG, CLIP Skip).
  8. Displaying full metadata for generated images in the UI.
  9. Support for JPEG format.
  10. Ability to save full metadata for generated images (as JSON or embedded in image, disabled by default).
  11. Ability to load prompt information from JSON and image files (if saved with metadata).
  12. Ability to change default values of UI settings (loaded from settings.json file - use settings-example.json as a template).
  13. Ability to retain input files names (when using Image-2-Image mode).
  14. Ability to generate multiple images using same seed (useful in Image-2-Image mode).
  15. Ability to generate images forever (ported from SD web UI - right-click on Generate button to start or stop this mode).
  16. Official list of SDXL resolutions (as defined in SDXL paper).
  17. Compact resolution and style selection (thx to runew0lf for hints).
  18. Support for custom resolutions list (loaded from resolutions.json - use resolutions-example.json as a template).
  19. Support for custom resolutions - you can just type it now in Resolution field, like "1280x640".
  20. Support for upscaling via Image-2-Image (see example in Wiki).
  21. Support for custom styles (loaded from sdxl_styles folder on start).
  22. Support for playing audio when generation is finished (ported from SD web UI - use notification.ogg or notification.mp3).
  23. Starting generation via Ctrl-ENTER hotkey (ported from SD web UI).
  24. Support for loading models from subfolders (ported from RuinedFooocus).
  25. Support for authentication in --share mode (credentials loaded from auth.json - use auth-example.json as a template).
  26. Support for wildcards (ported from RuinedFooocus - put them in wildcards folder, then try prompts like __color__ sports car with different seeds).
  27. Support for FreeU.
  28. Limited support for non-SDXL models (no refiner, Control-LoRAs, Revision, inpainting, outpainting).
  29. Style Iterator (iterates over selected style(s) combined with remaining styles - S1, S1 + S2, S1 + S3, S1 + S4, and so on; for comparing styles pick no initial style, and use same seed for all images).

Thanks

The codebase starts from an odd mixture of Stable Diffusion web UI and ComfyUI. (And they both use GPL license.) MoonRide Edition is based on the original Fooocus. Big thanks to AUTOMATIC1111, comfyanonymous, and lllyasviel for providing those fantastic tools.

Thanks to Stability AI for researching and opening their Stable Diffusion model series, OpenAI for CLIP and mlfoundations for OpenCLIP, and LAION AI for data sets on which those models could learn.

Special thanks to twri and 3Diva for creating additional SDXL styles available in Fooocus.

Update Log

The log for original version is here, and for enhancements added in MRE here.

fooocus-mre's People

Contributors

camenduru avatar lllyasviel avatar moonride303 avatar sngazm avatar tcmaps avatar ttio2tech avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

fooocus-mre's Issues

Save default setting for use next time

A good to have would be
An option to save settings like the resolution, style, batch and output as a default.
(for example, I need to select JPG with metadata and change the resolution every time), so that the next launch of the application, the same settings do not have to be applied again.

Support for Custom Styles

I have my own custom styles that I edit into the sdxl_styles.py file. But everytime I start Fooocus it updates everything and I lose my custom styles. Is there a way to keep my styles, and/or stop Fooocus from updating every start.

Feature Request: Custom path for models folder

Instead of having multiple copies of checkpoints, lora,s embeddings TI, I keep them in once central place on my fastest SSD drive (This will make Fooocus faster) then I instruct A1111 and Comfyui to look there.

Request for custom path for checkpoints, Loras, and outputs in settings.json

I believe the file to edit is Fooocus-MoonRide/modules/path.py

Priority is not urgent, I could just hard code in path.py. Only want to know if MoonRide is interested or has the time?

Feature request: Store prompt in exif tag named 'comment' instead of 'prompt'

Storing the prompt data in a tag called 'prompt' is logical but it's a non-standard tag and image viewers don't see it. At least not Photoshop, Macos-Preview or XnView.

If you store the prompt data in a tag called name 'comment' it's immediately available.

Given that were talking about 1 kilobyte of data, could you alter your code to store it in both tags 'prompt' AND 'comment' ?

Screenshot 2023-08-16 at 17 39 26

[Request] - Multi Gpu usage and manual

The software works fine with 1 gpu, but it completely ignores other. It would be nice if it could automaticly generate few images at the same time depending on the amount of gpus computer has. Also it's now hard to understand how to use ImgToImg mode properly for a new user, so I think it's worth adding some information about how to use that program in ReadMe or somewhere else. Thanks for your work!

Feature Request: Image Browser/Photo Album

Hello! First of all, thanks a million for putting this together!
I would love to see something like a Image Browser / Photo Album feature like the one available on InvokeAI.
For those running it on a headless server on Linux, it's a bit difficult to find the files after a while.
image

img2img works

Just starting to play with img2img and am so happy. Thank you

So far I only played with 2 sliders from the Advanced tab "Img2img Start Step" and "Img2img Denoise"

Also I want to as if you should make a "discussions" area on GitHub similar to what lllyasviel did here https://github.com/lllyasviel/Fooocus/discussions

I feel the "Issues" area should reserve for bug report.

Feature Request: Latent Composite with Visual Guide

ComfyUI features two primary nodes for latent compositing: 'Latent Composite' and 'Latent Composite Masked'. These nodes offer enhanced control over image composition.

Additionally, GitHub user Davemane42 has provided custom nodes that feature an intuitive graphical representation of latent placement within the image.

https://github.com/Davemane42/ComfyUI_Dave_CustomNode
image

Is it feasible to integrate such a feature into Fooocus-MRE?
I understand that implementing this request might not be straightforward. However, it would be an extremely valuable feature, especially if users could interactively set up the image composition in the GUI by simply dragging and resizing the separate latent areas.

AttributeError: 'NoneType' object has no attribute 'clip'

Is it necessary to opt for the SDXL Refiner model? Can it be omitted from selection?

Exception in thread Thread-2 (worker):
Traceback (most recent call last):
File "threading.py", line 1016, in _bootstrap_inner
File "threading.py", line 953, in run
File "E:\Fooocus_win64_1-1-10\moonride-main\Fooocus\modules\async_worker.py", line 115, in worker
handler(task)
File "E:\Fooocus_win64_1-1-10\moonride-main\Fooocus\modules\async_worker.py", line 73, in handler
imgs = pipeline.process(p_txt, n_txt, steps, switch, width, height, seed, sampler_name, cfg, base_clip_skip, refiner_clip_skip, callback=callback)
File "E:\Fooocus_win64_1-1-10\python_embeded\lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "E:\Fooocus_win64_1-1-10\moonride-main\Fooocus\modules\default_pipeline.py", line 127, in process
xl_refiner.clip.clip_layer(refiner_clip_skip)
AttributeError: 'NoneType' object has no attribute 'clip'

change `settings-example.json` to `settings.json`

I think it will be easier for everyone if settings-example.json became settings.json and people edit in it directly to change the defaults without renaming it and go through the documentations.

any way to see log.html with gradio?

I had focus installed on server with --share
How can I see log.html file with gradio link?

Can we have someting like this - https://[4afc4f0bf5bbcf259.gradio.live/internal/log.html

~help~ How to highlight options?

Hello, thank you for everything you have done, so that we can get the picture soon. Can you tell me the specific parameters of various options of the sampler used in the original FOC? If I want to display its options as highlights (such as orange). Used to refer to the comparison and then gradually reduce the parameters to adjust the appropriate time, because how to do it? (The original FOC picture quality is very good is very slow, the keywords are not very strict, and I am not familiar with XL new sampler, the picture is easy to feel plastic) Thank you ~

Feature Request: IPAdapter and Latent Upscale (Neural Network)

Ipadapter is a better and more robust Revision with more control. It can work as a real-time 1 image LoRa. People are doing revision + ipadapter and other combinations using Comfy. It would be really nice to have this in MRE.

ComfyUI Custom Node: https://github.com/laksjdjf/IPAdapter-ComfyUI

Link to the model: https://huggingface.co/h94/IP-Adapter/tree/main/sdxl_models

About IPAdapter: https://github.com/tencent-ailab/IP-Adapter

===

Upscale

SAG on Fooocus helps a lot, but MRE would need an effective upscale strategy to be even better. Some ideas:

The attempt by city96: https://github.com/city96/SD-Latent-Upscaler

By WAS/Ttl: https://github.com/Ttl/ComfyUi_NNLatentUpscale

These are some suggestions to create an upscaler for MRE; it would make this project even better.

Thank you, MoonRide, for this excellent project. You made Foocus much better, and thank you for the Colab edition. It's good for us folks that don't have a proper computer to run SDXL, and I'm using it every day.

Error accessing through a proxy server

Hello
I'm running Fooocus-MRE on a Cloud Instance and if it's accessed by the ip-address:port it works fine. But if I access it using a proxy sever (fooocus.example.com) it opens, but I see the error in the screenshot.

I'm starting the service with python launch.py --listen 0.0.0.0 --port 7866 but I have tried use the server public address, dns name, etc and get the same error.

Any tips?

Thanks for you fantatic work.
image

subfolders

have set custom paths in the paths.py but it doesnt recognise subfolders for checkpoints or loras

Adding a toggle button to the resolutions

There are a bunch of resolutions / image sizes , but as we can see now, the dropdown is TOO long.
I propose having either a toggle button for landscape/ portrait mode or an aspect ratio button to quickly switch without getting bogged down in the long dropdown.

image

image

Addendum.
I don't see any 16:9 / 9:16 ratios or 16:10, which are pretty common and widely used. For example
I feel those are necessary additions and should have been part of the original project too.

I've calculated some resolutions that might be useful
1152 x 720 - 16:10
1152 x 648 - 16:9
1280x720 - 16:9

Edit 2:
(Note from the GUI perspective - the dropdown is indeed way too long now, and serves as a distraction rather than a 'quick' generation.
For other members contributing to the project, it might be helpful to discuss / see if it might be more feasible to say enter a width or height value, and use the toggle to fill in the other resolution based on aspect, or just enter values in both fields (faster).

Ubuntu: /usr/bin/www-browser: line 216: /mnt/c/Windows/System32/reg.exe: No such file or directory

When staring in linux there are errors:

[error] WSL Interoperability is disabled. Please enable it before using WSL.
/usr/bin/www-browser: line 216: /mnt/c/Windows/System32/reg.exe: No such file or directory
/usr/bin/www-browser: line 149: /mnt/c/Windows/System32/chcp.com: No such file or directory
/usr/bin/www-browser: line 156: /mnt/c/Windows/System32/WindowsPowerShell/v1.0/powershell.exe: No such file or directory
/usr/bin/www-browser: line 149: /mnt/c/Windows/System32/chcp.com: No such file or directory

but works fine

Ability to save as JPG

Could you add the ability to save as JPG? (With EXIF info that contains metadata)
A1111 has this functionality.
I'm trying out MRE today, and it is great as a simple image generator.

I can't also find the DPM++ 2M scheduler in the dropdown.

Feature request: Generate Forever

I post this request at lllyasviel/Fooocus but I will post this here too. Pretty sure lllyasviel have higher-piority things to do

In A1111 you can right-click the generate button and select "generate forever" then right-click again when you want to stop.

This way you can have Fooocus working all night while you sleep or all day while you are at work.

This addition doesn't affect the simple and elegant interface.

Please, add to output time taken to render image. In the log also, if possible

start_time = time.time()
imgs = pipeline.process(p_txt, n_txt, steps, switch, width, height, seed, sampler_name, scheduler,
cfg, base_clip_skip, refiner_clip_skip, input_image_path, start_step, denoise, callback=callback)
end_time = time.time()
elapsed_time = end_time - start_time
print(f"\033[91mTime taken: {elapsed_time} seconds\033[0m")
shot_230823_214522

Request to bring a custom resolution box

@MoonRide303 I know you closed the toggle issue, but not a big deal.
I'm trying to work with F-MRE, but that long dropdown of resolutions is way too long. I showed this to a GUI/UX person who felt the same way, in what is otherwise a very very usable and clean interface.

A 'nice to have' : A custom resolution box be implemented for X / Y (Width, Height) resolution, and this will allow people to simply type in their required resolution and start working rather than scroll through the list each time they want a different res.

photo editing

Hi, first of all I have to say that your work is excellent.Is it possible to upload a photo (or send a link) and edit based on the uploaded photo, if there is such a feature, please guide me to use it.

Expose more options in the GUI

It would be useful to have a tickbox that exposes the CFG and sampler options on the front page. It great having this all hidden away but sometimes you want to experiment.

BUG: Symbolic Link gets wipe out

Got symbolic links working (thanks) but whenever Fooocus update, it deletes symbolic links and replace with actual folders and starts download missing safetensors. Link to Symbolic Link tutorial https://youtu.be/RDH5IuyPJtk

My current solution is hard code modules/path.py

Before
modelfile_path = os.path.abspath(os.path.join(os.path.dirname(file), '../models/checkpoints/'))
lorafile_path = os.path.abspath(os.path.join(os.path.dirname(file), '../models/loras/'))

After
modelfile_path = os.path.abspath(os.path.join(os.path.dirname(file), 'D:\stable-diffusion-shared\Ckpt\XL'))
lorafile_path = os.path.abspath(os.path.join(os.path.dirname(file), 'D:\stable-diffusion-shared\Lora\SDXL'))

Feature Request: Abort button

In the original repository maintained by Illyasviel, I submitted a feature request for an 'Abort Button' for image generation
(see Issue #85).
However, it appears that this request has since been overshadowed by 100+ subsequent issues. Would it be possible for this repository to consider implementing this feature?

img2img dosent follow selected resolution

The supported official resolutions are very low resolution, so to get high resolution output without bad composition I tried to give 1920x1080px in resolutions to do img2img from low resolution to high.
But this doesn't work and there is no high res fix option as well so all the output images with such low resolution are not usefull at all unless i do img2img in A1111 with output of foocus.
That just wastes time and i would rather just use A1111 at that point.
There should be some way to to do from low res to high in Fooocus.

Instructions for newbs

I'm sure this is NOT the best approach but this worked for me on Windows.

Assuming you already have the original Fooocus 7z file downloaded, installed and working.

Close Fooocus if its running.
cd into the parent folder that contains Foooccus / python-embeded / run.bat
git clone https://github.com/MoonRide303/Fooocus.git Fooocus-MoonRide

duplicate run.bat to run-moonride.bat

Copy (or move) your models folder from Fooocus/models to Fooocus-MoonRide/models

edit run-moonride.bat and change Fooocus to Fooocus-MoonRide

double click the run-moonride.bat

[ Request ] Integration of an upscaler and face restoration

Thank you Moonride303 for this fantastic upgrade of Fooocus. It is gold. The only thing which is still missing in Fooocus is an integration of an upscaler like Esrgan and a face restoration like Codeformer/Gfpgan. Thank you for your work!

Appeal for the style choosing Panel

Hi @MoonRide303 ,
There are more than 20 styles in the advanced panel, when choosing any of them, it's really hard to choose.
And if there are preview photo like this, that might be more easy to use:

2023-08-31 152946

I know that would not be simple any more..
But it's really hard for non-english user T T...

By the way, how shall I do to add translation package to this project, would you like to add the i-18 folder?

[Request] Full screen preview

After generating, click on the image and it will load in fullscreen in sort of a lightbox style. I always end up having to dig into the output folder, and then open them up to see the smaller details.

Would be awesome, while inspecting in fullscreen, one can reject/delete the image and auto advance to the next image

Feature request: outpainiting

Since Fooocus aims at matching mj's simplicity and popular functions, I am wondering if there is any plan to add outpainting in Fooocus/Foooocus-MRE.

Simple top, left, bottom, right outpainting with some default settings would be sufficient.

Great Job! but would it support SD 1.5 models?

Hi MoonRide303,

This version is sooooooo cooooooool!

I like it so much for new feathers like the prompt information in png, and image 2 image

Well, after I modified the path-example.json and wishing to use sd 1.5 models and loras in fooocus-mre, it seems do not work well.
Like this :example, I changed the model to reacatoonanimeV4.safetensors, and same seed, same prompt( which is fairy girl) and same negative prompt, result is like this:
2023-08-28_19-55-24_8667

and the same setting on webui, the result is like this:
00200-2065610250-Fairy girl

I'm not sure what's the problem, would you like to fix this issue?

Thank you so much : )
👍

docker of fooocus-mre

docker pull cgwyx/fooocusmre:cuda116-xformers

docker run -it --name fooocusmre-xformers --gpus /"device=1/" --network host /
-v /mnt/p1/diffusion/Fooocus-MREdocker:/Fooocus-MRE /
-v /mnt/p1/diffusion/models/Stable-diffusion:/Fooocus-MRE/models/checkpoints /
-v /mnt/p1/diffusion/models/loras:/Fooocus-MRE/models/loras /
-v /mnt/p1/diffusion/outputs:/Fooocus-MRE/outputs /
cgwyx/fooocusmre:cuda116-xformers bash

conda activate fooocus && python launch.py

log.html and gradio. lifehuck and suggestion

I use FocusMRE and it is great.
I use it locally (much faster then Automatic1111 on my RTX 2080 8GB) and also on linux server with gradio and --share.

May be I am stupid but I found no way to see log.html when using gradio.

Lifehuck: I see that Automatic1111 can open links like:
https://47c763360a19fc9500.gradio.live/file=/data/ptits/a1111/stable-diffusion-webui/outputs/txt2img-images/2023-09-07/

So, I set output path for FocusMRE то stable-diffusion-webui/outputs/txt2img-images/ and then I can open
https://47c763360a19fc9500.gradio.live/file=/data/ptits/a1111/stable-diffusion-webui/outputs/txt2img-images/2023-09-07/log.html with no problem

But I need to have Automatic working on sever.

I found no directory than gradio of FocusMRE can access, is there one? (like in Automatic). (searching for gr.Interface had now results)

Can you make default "outputs" directory be accessible from gradio?
Something like:
https://47c763360a19fc9500.gradio.live/file=/data/ptits/FocusMRE/outputs/2023-09-07/log.html

missing {'cond_stage_model.clip_l.text_projection'

Whenever I switch models and press Generate I get

missing {'cond_stage_model.clip_l.text_projection', 'cond_stage_model.clip_g.transformer.text_model.embeddings.position_ids', 'cond_stage_model.clip_l.logit_scale'}

Generate again it goes away. Only happens after you switch model

Also get it when 1st launch Fooocus

Print image meta data to screen for easy copy/paste

I appreciate we now have this data in the PNG but having it printed to screen somewhere would also be useful, especially the seed and image dimensions.

BTW if I'm saturating you with requests please just say!

Plot X/Y/Z

[Fooocus-MRE] is real top.
I switched from A1111 completely.
Last update of A1111 works with refiner in a very bad way, switching between model takes more time than rendering itself.
What I missed from A1111 is PlotXYZ.
For tweaking and experimenting it is a killer feature.
Now I need to tweak the code to run loops for denoising or other parameters and then selecting best ones
Is there any hope to see some kind of looping through parameters (no legend needed, log is fine)?

One day I made loop with style and got some funny results:
#from modules.sdxl_styles import style_keys
#for key in style_keys:
#style = key
#pipeline.clean_prompt_cond_caches()

I described it here - lllyasviel#249

Some more fun here:
https://cgevent.ru/imgs/collage/DynaVision/1table.html
https://cgevent.ru/imgs/collage/PhysGen/1table.html

All done with Fooocus-MRE.

So PlotXYZ can be real killer in Fooocus.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.