Giter VIP home page Giter VIP logo

stable-diffusion-webui-docker's Introduction

Stable Diffusion WebUI Docker

Run Stable Diffusion on your machine with a nice UI without any hassle!

Setup & Usage

Visit the wiki for Setup and Usage instructions, checkout the FAQ page if you face any problems, or create a new issue!

Features

This repository provides multiple UIs for you to play around with stable diffusion:

Full feature list here, Screenshots:

Text to image Image to image Extras

Full feature list here, Screenshots:

Text to image Image to image Extras

Full feature list here, Screenshot:

Workflow

Contributing

Contributions are welcome! Create a discussion first of what the problem is and what you want to contribute (before you implement anything)

Disclaimer

The authors of this project are not responsible for any content generated using this interface.

This license of this software forbids you from sharing any content that violates any laws, produce any harm to a person, disseminate any personal information that would be meant for harm, spread misinformation and target vulnerable groups. For the full list of restrictions please read the license.

Thanks

Special thanks to everyone behind these awesome projects, without them, none of this would have been possible:

stable-diffusion-webui-docker's People

Contributors

66li avatar abdbarho avatar adamflorizone avatar cloudaxes avatar cococig avatar devilan avatar divens avatar gmasil avatar jsjolund avatar kagurazakanyaa avatar lev145 avatar mashb1t avatar moulai avatar passivelemon avatar producermatt avatar redtopper avatar rgriffogoes avatar sejoung avatar simonmcnair avatar wookiefriseur avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

stable-diffusion-webui-docker's Issues

MacOS Support

Steps to Reproduce

  1. clone the repo
  2. copy model.ckpt to models
  3. cd ./AUTOMATIC1111
  4. docker compose up --build

There are no problems with the previous steps, but the last error is reported:

Use 'docker scan' to run Snyk tests against images to find vulnerabilities and learn how to fix them
[+] Running 2/2
 ⠿ Network automatic1111_default    Created                                                                                                 0.0s
 ⠿ Container automatic1111-model-1  Created                                                                                                 0.1s
Attaching to automatic1111-model-1
Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]

Hardware / Software:

  • MacOS 12.5 M1 Pro chip
  • GPU: No GPU

What do I need to do to fix this?

NSFW Filter

It's unclear to me whether or not this build has the NSFW filter option activated or not. Regardless of the default, it would be good if this could be toggled within the webUI so that users can decide whether or not they would like to generate those sorts of images. Is that an easy thing to do?

"failed to solve "Error on M2 Mac...

Hello!

I'm trying to get this running on an M2 Mac, and am getting this error... does anyone know what it means or how to solve it?

failed to solve: executor failed running [/bin/bash -ceuxo pipefail conda install python=3.8.5 && conda clean -a -y]: exit code: 1

Thanks!!

lstein tokenizer error

Has this issue been opened before? Check the FAQ, the issues

invoke-ai/InvokeAI#34
Running into this error, how would this be resolved in your setup?

Describe the bug

Which UI

hlky or auto or auto-cpu or lstein?

Steps to Reproduce

  1. Go to '...'
  2. Click on '....'
  3. Scroll down to '....'
  4. See error

Hardware / Software:

  • OS: [e.g. Windows / Ubuntu and version]
  • RAM:
  • GPU: [Nvidia 1660 / No GPU]
  • VRAM:
  • Docker Version, Docker compose version
  • Release version [e.g. 1.0.1]

Additional context
Any other context about the problem here. If applicable, add screenshots to help explain your problem.

Issue #21 is back: Execution bit lost on /docker/mount.sh

Building the container works. Starting the container fails (again) with the following message:

stable-diffusion-webui-docker-model-1 | + /docker/mount.sh
stable-diffusion-webui-docker-model-1 | /bin/bash: line 1: /docker/mount.sh: Permission denied

Steps to Reproduce

docker compose build
docker compose up --build
[error...]

Hardware / Software:

OS: Windows 11/WSL2
Docker version
GPU: [Nvidia 3090]
Docker version: v20.10.17

Additional context
Attempted to add a chmod +x /docker/mount.sh before the command is run in dockerfile, the command does execute but does not improve things. Docker seems to work otherwise. A different version seems to have been deployed.

AWS: Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]

Has this issue been opened before? Check the FAQ, the issues

Describe the bug
Build of hlky or auto fails in the final step with "Attaching to webui-docker-automatic1111-1
Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]" on an AWS g5.xlarge using an AWS Nvidia A10G Tensor Core GPU.

This occurs after step 12 in the build and web interface will not launch following the error. Error is still present after reboot and rebuild of images.

Behavior duplicated on Debian 11 and Ubuntu 22.04, both with and without adding Nvidia drivers per https://levelup.gitconnected.com/how-to-install-an-nvidia-gpu-driver-on-an-aws-ec2-instance-20185c1c578c (This shouldn't have been necessary in any case, since AWS g5 instances include the appropriate Nvidia drivers.)

Which UI
hlky or auto

Steps to Reproduce

  1. Install per instructions on AWS g5.xlarge
  2. Error will show on shell screen after step 12/12 of the auto build

Hardware / Software:

  • OS: Debian 11 and Ubuntu 22.04
  • RAM: 16GB
  • GPU: Nvidia A10G Tensor Core
  • VRAM: 24GB
  • Docker Version, Docker compose version: Latest
  • Release version [e.g. 1.0.1]: Latest

Failed Inference with GFPGAN

I get this error when trying to use GFPGAN:

Failed inference for GFPGAN: Input type (torch.FloatTensor) and weight type (torch.cuda.FloatTensor) should be the same or input should be a MKLDNN tensor and weight is a dense tensor.

Should some flag be set to avoid using Torch's FloatTensor?

AMD Support?

Hello,

First of all, brilliant idea to make a docker image for this, this is very cool 👍

When I started up the docker image, I got this error

[+] Building 0.9s (20/20) FINISHED
 => [internal] load build definition from Dockerfile                                              0.0s
 => => transferring dockerfile: 32B                                                               0.0s
 => [internal] load .dockerignore                                                                 0.0s
 => => transferring context: 2B                                                                   0.0s
 => resolve image config for docker.io/docker/dockerfile:1                                        0.3s
 => CACHED docker-image://docker.io/docker/dockerfile:1@sha256:443aab4ca21183e069e7d8b2dc6800659  0.0s
 => [internal] load .dockerignore                                                                 0.0s
 => [internal] load build definition from Dockerfile                                              0.0s
 => [internal] load metadata for docker.io/continuumio/miniconda3:4.12.0                          0.3s
 => [internal] load build context                                                                 0.0s
 => => transferring context: 29B                                                                  0.0s
 => [ 1/11] FROM docker.io/continuumio/miniconda3:4.12.0@sha256:977263e8d1e476972fddab1c75fe050d  0.0s
 => CACHED [ 2/11] RUN conda install python=3.8.5 && conda clean -a -y                            0.0s
 => CACHED [ 3/11] RUN conda install pytorch==1.11.0 torchvision==0.12.0 cudatoolkit=11.3 -c pyt  0.0s
 => CACHED [ 4/11] RUN git clone https://github.com/hlky/stable-diffusion.git && cd stable-diffu  0.0s
 => CACHED [ 5/11] RUN conda env update --file stable-diffusion/environment.yaml --name base &&   0.0s
 => CACHED [ 6/11] RUN apt-get update && apt install fonts-dejavu-core && apt-get clean           0.0s
 => CACHED [ 7/11] RUN cd stable-diffusion && git pull && git reset --hard f3ab556a0c25389cf76f8  0.0s
 => CACHED [ 8/11] RUN git clone https://github.com/hlky/stable-diffusion-webui.git && cd stable  0.0s
 => CACHED [ 9/11] COPY info.py /info.py                                                          0.0s
 => CACHED [10/11] RUN  python /info.py /stable-diffusion/frontend/frontend.py                    0.0s
 => CACHED [11/11] WORKDIR /stable-diffusion                                                      0.0s
 => exporting to image                                                                            0.0s
 => => exporting layers                                                                           0.0s
 => => writing image sha256:3e58109d2fb9a6751d41b1bf3b6b0a8cd16de27bfdf2b1ba7d68964c7a1efea9      0.0s
 => => naming to docker.io/library/stable-diffusion-webui-docker_model                            0.0s

Use 'docker scan' to run Snyk tests against images to find vulnerabilities and learn how to fix them
[+] Running 2/1
 - Network stable-diffusion-webui-docker_default    Created                                       0.8s
 - Container stable-diffusion-webui-docker-model-1  Created                                       0.1s
Attaching to stable-diffusion-webui-docker-model-1
Error response from daemon: OCI runtime create failed: container_linux.go:380: starting container process caused: process_linux.go:545: container init caused: Running hook #0:: error running hook: signal: segmentation fault, stdout: , stderr:: unknown

Obviously that doesn't tell us very much 😁 However, I have a theory that it may be the result of me using an AMD graphics card. The other day when I was trying to get SD working on my machine, I ran into a lot of problems around making SD work with AMD, so I was wondering if this has been a consideration at all... no problem at all if not, of course!! As I said it was a MASSVIE pain to try to get running when I tried it the other day (in fact, I didn't even get it!)

My graphics card is a Radeon RX 570, if that helps, also, I am running on Windows.

Thanks!!

no src folder in stable-diffusion directory after image creation

Following error poped up after building image.

ln: failed to create symbolic link '/stable-diffusion/src/gfpgan/experiments/pretrained_models/GFPGANv1.3.pth': No such file or directory
(base) root@f6e4124011d3:/stable-diffusion# ls 
 LICENSE                       README.md                           assets    data               latent_diffusion.egg-info   ldm.cmd   models                optimizedSD   scripts    txt2img.yaml   webuildm.cmd
'Launch Waifu Diffusion.lnk'   Stable_Diffusion_v1_Model_Card.md   configs   environment.yaml   ldm                         main.py   notebook_helpers.py   run.cmd       setup.py   webui.cmd

Output is a green image

Specs:

CPU: Ryzen 3600
RAM : 16 GB
GPU: 1660 Super 6GB

  • Running through WSL2 docker backend.
  • CUDA works inside docker (checked with nvidia-smi docker image)
  • Tried lowering resolution to 64x64 , closed all GPU accelerated programs.
  • Tried running with default args
  • Tried with --precision full and --precision full + default args:

Output for --precision full in both cases:

!!Runtime error (txt2img)!!
   expected scalar type Half but found Float
exiting...calling os._exit(0)

Getting an error due to missing checkpoint file

Getting following error after running docker-compose up --build

model_1  | Loading model from /models/model.ckpt
model_1  | Traceback (most recent call last):
model_1  |   File "scripts/webui.py", line 308, in <module>
model_1  |     sd = load_sd_from_config(opt.ckpt)
model_1  |   File "scripts/webui.py", line 140, in load_sd_from_config
model_1  |     pl_sd = torch.load(ckpt, map_location="cpu")
model_1  |   File "/opt/conda/lib/python3.8/site-packages/torch/serialization.py", line 699, in load
model_1  |     with _open_file_like(f, 'rb') as opened_file:
model_1  |   File "/opt/conda/lib/python3.8/site-packages/torch/serialization.py", line 231, in _open_file_like
model_1  |     return _open_file(name_or_buffer, mode)
model_1  |   File "/opt/conda/lib/python3.8/site-packages/torch/serialization.py", line 212, in __init__
model_1  |     super(_open_file, self).__init__(open(name, mode))
model_1  | FileNotFoundError: [Errno 2] No such file or directory: '/models/model.ckpt'

How to include a custom model?

Was working through the scripts, and found that there are explicit mountings in mount.sh for the apps.

I tried adding a custom entry but couldn't get the model to show up in the UI.
Is there something beyond adding a new entry into mount.sh that is required?

Example:

#!/bin/bash

set -e

declare -A MODELS

ROOT=/stable-diffusion/src

MODELS["${ROOT}/gfpgan/experiments/pretrained_models/GFPGANv1.3.pth"]=GFPGANv1.3.pth
MODELS["${ROOT}/realesrgan/experiments/pretrained_models/RealESRGAN_x4plus.pth"]=RealESRGAN_x4plus.pth
MODELS["${ROOT}/realesrgan/experiments/pretrained_models/RealESRGAN_x4plus_anime_6B.pth"]=RealESRGAN_x4plus_anime_6B.pth
MODELS["/latent-diffusion/experiments/pretrained_models/model.ckpt"]=LDSR.ckpt
MODELS["${ROOT}/waifu-diffusion/experiments/pretrained_models/waifu-diffusion.ckpt"]=waifu-diffusion.ckpt # 👈👈👈👈👈👈👈👈👈👈👈👈👈👈

MODELS_DIR=/cache/models

Unnecessary redownload of model_base_caption_capfilt_large.pth

Has this issue been opened before? Check the FAQ, the issues

Describe the bug
Each time new version of auto service has been build old files from cache/torch/hub/checkpoints are being ignored and running img2img interrogate option ends up with redownloading files.

Which UI
I play only with auto, so I can confirm it for auto.

Steps to Reproduce

  1. Go to img2img
  2. Upload image and click Interrogate
  3. Wait for necessary models to download
    load checkpoint from https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_caption_capfilt_large.pth
    21%|████████▏ | 187M/890M [01:34<05:29, 2.24MiB/s]
  4. Ctrl+C or other mean to stop docker compose.
  5. Remove container with docker compose rm
  6. docker compose --profile=auto up # spin up everything once again
  7. Repeat steps from 1 and see file being downloaded again despite having it in cache/torch/hub/checkpoints/ directory which is mounted as volume &v1.

Hardware / Software:
Does not matter

Additional context
N/A

[Suggestion] Add a guide for images filename pattern

Thanks for this amazing web ui! In the settings tab, I see a box for images filename pattern, but I have no idea how to use it. Could you include an explanation of how to use it or a link to a website that does?

Is CPU-Only Usage Possible?

Hello,
I really love what you're doing here. Unfortunately I don't have a system with a GPU to throw at this. Is it possible to do it only in CPU? I saw that there is the vino project for that however it doesn't have nearly the same features of this project.

[Request] Add --allow-code to auto's ui Dockerfile

Thanks for this great use of docker! The --allow-code flag allows custom code to be run in auto's ui, and I don't see a reason why it shouldn't be enabled by default. I just changed the last line of the Dockerfile from
python3 -u ../../webui.py --listen --port 7860 --hide-ui-dir-config --ckpt-dir /cache/custom-models --ckpt /cache/models/model.ckpt ${CLI_ARGS}
to
python3 -u ../../webui.py --allow-code --listen --port 7860 --hide-ui-dir-config --ckpt-dir /cache/custom-models --ckpt /cache/models/model.ckpt ${CLI_ARGS}
and the feature becomes enabled.

It looks like the config file "..." is not a valid JSON file.

I had a previous working build. I get this same error in main folder and the automatic subfolder:

automatic1111-model-1  |     f"It looks like the config file at '{resolved_config_file}' is not a valid JSON file."
automatic1111-model-1  | OSError: It looks like the config file at '/cache/transformers/9c24e6cd9f499d02c4f21a033736dabd365962dc80fe3aeb57a8f85ea45a20a3.26fead7ea4f0f843f6eb4055dfd25693f1a71f3c6871b184042d4b126244e142' is not a valid JSON file.

AMD GPUs

Describe the bug

I have a AMD Radeon RX 6800 XT. Stable diffusion supports this GPU.

After building this image, it fails to run:

 => => naming to docker.io/library/webui-docker-automatic1111                                                                                                                                                0.0s
[+] Running 1/1
 ⠿ Container webui-docker-automatic1111-1  Created                                                                                                                                                           0.2s
Attaching to webui-docker-automatic1111-1
Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]

Steps to Reproduce

  1. Run docker compose --profile auto up --build (after download)

Hardware / Software:

  • OS: Arch Linux (up-to-date)
  • GPU: AMD Radeon RX 6800 XT
  • Version 1.0.1

`hlky` FixFaces is broken (GFPGAN throws error) if `--extra-models-cpu` is given

FixFaces is broken and returns «RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same»

stable-diffusion-webui-docker-master-model-1  | Loaded GFPGAN
stable-diffusion-webui-docker-master-model-1  | Couldn't find metadata on image
stable-diffusion-webui-docker-master-model-1  | Couldn't find metadata on image
stable-diffusion-webui-docker-master-model-1  | Traceback (most recent call last):
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
stable-diffusion-webui-docker-master-model-1  |     output = await app.blocks.process_api(
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 641, in process_api
stable-diffusion-webui-docker-master-model-1  |     predictions, duration = await self.call_function(fn_index, processed_input)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 556, in call_function
stable-diffusion-webui-docker-master-model-1  |     prediction = await anyio.to_thread.run_sync(
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
stable-diffusion-webui-docker-master-model-1  |     return await get_asynclib().run_sync_in_worker_thread(
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
stable-diffusion-webui-docker-master-model-1  |     return await future
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
stable-diffusion-webui-docker-master-model-1  |     result = context.run(func, *args)
stable-diffusion-webui-docker-master-model-1  |   File "scripts/webui.py", line 2027, in imgproc
stable-diffusion-webui-docker-master-model-1  |     image = processGFPGAN(image,imgproc_gfpgan_strength)
stable-diffusion-webui-docker-master-model-1  |   File "scripts/webui.py", line 1732, in processGFPGAN
stable-diffusion-webui-docker-master-model-1  |     cropped_faces, restored_faces, restored_img = GFPGAN.enhance(np.array(image, dtype=np.uint8), has_aligned=False, only_center_face=False, paste_back=True)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context
stable-diffusion-webui-docker-master-model-1  |     return func(*args, **kwargs)
stable-diffusion-webui-docker-master-model-1  |   File "/stable-diffusion/src/gfpgan/gfpgan/utils.py", line 108, in enhance
stable-diffusion-webui-docker-master-model-1  |     self.face_helper.get_face_landmarks_5(only_center_face=only_center_face, eye_dist_threshold=5)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/facexlib/utils/face_restoration_helper.py", line 139, in get_face_landmarks_5
stable-diffusion-webui-docker-master-model-1  |     bboxes = self.face_det.detect_faces(input_img, 0.97) * scale
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/facexlib/detection/retinaface.py", line 205, in detect_faces
stable-diffusion-webui-docker-master-model-1  |     loc, conf, landmarks, priors = self.__detect_faces(image)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/facexlib/detection/retinaface.py", line 156, in __detect_faces
stable-diffusion-webui-docker-master-model-1  |     loc, conf, landmarks = self(inputs)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1110, in _call_impl
stable-diffusion-webui-docker-master-model-1  |     return forward_call(*input, **kwargs)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/facexlib/detection/retinaface.py", line 121, in forward
stable-diffusion-webui-docker-master-model-1  |     out = self.body(inputs)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1110, in _call_impl
stable-diffusion-webui-docker-master-model-1  |     return forward_call(*input, **kwargs)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/torchvision/models/_utils.py", line 63, in forward
stable-diffusion-webui-docker-master-model-1  |     x = module(x)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1110, in _call_impl
stable-diffusion-webui-docker-master-model-1  |     return forward_call(*input, **kwargs)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/conv.py", line 447, in forward
stable-diffusion-webui-docker-master-model-1  |     return self._conv_forward(input, self.weight, self.bias)
stable-diffusion-webui-docker-master-model-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/conv.py", line 443, in _conv_forward
stable-diffusion-webui-docker-master-model-1  |     return F.conv2d(input, weight, bias, self.stride,
stable-diffusion-webui-docker-master-model-1  | RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same

Steps to Reproduce

  1. Generate image
  2. Send to ImageLab
  3. FixFaces

Hardware / Software:

  • OS: Ubuntu 22.04
  • GPU: RTX 3060

GoBig and GoLatent upscaling models fail with IndexError

Hello, and thank you for this amazing docker image!

Upsacling with GoBig or GoLatent fail with the fllowing exception:

Traceback (most recent call last):
  File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
    output = await app.blocks.process_api(
  File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 641, in process_api
    predictions, duration = await self.call_function(fn_index, processed_input)
  File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 556, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
    return await future
  File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
    result = context.run(func, *args)
  File "scripts/webui.py", line 1731, in imgproc
    image = processGoBig(image)
  File "scripts/webui.py", line 1665, in processGoBig
    work_results.append(output_images[0])
IndexError: list index out of range

Steps to Reproduce

  1. Go to 'Image Lab Tab'
  2. Upload 'Single Image'
  3. Select 'GoBig' upscaling model
  4. Press 'Process'

Hardware / Software:

  • OS: Archlinux
  • GPU: RTX3080

Additional context
Any other context about the problem here. If applicable, add screenshots to help explain your problem.

The following packages are not available from current channels:

running this cammand: docker compose --profile auto up --build

#0 7.553 Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
#0 7.554 
#0 7.554 PackagesNotFoundError: The following packages are not available from current channels:
#0 7.554 
#0 7.554   - python=3.8.5
#0 7.554 
#0 7.554 Current channels:
#0 7.554 
#0 7.554   - https://repo.anaconda.com/pkgs/main/linux-aarch64
#0 7.554   - https://repo.anaconda.com/pkgs/main/noarch
#0 7.554   - https://repo.anaconda.com/pkgs/r/linux-aarch64
#0 7.554   - https://repo.anaconda.com/pkgs/r/noarch
#0 7.554 
#0 7.554 To search for alternate channels that may provide the conda package you're
#0 7.554 looking for, navigate to
#0 7.554 
#0 7.554     https://anaconda.org
#0 7.554 
#0 7.554 and use the search bar at the top of the page.

/bin/bash: line 1: /docker/mount.sh: Permission denied on container startup

Building the container works. Starting the container fails with the following message:

stable-diffusion-webui-docker-model-1 | + /docker/mount.sh
stable-diffusion-webui-docker-model-1 | /bin/bash: line 1: /docker/mount.sh: Permission denied

Steps to Reproduce

  1. docker compose build
  2. docker compose up --build
  3. [error...]

Hardware / Software:

  • OS: Windows 11/WSL2
  • Docker version
  • GPU: [Nvidia 3090]
  • Docker version: v20.10.17

Additional context
Attempted to add a chmod +x /docker/mount.sh before the command is run in dockerfile, the command does execute but does not improve things. Docker seems to work otherwise.

CPU only fails because of missing NVIDIA driver

Describe the bug

Right after txt2img finishes I receive the error

Total progress: 100%|██████████| 20/20 [07:53<00:00, 23.66s/it]
webui-docker-automatic1111-cpu-1  | Traceback (most recent call last):
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 273, in run_predict
webui-docker-automatic1111-cpu-1  |     output = await app.blocks.process_api(
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 753, in process_api
webui-docker-automatic1111-cpu-1  |     result = await self.call_function(fn_index, inputs, iterator)
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 630, in call_function
webui-docker-automatic1111-cpu-1  |     prediction = await anyio.to_thread.run_sync(
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
webui-docker-automatic1111-cpu-1  |     return await get_asynclib().run_sync_in_worker_thread(
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
webui-docker-automatic1111-cpu-1  |     return await future
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
webui-docker-automatic1111-cpu-1  |     result = context.run(func, *args)
webui-docker-automatic1111-cpu-1  |   File "/stable-diffusion-webui/modules/ui.py", line 139, in f
webui-docker-automatic1111-cpu-1  |     mem_stats = {k: -(v//-(1024*1024)) for k,v in shared.mem_mon.stop().items()}
webui-docker-automatic1111-cpu-1  |   File "/stable-diffusion-webui/modules/memmon.py", line 77, in stop
webui-docker-automatic1111-cpu-1  |     return self.read()
webui-docker-automatic1111-cpu-1  |   File "/stable-diffusion-webui/modules/memmon.py", line 65, in read
webui-docker-automatic1111-cpu-1  |     free, total = torch.cuda.mem_get_info()
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/cuda/memory.py", line 583, in mem_get_info
webui-docker-automatic1111-cpu-1  |     device = torch.cuda.current_device()
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/cuda/__init__.py", line 481, in current_device
webui-docker-automatic1111-cpu-1  |     _lazy_init()
webui-docker-automatic1111-cpu-1  |   File "/opt/conda/lib/python3.8/site-packages/torch/cuda/__init__.py", line 216, in _lazy_init
webui-docker-automatic1111-cpu-1  |     torch._C._cuda_init()
webui-docker-automatic1111-cpu-1  | RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx

Which UI

auto-cpu

Steps to Reproduce

  1. From a fresh master git clone a few hours ago, I did docker compose --profile download up --build
  2. docker compose --profile auto-cpu up --build
  3. After this finishes, run the GUI at http://localhost:7860/ and without any modifications to settings use cat cartoon handrawing
  4. A few images are generating during the processing which are shown in the WebUI, but at the very end (100%) I get above error.

Hardware / Software:

  • OS: Ubuntu 22.04.1 LTS
  • RAM: 16 GB
  • GPU: NVIDIA Corporation GM107 [GeForce GTX 750 Ti]
  • VRAM: 2GB (thus I'm using auto-cpu instead of gpu)
  • Docker Version 20.10.18, Docker compose version 2.4.1
  • Release version: master (6a66ff6)

Additional context

The WebUI only shows the message Error and in browser console I have Uncaught (in promise) API Error .
The final image is successfully saved to output directory.

@AbdBarho Many thanks for your efforts on this docker setup! You're doing an awesome job

SD upscale limited to 1024x1024?

Hello, firstly thank you for this great webui.

I've been playing about using auto-cpu on an Epyc CPU.

I had reached memory limitations (32gb) when trying to upscale images past 896x896 and found the improvements form Doggettx (basujindal/stable-diffusion#117) and applied this.

The improvements now let me scale images at higher resolutions but when using img2img and the "SD upscale" option anything over 1024x1024 gets ignored (although upon completion it reports the resolution chosen).

Using "Redraw whole image" is fine at higher image sizes until I run out of memory again.

Is there a limit set somewhere or am I using the wrong options somewhere?

Building lstein failed

#0 201.6   gcc -pthread -B /opt/conda/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -DPY_EXTENSION -DHAVE_C99_COMPLEX -Ipywt/_extensions/c -I/opt/conda/include/python3.8 -c pywt/_extensions/c/common.c -o build/temp.linux-x86_64-cpython-38/pywt/_extensions/c/common.o
#0 201.6   error: command 'gcc' failed: No such file or directory
#0 201.6   ----------------------------------------
#0 201.6   ERROR: Failed building wheel for PyWavelets
#0 201.6 ERROR: Could not build wheels for PyWavelets which use PEP 517 and cannot be installed directly
#0 201.6 
#0 201.6 
#0 201.6 CondaEnvException: Pip failed
#0 201.6 
------
failed to solve: executor failed running [/bin/bash -ceuxo pipefail git clone https://github.com/lstein/stable-diffusion.git
cd stable-diffusion
git reset --hard 751283a2de81bee4bb571fbabe4adb19f1d85b97
conda env update --file environment.yaml -n base
conda clean -a -y
]: exit code: 1

Any ideas how to fix it?

Big thanks :)

Your first issue is a big thank you (and less than 24 hrs since you made the repo). 🎉 The webui is a bit of a mission to get up and running on e.g. Linux without the author's Windows scripts. This was an absolute pleasure! 🚀 So thanks for making and publishing.

Docker fails to start with a no such file or directory

This was working with the version pulled on Friday and stopped working this morning

Building the container works. Starting the container fails (again) with the following message:

docker compose --profile download up
[+] Running 1/0
⠿ Container webui-docker-automatic1111-1 Created 0.0s
Attaching to webui-docker-automatic1111-1, webui-docker-download-1, webui-docker-hlky-1
webui-docker-download-1 | Downloading...
webui-docker-download-1 | - GFPGANv1.3.pth exists
webui-docker-download-1 | - RealESRGAN_x4plus.pth exists
webui-docker-download-1 | - LDSR.ckpt exists
webui-docker-download-1 | - LDSR.yaml exists
webui-docker-download-1 | - RealESRGAN_x4plus_anime_6B.pth exists
webui-docker-download-1 | - model.ckpt exists
webui-docker-download-1 | Checking SHAs...
Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error mounting "/run/desktop/mnt/host/wsl/docker-desktop-bind-mounts/Ubuntu/0c85a38b66f627aacc67ce658a895bc6a612c6eba2cb018c2750e2a2332febb0" to rootfs at "/stable-diffusion-webui/config.json": mount /run/desktop/mnt/host/wsl/docker-desktop-bind-mounts/Ubuntu/0c85a38b66f627aacc67ce658a895bc6a612c6eba2cb018c2750e2a2332febb0:/stable-diffusion-webui/config.json (via /proc/self/fd/14), flags: 0x5000: no such file or directory: unknown

Steps to Reproduce

docker compose --profile download up
[error...]

Hardware / Software:

OS: Windows 11/WSL2
Docker version
GPU: [Nvidia 3090]
Docker version: v20.10.17

Build fails with credentials errors when logged in via SSH on WSL2/Windows

Had trouble controlling Docker Compose remotely via SSH on Win/WSL2 while there were no errors with regular Windows desktop session. Seems like logging in via SSH messes up the Docker Desktop's credit store access with WSL somehow where docker compose --profile hlky up --build, docker compose --profile auto restart or just docker login would end with:

Error saving credentials: error storing credentials - err: exit status 1, out: `A specified logon session does not exist. It may already have been terminated.`

Got it to work by forcing plain text auth creds storage by removing line with "credStore" from ~/.docker/config.json (in WSL) and then docker login and docker compose commands worked. Used a read-only access token made on Docker Hub site for safety as a password. Might need to be redone after reboot as config.json file is controlled by Docker Desktop app where some seem to have set that file read-only to keep it from changing with some possible side effects.

There were some open issues on Docker and Docker Compose issue trackers with this workaround but leaving this as a note here, could maybe moved to wiki.

My build issues resolved

update to latest windows 10 release 2021H2 (required for WSL to see the GPU)
update to WSL2 (download and switch to version 2)
install docker (duh, Docker Desktop)
wsl2 dos2unix dockerfile ( these were not so obvious, some characters in the files werent going well)
wsl2 dos2unix ./build/mount.sh

Run on CPU

Hello, how can I run this without GPU support?

I disabled the deploy resources section

    # environment:
    #   - CLI_ARGS=--extra-models-cpu --optimized-turbo
    # deploy:
    #   resources:
    #     reservations:
    #       devices:
    #           - driver: nvidia
    #             device_ids: ['0']
    #             capabilities: [gpu]

But I am still getting an error when trying to run:

stable-diffusion-webui-docker-model-1 | Traceback (most recent call last):
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 453, in
stable-diffusion-webui-docker-model-1 | model, device,config = load_SD_model()
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 444, in load_SD_model
stable-diffusion-webui-docker-model-1 | model = load_model_from_config(config, opt.ckpt)
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 159, in load_model_from_config
stable-diffusion-webui-docker-model-1 | model.cuda()
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/pytorch_lightning/core/mixins/device_dtype_mixin.py", line 127, in cuda
stable-diffusion-webui-docker-model-1 | return super().cuda(device=device)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 688, in cuda
stable-diffusion-webui-docker-model-1 | return self._apply(lambda t: t.cuda(device))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 578, in _apply
stable-diffusion-webui-docker-model-1 | module._apply(fn)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 578, in _apply
stable-diffusion-webui-docker-model-1 | module._apply(fn)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 578, in _apply
stable-diffusion-webui-docker-model-1 | module._apply(fn)
stable-diffusion-webui-docker-model-1 | [Previous line repeated 1 more time]
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 601, in _apply
stable-diffusion-webui-docker-model-1 | param_applied = fn(param)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 688, in
stable-diffusion-webui-docker-model-1 | return self._apply(lambda t: t.cuda(device))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/cuda/init.py", line 216, in _lazy_init
stable-diffusion-webui-docker-model-1 | torch._C._cuda_init()
stable-diffusion-webui-docker-model-1 | RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx
stable-diffusion-webui-docker-model-1 exited with code 1

Thank you

Questions about Docker

I haven't used Docker before, but I like how you've packaged the best repos together to try and switch between. I'm trying to get multiple repos running on my machine, and this could be the ticket. I have some questions:

How do you update the repos? Does docker update them automatically, or do you run a git pull?

Do the repos share package libraries, or are they all in their separate independent environments?

Is there any memory overhead running them with Docker?

Do the output files land in their regular output folders within the repos on the Windows file system? I read somewhere that WSL2 uses a separate Linux file system and doesn't interoperate with the Windows file system.

Thanks!

Compose won't build

Has this issue been opened before? Check issues here and in this one as well

Describe the bug
running docker-compose build causes

Error response from daemon: dockerfile parse error line 33: unknown instruction: GIT
ERROR: Service 'model' failed to build : Build failed

Hardware / Software:

  • Ubuntu 22

Python module error while using Image Lab in hlky interface : ModuleNotFoundError: No module named 'gfpgan.archs.stylegan2_cleanonnx_arch'

Has this issue been opened before? Check the FAQ, the issues and in the issues in the WebUI repo

Nothing related found

Describe the bug
Python module error when trying to use Image Lab tab under hulk interface

ModuleNotFoundError: No module named 'gfpgan.archs.stylegan2_cleanonnx_arch'

Steps to Reproduce

  • Build and run hlky interface
docker compose --profile download up --build
docker compose --profile hlky up --build
  • Browse to Image Lab tab of web interface
  • Upload image to left hand "Single Image" tab
  • Select "Fix Face" & "Upscale" on Processor Selection right hand. tab.
  • Click "Process" button.
  • Error appears in docker compose logs
webui-docker-hlky-1  | Traceback (most recent call last):
webui-docker-hlky-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
webui-docker-hlky-1  |     output = await app.blocks.process_api(
webui-docker-hlky-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 641, in process_api
webui-docker-hlky-1  |     predictions, duration = await self.call_function(fn_index, processed_input)
webui-docker-hlky-1  |   File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 556, in call_function
webui-docker-hlky-1  |     prediction = await anyio.to_thread.run_sync(
webui-docker-hlky-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
webui-docker-hlky-1  |     return await get_asynclib().run_sync_in_worker_thread(
webui-docker-hlky-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
webui-docker-hlky-1  |     return await future
webui-docker-hlky-1  |   File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
webui-docker-hlky-1  |     result = context.run(func, *args)
webui-docker-hlky-1  |   File "scripts/webui.py", line 1838, in imgproc
webui-docker-hlky-1  |     ModelLoader(['GFPGAN'],True,False) # Load used models
webui-docker-hlky-1  |   File "scripts/webui.py", line 1935, in ModelLoader
webui-docker-hlky-1  |     global_vars[m] = load_GFPGAN()
webui-docker-hlky-1  |   File "scripts/webui.py", line 317, in load_GFPGAN
webui-docker-hlky-1  |     from gfpgan import GFPGANer
webui-docker-hlky-1  |   File "/stable-diffusion/src/gfpgan/gfpgan/__init__.py", line 2, in <module>
webui-docker-hlky-1  |     from .archs import *
webui-docker-hlky-1  |   File "/stable-diffusion/src/gfpgan/gfpgan/archs/__init__.py", line 10, in <module>
webui-docker-hlky-1  |     _arch_modules = [importlib.import_module(f'gfpgan.archs.{file_name}') for file_name in arch_filenames]
webui-docker-hlky-1  |   File "/stable-diffusion/src/gfpgan/gfpgan/archs/__init__.py", line 10, in <listcomp>
webui-docker-hlky-1  |     _arch_modules = [importlib.import_module(f'gfpgan.archs.{file_name}') for file_name in arch_filenames]
webui-docker-hlky-1  |   File "/opt/conda/lib/python3.8/importlib/__init__.py", line 127, in import_module
webui-docker-hlky-1  |     return _bootstrap._gcd_import(name[level:], package, level)
webui-docker-hlky-1  |   File "/stable-diffusion/src/gfpgan/gfpgan/archs/gfpganv1_cleanonnx_arch.py", line 8, in <module>
webui-docker-hlky-1  |     from .stylegan2_cleanonnx_arch import StyleGAN2GeneratorCleanONNX
webui-docker-hlky-1  | ModuleNotFoundError: No module named 'gfpgan.archs.stylegan2_cleanonnx_arch'

Hardware / Software:

  • OS: Ubuntu 22.04.1 LTS
  • GPU: RTX 2060
  • Version: Docker 20.10.17, latest pull of AbdBarho/stable-diffusion-webui-docker repo Mon 12th Sep '22

Additional context
Text-to-image prompts work fine.

Web UI container exits with code 137

Describe the bug
Running the web UI does not work. It hangs on making attention of type 'vanilla' with 512 in_channels for a while and then exits with code 137.

Console output:

[+] Running 1/0
 - Container webui-docker-automatic1111-1  Created                                                                 0.0s
Attaching to webui-docker-automatic1111-1
webui-docker-automatic1111-1  | + /docker/mount.sh
webui-docker-automatic1111-1  | Mounted model.ckpt
webui-docker-automatic1111-1  | Mounted GFPGANv1.3.pth
webui-docker-automatic1111-1  | + python3 -u ../../webui.py --listen --port 7860 --hide-ui-dir-config --medvram --opt-split-attention
webui-docker-automatic1111-1  | Loading model [7460a6fa] from /stable-diffusion-webui/repositories/stable-diffusion/models/ldm/stable-diffusion-v1/model.ckpt
webui-docker-automatic1111-1  | Global Step: 470000
webui-docker-automatic1111-1  | LatentDiffusion: Running in eps-prediction mode
webui-docker-automatic1111-1  | DiffusionWrapper has 859.52 M params.
webui-docker-automatic1111-1  | making attention of type 'vanilla' with 512 in_channels
webui-docker-automatic1111-1  | Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
webui-docker-automatic1111-1  | making attention of type 'vanilla' with 512 in_channels
webui-docker-automatic1111-1 exited with code 137

Steps to Reproduce

  1. Run docker compose --profile download up --build
  2. Run docker compose --profile auto up --build
  3. 2nd command builds container but exits with code 137 after attempting to run

Hardware / Software:

  • OS: Windows 11 Pro (Build 22000)
  • Docker version 20.10.17
  • Docker Desktop version 4.12.0
  • GPU: NVIDIA RTX 3090

Image to image generates an error

Has this issue been opened before? Check the FAQ, the issues and in the issues in the WebUI repo

Describe the bug
Img 2 img generates an error.
Traceback (most recent call last):
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
stable-diffusion-webui-docker-model-1 | output = await app.blocks.process_api(
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 639, in process_api
stable-diffusion-webui-docker-model-1 | processed_input = self.preprocess_data(fn_index, raw_input, state)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 543, in preprocess_data
stable-diffusion-webui-docker-model-1 | processed_input.append(block.preprocess(raw_input[i]))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/components.py", line 1546, in preprocess
stable-diffusion-webui-docker-model-1 | x, mask = x["image"], x["mask"]
stable-diffusion-webui-docker-model-1 | TypeError: string indices must be integers

Steps to Reproduce

  1. Go to img 2 img unified and enter a sample image in img2img and a prompt
  2. Click on generate
  3. Look in docker log
  4. See error

Hardware / Software:

  • OS: Windows 10 - WSL2 - Ubuntu
  • GPU: Nvidia 2070 Super
  • Version Built 9/4/2022

Additional context
Any other context about the problem here. If applicable, add screenshots to help explain your problem.

M1 Mac issue - PackagesNotFoundError: The following packages are not available from current channels: pytorch==1.11.0

i checked that this issue has not been reported before.

Describe the bug

docker-compose up produces this:

 => => transferring context: 4.20kB                                        0.0s
 => ERROR [ 2/12] RUN conda install python=3.8.5 && conda clean -a -y      3.0s
------
 > [ 2/12] RUN conda install python=3.8.5 && conda clean -a -y:
#0 0.241 + conda install python=3.8.5
#0 0.430 Collecting package metadata (current_repodata.json): ...working... done
#0 1.315 Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
#0 1.316 Collecting package metadata (repodata.json): ...working... done
#0 2.806 Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
#0 2.807
#0 2.807 PackagesNotFoundError: The following packages are not available from current channels:
#0 2.807
#0 2.807   - python=3.8.5
#0 2.807
#0 2.807 Current channels:
#0 2.807
#0 2.807   - https://repo.anaconda.com/pkgs/main/linux-aarch64
#0 2.807   - https://repo.anaconda.com/pkgs/main/noarch
#0 2.807   - https://repo.anaconda.com/pkgs/r/linux-aarch64
#0 2.807   - https://repo.anaconda.com/pkgs/r/noarch
#0 2.807
#0 2.807 To search for alternate channels that may provide the conda package you're
#0 2.807 looking for, navigate to
#0 2.807
#0 2.807     https://anaconda.org
#0 2.807
#0 2.807 and use the search bar at the top of the page.
#0 2.807
#0 2.807
------
failed to solve: executor failed running [/bin/bash -ceuxo pipefail conda install python=3.8.5 && conda clean -a -y]: exit code: 1

Steps to Reproduce

image

Hardware / Software:

  • OS: macos 12.5.1
  • GPU: m1 max, macbook pro

Custom WebUI SHA

It looks like you set the webui repo to a specific stable version.

When I go look at that project it is frequently being updated.

How can I easily update my container? Do I need to rebuild it from main each time?

be a nice option, thanks

No file or directory for latent-diffusion

image

stable-diffusion-webui-docker-model-1  | /docker/mount.sh: line 25: /latentdiffusion/experiments/pretrained_models/project.yaml: No such file or directory 
stable-diffusion-webui-docker-model-1 exited with code 1

EDIT: Fixed the issue in the hlky mount.sh file by removing the # from the project.yaml line but now facing this issue

image

stable-diffusion-webui-docker-model-1  |  Error loading LDSR
stable-diffusion-webui-docker-model-1  |  Loading model from /models/model.ckpt
stable-diffusion-webui-docker-model-1  |  Traceback (most recent call last):
stable-diffusion-webui-docker-model-1  |  File "scripts/webui.py", line 382, in try_loading_LDSR
stable-diffusion-webui-docker-model-1  |  LDSR = load_LDSR(checking=True) # TODO: Should try to load both models before giving up
stable-diffusion-webui-docker-model-1  |  File "scripts/webui.py", line 297, in load_LDSR
stable-diffusion-webui-docker-model-1  |  raise Exception("LDSR model not found at path "+model_path)
stable-diffusion-webui-docker-model-1  |  Exception: LDSR model not found at path /latent diffusion/experiments/pretrained_models/model.ckpt  

stable-diffusion-webui-docker-model-1  | Global Step: 470000
stable-diffusion-webui-docker-model-1  | UNet: Running in eps-prediction mode
stable-diffusion-webui-docker-model-1 exited with code 137

Proxmox / LXC: Unable to start docker container, nvidia-container-cli: initialization error: load library failed: libnvidia-ml.so.1

Hi,

When I docker compose up -d

I get the following

root@stablediffusion:/opt/stable-diffusion-webui-docker# docker compose up -d
[+] Running 0/1
 ⠹ Container stable-diffusion-webui-docker-model-1  Starting                                                                                                            68.2s
Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: initialization error: load library failed: libnvidia-ml.so.1: cannot open shared object file: no such file or directory: unknown

My setup is as follows, I have a x64 computer running latest proxmox
On it I have created an LXC container with Ubuntu 22.04.1 LTS
This container is updated upgraded, docker-compose initially installed

The initial install failed due to old version 20.10.12 was replaced with 20.10.18

Installed with the following commands

apt remove docker-compose
curl -fsSL https://get.docker.com -o get-docker.sh
sh get-docker.sh

This didn't work, so I installed nvidia-docker2 as follows

distribution=$(. /etc/os-release;echo $ID$VERSION_ID)    && curl -s -L https://nvidia.github.io/nvidia-docker/gpgkey | sudo apt-key add -    && curl -s -L https://nvidia.github.io/nvidia-docker/$distribution/nvidia-docker.list | sudo tee /etc/apt/sources.list.d/nvidia-docker.list
apt-get update
apt-get install -y nvidia-docker2

Here are my model files

root@stablediffusion:/opt/stable-diffusion-webui-docker/models# sha1sum *
e5102721808b8f0e421e6c3a8b73248d3b08373f GFPGANv1.3.pth
60482fe909787328a7bf5a7edffb96e390465915 LDSR.ckpt
f7f8941c3117649042047220b762fe3d8b5f616c LDSR.yaml
a1b5c65842f69da7c857479b501fb7125022fabf RealESRGAN_x4plus.pth
d91e1688357e2d2238911253c57a94aeac4625a4 RealESRGAN_x4plus_anime_6B.pth
210783247af4f65a3d23d026490cc37a670964dd model.ckpt

While writing this issue, I have found one thing I have not yet tried, and I will after posting this.
Hopefully anyone walking in my footsteps will find this and not waste so much time !

apt-add-repository ppa:graphics-drivers/ppa
apt-get install nvidia-370 nvidia-prime

Maybe that will work, the version 370 is probably very old, looks like what we need here is "nvidia-smi" and that may be part of the drivers. I suspect this is going to install a whole X server ...

Also this LXC container is unprivileged so I suspect I will also run into some other problems later down the road and I don't know if I have the IOMMU stuff working at all, this is my first time using the graphics card in this computer (5 year old 1080ti)


ok installed the nvidia drivers with the following

apt install software-properties-common
apt-add-repository ppa:graphics-drivers/ppa
apt install -y nvidia-prime nvidia-driver-515-server

The nvidia repository recommends running
apt-get install phoronix-test-suite
phoronix-test-suite default-benchmark openarena xonotic tesseract gputest unigine-valley
However this package is nowhere to be found ( Unable to locate package phoronix-test-suite )

I tried installing only nvidia-prime, it changed nothing, I don't know if this package is required at all

root@stablediffusion:/opt/stable-diffusion-webui-docker# docker compose up -d
[+] Running 0/1
 ⠦ Container stable-diffusion-webui-docker-model-1  Starting                                                                                                             1.6s
Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: initialization error: nvml error: driver not loaded: unknown

Ok, well, on to the next error !

WebUI won't recover when running out of VRAM

Not sure where to ask and whether it's a known issue but the Web UI's don't seem to recover when increasing resolution enough to run out of VRAM. Having to interrupt Docker with ctrl+c and restart it currently to recover from out of memory situation with the RuntimeError: CUDA error: an illegal memory access was encountered error. Same behavior with the earlier default UI a couple days ago, and the 'auto' UI-profile of the latest 1.0.0 release on WSL2/Docker Desktop/RTX3060.

Good job with this project, very convenient way to get experimenting with SD web UI's, thank you.

Question: can I start a rendering from the command line?

Hi,

I've been trying a number of stable-diffusion installation methods that should work without GPU, but this one is the only one that I got working (out of the box!) without any breaking issue.

That's very good, but problem is that I was actually looking for an installation that would allow renderings to be run from the command line (scripted)...

So my question (which is definitely not a bug, as the repo says it's about offering a GUI): is it still possible somehow to render images without using the GUI?

Best regards,
Vic

Failure to send image to lab

This seems to have been functioningearlier in the week week but no longer. The expected behavior is to bring the image up in the image lab tab.

Describe the bug

stable-diffusion-webui-docker-model-1 | Traceback (most recent call last):
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
stable-diffusion-webui-docker-model-1 | output = await app.blocks.process_api(
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 639, in process_api
stable-diffusion-webui-docker-model-1 | processed_input = self.preprocess_data(fn_index, raw_input, state)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 543, in preprocess_data
stable-diffusion-webui-docker-model-1 | processed_input.append(block.preprocess(raw_input[i]))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/components.py", line 1270, in preprocess
stable-diffusion-webui-docker-model-1 | return self.choices.index(x)
stable-diffusion-webui-docker-model-1 | ValueError: [['frankfurt in the rain moody lighting\n', None], ['seed:', '4146656096'], [' ', None], ['width:', '768'], [' ', None], ['height:', '384'], [' ', None], ['steps:', '50'], [' ', None], ['cfg_scale:', '7.5'], [' ', None], ['sampler:', 'k_lms'], ['', None]] is not in list

Steps to Reproduce

  1. Change size to 384x768
  2. Prompt string: frankfurt in the rain moody lighting
  3. Click on Generate
  4. Wait for images to be generated (*4)
  5. Select one image
  6. Click on "Send to Lab"
  7. Attempt to view in image lab. Nothing there.

Hardware / Software:
OS: Windows 11/WSL2
Docker version: v20.10.17
GPU: Nvidia 3090
Docker version: v20.10.17

Additional context

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.