[日本語版 README はこちら](README_ja.md) # Stable Diffusion CLI on Modal This is a Diffusers-based script for running Stable Diffusion on [Modal](https://modal.com/). This script has no WebUI and only works with CLI. It can perform txt2img inference and has the ability to increase resolution using ControlNet Tile and Upscaler. ## Features 1. Image generation using txt2img ![](assets/20230902_tile_imgs.png) 2. Upscaling | Before upscaling | After upscaling | | ---------------------------------------------------------------- | ---------------------------------------------------------------- | | | | ## Requirements The app requires the following to run: - python: > 3.10 - modal-client - A token for Modal. The `modal-client` is the Python library. In order to install that: ```bash pip install modal-client ``` And you need a modal token to use this script: ```bash modal token new ``` Please see [the documentation of Modal](https://modal.com/docs/guide) for modals and tokens. ## Getting Started To use the script, execute the below. 1. git clone the repository. 2. Copy `./setup_files/config.sample.yml` to `./setup_files/config.yml` 3. Open the Makefile and set prompts. 4. Execute `make deploy` command. An application will be deployed to Modal. 5. Execute `make run` command. Images are generated and output to the `outputs/` directory. ## Directory structure ```txt . ├── .env # Secrets manager ├── Makefile ├── README.md ├── sdcli/ # A directory with scripts to run inference. │   ├── outputs/ # Images are outputted this directory. │   ├── txt2img.py # A script to run txt2img inference. │   └── util.py └── setup_files/ # A directory with config files. ├── __main__.py # A main script to run inference. ├── Dockerfile # To build a base image. ├── config.yml # To set a model, vae and some tools. ├── requirements.txt ├── setup.py # Build an application to deploy on Modal. └── txt2img.py # There is a class to run inference. ``` ## How to use ### 1. `git clone` the repository ```bash git clone https://github.com/hodanov/stable-diffusion-modal.git cd stable-diffusion-modal ``` ### 2. Add hugging_face_token to .env file Hugging Add hugging_face_token to .env file. This script downloads and uses a model from HuggingFace, but if you want to use a model in a private repository, you will need to set this environment variable. ```txt HUGGING_FACE_TOKEN="Write your hugging face token here." ``` ### 3. Add the model to ./setup_files/config.yml Add the model used for inference. Use the Safetensors file as is. VAE, LoRA, and Textual Inversion are also configurable. ```yml # ex) model: name: stable-diffusion-1-5 url: https://huggingface.co/runwayml/stable-diffusion-v1-5/blob/main/v1-5-pruned.safetensors # Specify URL for the safetensor file. vae: name: sd-vae-ft-mse url: https://huggingface.co/stabilityai/sd-vae-ft-mse-original/blob/main/vae-ft-mse-840000-ema-pruned.safetensors controlnets: - name: control_v11f1e_sd15_tile repo_id: lllyasviel/control_v11f1e_sd15_tile ``` If you want to use LoRA and Textual Inversion, configure as follows. ```yml # Example loras: - name: lora_name.safetensors # Specify the LoRA file name. Any name is fine, but the extension `.safetensors` is required. url: download_link_here # Specify the download link for the safetensor file. ``` ### 4. Setting prompts Set the prompt to Makefile. ```makefile # ex) run: cd ./sdcli && modal run txt2img.py \ --prompt "hogehoge" \ --n-prompt "mogumogu" \ --height 768 \ --width 512 \ --samples 1 \ --steps 30 \ --seed 12321 | --upscaler "RealESRGAN_x2plus" \ --use-face-enhancer "False" \ --fix-by-controlnet-tile "True" ``` ### 5. make deploy Execute the below command. An application will be deployed on Modal. ```bash make deploy ``` ### 6. make run The txt2img inference is executed with the following command. ```bash make run ``` Thank you.