docs: cleanup of runway model (#12503)
* cleanup of runway model * quality fixes
This commit is contained in:
parent
af769881d3
commit
1b456bd5d5
@ -171,7 +171,7 @@ Also, say 👋 in our public Discord channel <a href="https://discord.gg/G7tWnz9
|
|||||||
<tr style="border-top: 2px solid black">
|
<tr style="border-top: 2px solid black">
|
||||||
<td>Text-guided Image Inpainting</td>
|
<td>Text-guided Image Inpainting</td>
|
||||||
<td><a href="https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion/inpaint">Stable Diffusion Inpainting</a></td>
|
<td><a href="https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion/inpaint">Stable Diffusion Inpainting</a></td>
|
||||||
<td><a href="https://huggingface.co/runwayml/stable-diffusion-inpainting"> runwayml/stable-diffusion-inpainting </a></td>
|
<td><a href="https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting"> stable-diffusion-v1-5/stable-diffusion-inpainting </a></td>
|
||||||
</tr>
|
</tr>
|
||||||
<tr style="border-top: 2px solid black">
|
<tr style="border-top: 2px solid black">
|
||||||
<td>Image Variation</td>
|
<td>Image Variation</td>
|
||||||
|
|||||||
@ -39,7 +39,7 @@ mask_url = "https://huggingface.co/datasets/hf-internal-testing/diffusers-images
|
|||||||
original_image = load_image(img_url).resize((512, 512))
|
original_image = load_image(img_url).resize((512, 512))
|
||||||
mask_image = load_image(mask_url).resize((512, 512))
|
mask_image = load_image(mask_url).resize((512, 512))
|
||||||
|
|
||||||
pipe = StableDiffusionInpaintPipeline.from_pretrained("runwayml/stable-diffusion-inpainting")
|
pipe = StableDiffusionInpaintPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-inpainting")
|
||||||
pipe.vae = AsymmetricAutoencoderKL.from_pretrained("cross-attention/asymmetric-autoencoder-kl-x-1-5")
|
pipe.vae = AsymmetricAutoencoderKL.from_pretrained("cross-attention/asymmetric-autoencoder-kl-x-1-5")
|
||||||
pipe.to("cuda")
|
pipe.to("cuda")
|
||||||
|
|
||||||
|
|||||||
@ -21,7 +21,7 @@ The Stable Diffusion model can also infer depth based on an image using [MiDaS](
|
|||||||
> [!TIP]
|
> [!TIP]
|
||||||
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
||||||
>
|
>
|
||||||
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis), [Runway](https://huggingface.co/runwayml), and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis) and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
||||||
|
|
||||||
## StableDiffusionDepth2ImgPipeline
|
## StableDiffusionDepth2ImgPipeline
|
||||||
|
|
||||||
|
|||||||
@ -21,14 +21,14 @@ The Stable Diffusion model can also be applied to inpainting which lets you edit
|
|||||||
## Tips
|
## Tips
|
||||||
|
|
||||||
It is recommended to use this pipeline with checkpoints that have been specifically fine-tuned for inpainting, such
|
It is recommended to use this pipeline with checkpoints that have been specifically fine-tuned for inpainting, such
|
||||||
as [runwayml/stable-diffusion-inpainting](https://huggingface.co/runwayml/stable-diffusion-inpainting). Default
|
as [stable-diffusion-v1-5/stable-diffusion-inpainting](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting). Default
|
||||||
text-to-image Stable Diffusion checkpoints, such as
|
text-to-image Stable Diffusion checkpoints, such as
|
||||||
[stable-diffusion-v1-5/stable-diffusion-v1-5](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) are also compatible but they might be less performant.
|
[stable-diffusion-v1-5/stable-diffusion-v1-5](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) are also compatible but they might be less performant.
|
||||||
|
|
||||||
> [!TIP]
|
> [!TIP]
|
||||||
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
||||||
>
|
>
|
||||||
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis), [Runway](https://huggingface.co/runwayml), and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis) and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
||||||
|
|
||||||
## StableDiffusionInpaintPipeline
|
## StableDiffusionInpaintPipeline
|
||||||
|
|
||||||
|
|||||||
@ -17,7 +17,7 @@ The Stable Diffusion latent upscaler model was created by [Katherine Crowson](ht
|
|||||||
> [!TIP]
|
> [!TIP]
|
||||||
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
||||||
>
|
>
|
||||||
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis), [Runway](https://huggingface.co/runwayml), and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis) and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
||||||
|
|
||||||
## StableDiffusionLatentUpscalePipeline
|
## StableDiffusionLatentUpscalePipeline
|
||||||
|
|
||||||
|
|||||||
@ -22,7 +22,7 @@ Stable Diffusion is trained on 512x512 images from a subset of the LAION-5B data
|
|||||||
|
|
||||||
For more details about how Stable Diffusion works and how it differs from the base latent diffusion model, take a look at the Stability AI [announcement](https://stability.ai/blog/stable-diffusion-announcement) and our own [blog post](https://huggingface.co/blog/stable_diffusion#how-does-stable-diffusion-work) for more technical details.
|
For more details about how Stable Diffusion works and how it differs from the base latent diffusion model, take a look at the Stability AI [announcement](https://stability.ai/blog/stable-diffusion-announcement) and our own [blog post](https://huggingface.co/blog/stable_diffusion#how-does-stable-diffusion-work) for more technical details.
|
||||||
|
|
||||||
You can find the original codebase for Stable Diffusion v1.0 at [CompVis/stable-diffusion](https://github.com/CompVis/stable-diffusion) and Stable Diffusion v2.0 at [Stability-AI/stablediffusion](https://github.com/Stability-AI/stablediffusion) as well as their original scripts for various tasks. Additional official checkpoints for the different Stable Diffusion versions and tasks can be found on the [CompVis](https://huggingface.co/CompVis), [Runway](https://huggingface.co/runwayml), and [Stability AI](https://huggingface.co/stabilityai) Hub organizations. Explore these organizations to find the best checkpoint for your use-case!
|
You can find the original codebase for Stable Diffusion v1.0 at [CompVis/stable-diffusion](https://github.com/CompVis/stable-diffusion) and Stable Diffusion v2.0 at [Stability-AI/stablediffusion](https://github.com/Stability-AI/stablediffusion) as well as their original scripts for various tasks. Additional official checkpoints for the different Stable Diffusion versions and tasks can be found on the [CompVis](https://huggingface.co/CompVis) and [Stability AI](https://huggingface.co/stabilityai) Hub organizations. Explore these organizations to find the best checkpoint for your use-case!
|
||||||
|
|
||||||
The table below summarizes the available Stable Diffusion pipelines, their supported tasks, and an interactive demo:
|
The table below summarizes the available Stable Diffusion pipelines, their supported tasks, and an interactive demo:
|
||||||
|
|
||||||
@ -64,7 +64,7 @@ The table below summarizes the available Stable Diffusion pipelines, their suppo
|
|||||||
<a href="./inpaint">StableDiffusionInpaint</a>
|
<a href="./inpaint">StableDiffusionInpaint</a>
|
||||||
</td>
|
</td>
|
||||||
<td class="px-4 py-2 text-gray-700">inpainting</td>
|
<td class="px-4 py-2 text-gray-700">inpainting</td>
|
||||||
<td class="px-4 py-2"><a href="https://huggingface.co/spaces/runwayml/stable-diffusion-inpainting"><img src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Spaces-blue"/></a>
|
<td class="px-4 py-2"><a href="https://huggingface.co/spaces/stable-diffusion-v1-5/stable-diffusion-inpainting"><img src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Spaces-blue"/></a>
|
||||||
</td>
|
</td>
|
||||||
</tr>
|
</tr>
|
||||||
<tr>
|
<tr>
|
||||||
|
|||||||
@ -36,7 +36,7 @@ Here are some examples for how to use Stable Diffusion 2 for each task:
|
|||||||
> [!TIP]
|
> [!TIP]
|
||||||
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
||||||
>
|
>
|
||||||
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis), [Runway](https://huggingface.co/runwayml), and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis) and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
||||||
|
|
||||||
## Text-to-image
|
## Text-to-image
|
||||||
|
|
||||||
|
|||||||
@ -25,7 +25,7 @@ The abstract from the paper is:
|
|||||||
> [!TIP]
|
> [!TIP]
|
||||||
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
||||||
>
|
>
|
||||||
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis), [Runway](https://huggingface.co/runwayml), and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis) and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
||||||
|
|
||||||
## StableDiffusionPipeline
|
## StableDiffusionPipeline
|
||||||
|
|
||||||
|
|||||||
@ -21,7 +21,7 @@ The Stable Diffusion upscaler diffusion model was created by the researchers and
|
|||||||
> [!TIP]
|
> [!TIP]
|
||||||
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
> Make sure to check out the Stable Diffusion [Tips](overview#tips) section to learn how to explore the tradeoff between scheduler speed and quality, and how to reuse pipeline components efficiently!
|
||||||
>
|
>
|
||||||
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis), [Runway](https://huggingface.co/runwayml), and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
> If you're interested in using one of the official checkpoints for a task, explore the [CompVis](https://huggingface.co/CompVis) and [Stability AI](https://huggingface.co/stabilityai) Hub organizations!
|
||||||
|
|
||||||
## StableDiffusionUpscalePipeline
|
## StableDiffusionUpscalePipeline
|
||||||
|
|
||||||
|
|||||||
@ -16,12 +16,12 @@ pipeline.unet.config["in_channels"]
|
|||||||
4
|
4
|
||||||
```
|
```
|
||||||
|
|
||||||
Inpainting requires 9 channels in the input sample. You can check this value in a pretrained inpainting model like [`runwayml/stable-diffusion-inpainting`](https://huggingface.co/runwayml/stable-diffusion-inpainting):
|
Inpainting requires 9 channels in the input sample. You can check this value in a pretrained inpainting model like [`stable-diffusion-v1-5/stable-diffusion-inpainting`](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting):
|
||||||
|
|
||||||
```py
|
```py
|
||||||
from diffusers import StableDiffusionPipeline
|
from diffusers import StableDiffusionPipeline
|
||||||
|
|
||||||
pipeline = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-inpainting", use_safetensors=True)
|
pipeline = StableDiffusionPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-inpainting", use_safetensors=True)
|
||||||
pipeline.unet.config["in_channels"]
|
pipeline.unet.config["in_channels"]
|
||||||
9
|
9
|
||||||
```
|
```
|
||||||
|
|||||||
@ -215,7 +215,7 @@ from diffusers import AutoPipelineForInpainting, LCMScheduler
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipe = AutoPipelineForInpainting.from_pretrained(
|
pipe = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting",
|
"stable-diffusion-v1-5/stable-diffusion-inpainting",
|
||||||
torch_dtype=torch.float16,
|
torch_dtype=torch.float16,
|
||||||
variant="fp16",
|
variant="fp16",
|
||||||
).to("cuda")
|
).to("cuda")
|
||||||
|
|||||||
@ -112,7 +112,7 @@ blurred_mask
|
|||||||
|
|
||||||
## Popular models
|
## Popular models
|
||||||
|
|
||||||
[Stable Diffusion Inpainting](https://huggingface.co/runwayml/stable-diffusion-inpainting), [Stable Diffusion XL (SDXL) Inpainting](https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1), and [Kandinsky 2.2 Inpainting](https://huggingface.co/kandinsky-community/kandinsky-2-2-decoder-inpaint) are among the most popular models for inpainting. SDXL typically produces higher resolution images than Stable Diffusion v1.5, and Kandinsky 2.2 is also capable of generating high-quality images.
|
[Stable Diffusion Inpainting](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting), [Stable Diffusion XL (SDXL) Inpainting](https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1), and [Kandinsky 2.2 Inpainting](https://huggingface.co/kandinsky-community/kandinsky-2-2-decoder-inpaint) are among the most popular models for inpainting. SDXL typically produces higher resolution images than Stable Diffusion v1.5, and Kandinsky 2.2 is also capable of generating high-quality images.
|
||||||
|
|
||||||
### Stable Diffusion Inpainting
|
### Stable Diffusion Inpainting
|
||||||
|
|
||||||
@ -124,7 +124,7 @@ from diffusers import AutoPipelineForInpainting
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -244,7 +244,7 @@ make_image_grid([init_image, image], rows=1, cols=2)
|
|||||||
```
|
```
|
||||||
|
|
||||||
</hfoption>
|
</hfoption>
|
||||||
<hfoption id="runwayml/stable-diffusion-inpainting">
|
<hfoption id="stable-diffusion-v1-5/stable-diffusion-inpainting">
|
||||||
|
|
||||||
```py
|
```py
|
||||||
import torch
|
import torch
|
||||||
@ -252,7 +252,7 @@ from diffusers import AutoPipelineForInpainting
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -278,7 +278,7 @@ make_image_grid([init_image, image], rows=1, cols=2)
|
|||||||
</div>
|
</div>
|
||||||
<div>
|
<div>
|
||||||
<img class="rounded-xl" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/inpaint-specific.png"/>
|
<img class="rounded-xl" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/inpaint-specific.png"/>
|
||||||
<figcaption class="mt-2 text-center text-sm text-gray-500">runwayml/stable-diffusion-inpainting</figcaption>
|
<figcaption class="mt-2 text-center text-sm text-gray-500">stable-diffusion-v1-5/stable-diffusion-inpainting</figcaption>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
@ -308,7 +308,7 @@ make_image_grid([init_image, image], rows=1, cols=2)
|
|||||||
```
|
```
|
||||||
|
|
||||||
</hfoption>
|
</hfoption>
|
||||||
<hfoption id="runwayml/stable-diffusion-inpaint">
|
<hfoption id="stable-diffusion-v1-5/stable-diffusion-inpaint">
|
||||||
|
|
||||||
```py
|
```py
|
||||||
import torch
|
import torch
|
||||||
@ -316,7 +316,7 @@ from diffusers import AutoPipelineForInpainting
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -340,7 +340,7 @@ make_image_grid([init_image, image], rows=1, cols=2)
|
|||||||
</div>
|
</div>
|
||||||
<div>
|
<div>
|
||||||
<img class="rounded-xl" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/specific-inpaint-basic.png"/>
|
<img class="rounded-xl" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/specific-inpaint-basic.png"/>
|
||||||
<figcaption class="mt-2 text-center text-sm text-gray-500">runwayml/stable-diffusion-inpainting</figcaption>
|
<figcaption class="mt-2 text-center text-sm text-gray-500">stable-diffusion-v1-5/stable-diffusion-inpainting</figcaption>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
@ -358,7 +358,7 @@ from diffusers.utils import load_image, make_image_grid
|
|||||||
|
|
||||||
device = "cuda"
|
device = "cuda"
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting",
|
"stable-diffusion-v1-5/stable-diffusion-inpainting",
|
||||||
torch_dtype=torch.float16,
|
torch_dtype=torch.float16,
|
||||||
variant="fp16"
|
variant="fp16"
|
||||||
)
|
)
|
||||||
@ -396,7 +396,7 @@ from diffusers import AutoPipelineForInpainting
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -441,7 +441,7 @@ from diffusers import AutoPipelineForInpainting
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -481,7 +481,7 @@ from diffusers import AutoPipelineForInpainting
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -606,7 +606,7 @@ from diffusers import AutoPipelineForInpainting, AutoPipelineForImage2Image
|
|||||||
from diffusers.utils import load_image, make_image_grid
|
from diffusers.utils import load_image, make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -683,7 +683,7 @@ from diffusers import AutoPipelineForInpainting
|
|||||||
from diffusers.utils import make_image_grid
|
from diffusers.utils import make_image_grid
|
||||||
|
|
||||||
pipeline = AutoPipelineForInpainting.from_pretrained(
|
pipeline = AutoPipelineForInpainting.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16,
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16,
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
@ -714,7 +714,7 @@ controlnet = ControlNetModel.from_pretrained("lllyasviel/control_v11p_sd15_inpai
|
|||||||
|
|
||||||
# pass ControlNet to the pipeline
|
# pass ControlNet to the pipeline
|
||||||
pipeline = StableDiffusionControlNetInpaintPipeline.from_pretrained(
|
pipeline = StableDiffusionControlNetInpaintPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", controlnet=controlnet, torch_dtype=torch.float16, variant="fp16"
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", controlnet=controlnet, torch_dtype=torch.float16, variant="fp16"
|
||||||
)
|
)
|
||||||
pipeline.enable_model_cpu_offload()
|
pipeline.enable_model_cpu_offload()
|
||||||
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
# remove following line if xFormers is not installed or you have PyTorch 2.0 or higher installed
|
||||||
|
|||||||
@ -173,7 +173,7 @@ mask_url = "https://raw.githubusercontent.com/CompVis/latent-diffusion/main/data
|
|||||||
init_image = download_image(img_url).resize((512, 512))
|
init_image = download_image(img_url).resize((512, 512))
|
||||||
mask_image = download_image(mask_url).resize((512, 512))
|
mask_image = download_image(mask_url).resize((512, 512))
|
||||||
|
|
||||||
path = "runwayml/stable-diffusion-inpainting"
|
path = "stable-diffusion-v1-5/stable-diffusion-inpainting"
|
||||||
|
|
||||||
run_compile = True # Set True / False
|
run_compile = True # Set True / False
|
||||||
|
|
||||||
|
|||||||
@ -28,12 +28,12 @@ pipeline.unet.config["in_channels"]
|
|||||||
4
|
4
|
||||||
```
|
```
|
||||||
|
|
||||||
인페인팅은 입력 샘플에 9개의 채널이 필요합니다. [`runwayml/stable-diffusion-inpainting`](https://huggingface.co/runwayml/stable-diffusion-inpainting)와 같은 사전학습된 인페인팅 모델에서 이 값을 확인할 수 있습니다:
|
인페인팅은 입력 샘플에 9개의 채널이 필요합니다. [`stable-diffusion-v1-5/stable-diffusion-inpainting`](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting)와 같은 사전학습된 인페인팅 모델에서 이 값을 확인할 수 있습니다:
|
||||||
|
|
||||||
```py
|
```py
|
||||||
from diffusers import StableDiffusionPipeline
|
from diffusers import StableDiffusionPipeline
|
||||||
|
|
||||||
pipeline = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-inpainting")
|
pipeline = StableDiffusionPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-inpainting")
|
||||||
pipeline.unet.config["in_channels"]
|
pipeline.unet.config["in_channels"]
|
||||||
9
|
9
|
||||||
```
|
```
|
||||||
|
|||||||
@ -14,7 +14,7 @@ specific language governing permissions and limitations under the License.
|
|||||||
|
|
||||||
[[open-in-colab]]
|
[[open-in-colab]]
|
||||||
|
|
||||||
[`StableDiffusionInpaintPipeline`]은 마스크와 텍스트 프롬프트를 제공하여 이미지의 특정 부분을 편집할 수 있도록 합니다. 이 기능은 인페인팅 작업을 위해 특별히 훈련된 [`runwayml/stable-diffusion-inpainting`](https://huggingface.co/runwayml/stable-diffusion-inpainting)과 같은 Stable Diffusion 버전을 사용합니다.
|
[`StableDiffusionInpaintPipeline`]은 마스크와 텍스트 프롬프트를 제공하여 이미지의 특정 부분을 편집할 수 있도록 합니다. 이 기능은 인페인팅 작업을 위해 특별히 훈련된 [`stable-diffusion-v1-5/stable-diffusion-inpainting`](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting)과 같은 Stable Diffusion 버전을 사용합니다.
|
||||||
|
|
||||||
먼저 [`StableDiffusionInpaintPipeline`] 인스턴스를 불러옵니다:
|
먼저 [`StableDiffusionInpaintPipeline`] 인스턴스를 불러옵니다:
|
||||||
|
|
||||||
@ -27,7 +27,7 @@ from io import BytesIO
|
|||||||
from diffusers import StableDiffusionInpaintPipeline
|
from diffusers import StableDiffusionInpaintPipeline
|
||||||
|
|
||||||
pipeline = StableDiffusionInpaintPipeline.from_pretrained(
|
pipeline = StableDiffusionInpaintPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting",
|
"stable-diffusion-v1-5/stable-diffusion-inpainting",
|
||||||
torch_dtype=torch.float16,
|
torch_dtype=torch.float16,
|
||||||
)
|
)
|
||||||
pipeline = pipeline.to("cuda")
|
pipeline = pipeline.to("cuda")
|
||||||
@ -61,12 +61,3 @@ image = pipe(prompt=prompt, image=init_image, mask_image=mask_image).images[0]
|
|||||||
|
|
||||||
> [!WARNING]
|
> [!WARNING]
|
||||||
> 이전의 실험적인 인페인팅 구현에서는 품질이 낮은 다른 프로세스를 사용했습니다. 이전 버전과의 호환성을 보장하기 위해 새 모델이 포함되지 않은 사전학습된 파이프라인을 불러오면 이전 인페인팅 방법이 계속 적용됩니다.
|
> 이전의 실험적인 인페인팅 구현에서는 품질이 낮은 다른 프로세스를 사용했습니다. 이전 버전과의 호환성을 보장하기 위해 새 모델이 포함되지 않은 사전학습된 파이프라인을 불러오면 이전 인페인팅 방법이 계속 적용됩니다.
|
||||||
|
|
||||||
아래 Space에서 이미지 인페인팅을 직접 해보세요!
|
|
||||||
|
|
||||||
<iframe
|
|
||||||
src="https://runwayml-stable-diffusion-inpainting.hf.space"
|
|
||||||
frameborder="0"
|
|
||||||
width="850"
|
|
||||||
height="500"
|
|
||||||
></iframe>
|
|
||||||
|
|||||||
@ -16,12 +16,12 @@ pipeline.unet.config["in_channels"]
|
|||||||
4
|
4
|
||||||
```
|
```
|
||||||
|
|
||||||
而图像修复任务需要输入样本具有9个通道。您可以在 [`runwayml/stable-diffusion-inpainting`](https://huggingface.co/runwayml/stable-diffusion-inpainting) 这样的预训练修复模型中验证此参数:
|
而图像修复任务需要输入样本具有9个通道。您可以在 [`stable-diffusion-v1-5/stable-diffusion-inpainting`](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting) 这样的预训练修复模型中验证此参数:
|
||||||
|
|
||||||
```python
|
```python
|
||||||
from diffusers import StableDiffusionPipeline
|
from diffusers import StableDiffusionPipeline
|
||||||
|
|
||||||
pipeline = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-inpainting", use_safetensors=True)
|
pipeline = StableDiffusionPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-inpainting", use_safetensors=True)
|
||||||
pipeline.unet.config["in_channels"]
|
pipeline.unet.config["in_channels"]
|
||||||
9
|
9
|
||||||
```
|
```
|
||||||
|
|||||||
@ -1328,7 +1328,7 @@ model = CLIPSegForImageSegmentation.from_pretrained("CIDAS/clipseg-rd64-refined"
|
|||||||
|
|
||||||
# Load Stable Diffusion Inpainting Pipeline with custom pipeline
|
# Load Stable Diffusion Inpainting Pipeline with custom pipeline
|
||||||
pipe = DiffusionPipeline.from_pretrained(
|
pipe = DiffusionPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting",
|
"stable-diffusion-v1-5/stable-diffusion-inpainting",
|
||||||
custom_pipeline="text_inpainting",
|
custom_pipeline="text_inpainting",
|
||||||
segmentation_model=model,
|
segmentation_model=model,
|
||||||
segmentation_processor=processor
|
segmentation_processor=processor
|
||||||
|
|||||||
@ -126,7 +126,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
... "lllyasviel/control_v11p_sd15_inpaint", torch_dtype=torch.float16
|
... "lllyasviel/control_v11p_sd15_inpaint", torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
>>> pipe = StableDiffusionControlNetInpaintPipeline.from_pretrained(
|
>>> pipe = StableDiffusionControlNetInpaintPipeline.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
|
|
||||||
>>> pipe.scheduler = DDIMScheduler.from_config(pipe.scheduler.config)
|
>>> pipe.scheduler = DDIMScheduler.from_config(pipe.scheduler.config)
|
||||||
@ -347,7 +347,7 @@ class AdaptiveMaskInpaintPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
@ -429,8 +429,8 @@ class AdaptiveMaskInpaintPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
@ -970,7 +970,7 @@ class AdaptiveMaskInpaintPipeline(
|
|||||||
>>> default_mask_image = download_image(mask_url).resize((512, 512))
|
>>> default_mask_image = download_image(mask_url).resize((512, 512))
|
||||||
|
|
||||||
>>> pipe = AdaptiveMaskInpaintPipeline.from_pretrained(
|
>>> pipe = AdaptiveMaskInpaintPipeline.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-inpainting", torch_dtype=torch.float16
|
... "stable-diffusion-v1-5/stable-diffusion-inpainting", torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
|
|
||||||
@ -1095,7 +1095,7 @@ class AdaptiveMaskInpaintPipeline(
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -62,7 +62,7 @@ class ComposableStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin)
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -145,8 +145,8 @@ class ComposableStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin)
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -1276,7 +1276,7 @@ class FrescoV2VPipeline(StableDiffusionControlNetImg2ImgPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
|
|||||||
@ -678,7 +678,7 @@ class StableDiffusionHDPainterPipeline(StableDiffusionInpaintPipeline):
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -78,7 +78,7 @@ class ImageToImageInpaintingPipeline(DiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -86,7 +86,7 @@ class InstaFlowPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
@ -165,8 +165,8 @@ class InstaFlowPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -166,7 +166,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
@ -247,8 +247,8 @@ class IPAdapterFaceIDStableDiffusionPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -414,7 +414,7 @@ class StableDiffusionHighResFixPipeline(StableDiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
|
|||||||
@ -222,7 +222,7 @@ class LatentConsistencyModelWalkPipeline(
|
|||||||
supports [`LCMScheduler`].
|
supports [`LCMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
|
|||||||
@ -302,7 +302,7 @@ class LLMGroundedDiffusionPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
@ -392,8 +392,8 @@ class LLMGroundedDiffusionPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -552,8 +552,8 @@ class StableDiffusionLongPromptWeightingPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -1765,7 +1765,7 @@ class SDXLLongPromptWeightingPipeline(
|
|||||||
|
|
||||||
# Check that sizes of mask, masked image and latents match
|
# Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != num_channels_unet:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != num_channels_unet:
|
||||||
|
|||||||
@ -3729,8 +3729,8 @@ class MatryoshkaPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -78,7 +78,7 @@ class MultilingualStableDiffusion(DiffusionPipeline, StableDiffusionMixin):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -1607,7 +1607,7 @@ class KolorsControlNetInpaintPipeline(
|
|||||||
|
|
||||||
# 9. Check that sizes of mask, masked image and latents match
|
# 9. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -135,7 +135,7 @@ class FabricPipeline(DiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@ -163,8 +163,8 @@ class FabricPipeline(DiffusionPipeline):
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -1487,7 +1487,7 @@ class KolorsInpaintPipeline(
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -106,7 +106,7 @@ class Prompt2PromptPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
@ -187,8 +187,8 @@ class Prompt2PromptPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -1730,7 +1730,7 @@ class StyleAlignedSDXLPipeline(
|
|||||||
|
|
||||||
# Check that sizes of mask, masked image and latents match
|
# Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != num_channels_unet:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != num_channels_unet:
|
||||||
|
|||||||
@ -59,7 +59,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> import torch
|
>>> import torch
|
||||||
>>> from diffusers import StableDiffusionPipeline
|
>>> from diffusers import StableDiffusionPipeline
|
||||||
|
|
||||||
>>> pipe = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16)
|
>>> pipe = StableDiffusionPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5", torch_dtype=torch.float16)
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
|
|
||||||
>>> prompt = "a photo of an astronaut riding a horse on mars"
|
>>> prompt = "a photo of an astronaut riding a horse on mars"
|
||||||
@ -392,7 +392,7 @@ class StableDiffusionBoxDiffPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
@ -473,8 +473,8 @@ class StableDiffusionBoxDiffPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -42,7 +42,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
```py
|
```py
|
||||||
>>> import torch
|
>>> import torch
|
||||||
>>> from diffusers import StableDiffusionPipeline
|
>>> from diffusers import StableDiffusionPipeline
|
||||||
>>> pipe = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16)
|
>>> pipe = StableDiffusionPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5", torch_dtype=torch.float16)
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
>>> prompt = "a photo of an astronaut riding a horse on mars"
|
>>> prompt = "a photo of an astronaut riding a horse on mars"
|
||||||
>>> image = pipe(prompt).images[0]
|
>>> image = pipe(prompt).images[0]
|
||||||
@ -359,7 +359,7 @@ class StableDiffusionPAGPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
@ -440,8 +440,8 @@ class StableDiffusionPAGPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -100,7 +100,7 @@ class StableDiffusionUpscaleLDM3DPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
|
|||||||
@ -2042,7 +2042,7 @@ class StableDiffusionXL_AE_Pipeline(
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -188,7 +188,7 @@ class StableDiffusionXLControlNetAdapterPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -330,7 +330,7 @@ class StableDiffusionXLControlNetAdapterInpaintPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
requires_aesthetics_score (`bool`, *optional*, defaults to `"False"`):
|
requires_aesthetics_score (`bool`, *optional*, defaults to `"False"`):
|
||||||
@ -1569,7 +1569,7 @@ class StableDiffusionXLControlNetAdapterInpaintPipeline(
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -46,7 +46,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> import torch
|
>>> import torch
|
||||||
>>> from diffusers import StableDiffusionPipeline
|
>>> from diffusers import StableDiffusionPipeline
|
||||||
|
|
||||||
>>> pipe = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16)
|
>>> pipe = StableDiffusionPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5", torch_dtype=torch.float16)
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
|
|
||||||
>>> prompt = "a photo of an astronaut riding a horse on mars"
|
>>> prompt = "a photo of an astronaut riding a horse on mars"
|
||||||
@ -86,7 +86,7 @@ class Zero1to3StableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
cc_projection ([`CCProjection`]):
|
cc_projection ([`CCProjection`]):
|
||||||
@ -164,8 +164,8 @@ class Zero1to3StableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -288,7 +288,7 @@ class RerenderAVideoPipeline(StableDiffusionControlNetImg2ImgPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -54,7 +54,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> # load control net and stable diffusion v1-5
|
>>> # load control net and stable diffusion v1-5
|
||||||
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
||||||
>>> pipe = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(
|
>>> pipe = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
|
|
||||||
>>> # speed up diffusion process with faster scheduler and memory optimization
|
>>> # speed up diffusion process with faster scheduler and memory optimization
|
||||||
|
|||||||
@ -158,7 +158,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> # load control net and stable diffusion v1-5
|
>>> # load control net and stable diffusion v1-5
|
||||||
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
||||||
>>> pipe = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(
|
>>> pipe = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
|
|
||||||
>>> # speed up diffusion process with faster scheduler and memory optimization
|
>>> # speed up diffusion process with faster scheduler and memory optimization
|
||||||
|
|||||||
@ -64,7 +64,7 @@ class StableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -114,7 +114,7 @@ class SdeDragPipeline(DiffusionPipeline):
|
|||||||
>>> from diffusers import DDIMScheduler, DiffusionPipeline
|
>>> from diffusers import DDIMScheduler, DiffusionPipeline
|
||||||
|
|
||||||
>>> # Load the pipeline
|
>>> # Load the pipeline
|
||||||
>>> model_path = "runwayml/stable-diffusion-v1-5"
|
>>> model_path = "stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
>>> scheduler = DDIMScheduler.from_pretrained(model_path, subfolder="scheduler")
|
>>> scheduler = DDIMScheduler.from_pretrained(model_path, subfolder="scheduler")
|
||||||
>>> pipe = DiffusionPipeline.from_pretrained(model_path, scheduler=scheduler, custom_pipeline="sde_drag")
|
>>> pipe = DiffusionPipeline.from_pretrained(model_path, scheduler=scheduler, custom_pipeline="sde_drag")
|
||||||
>>> pipe.to('cuda')
|
>>> pipe.to('cuda')
|
||||||
|
|||||||
@ -46,7 +46,7 @@ class StableDiffusionComparisonPipeline(DiffusionPipeline, StableDiffusionMixin)
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionMegaSafetyChecker`]):
|
safety_checker ([`StableDiffusionMegaSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -36,7 +36,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
||||||
|
|
||||||
>>> pipe_controlnet = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(
|
>>> pipe_controlnet = StableDiffusionControlNetImg2ImgPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-v1-5",
|
"stable-diffusion-v1-5/stable-diffusion-v1-5",
|
||||||
controlnet=controlnet,
|
controlnet=controlnet,
|
||||||
safety_checker=None,
|
safety_checker=None,
|
||||||
torch_dtype=torch.float16
|
torch_dtype=torch.float16
|
||||||
|
|||||||
@ -81,7 +81,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-seg", torch_dtype=torch.float16)
|
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-seg", torch_dtype=torch.float16)
|
||||||
|
|
||||||
>>> pipe = StableDiffusionControlNetInpaintPipeline.from_pretrained(
|
>>> pipe = StableDiffusionControlNetInpaintPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", controlnet=controlnet, safety_checker=None, torch_dtype=torch.float16
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", controlnet=controlnet, safety_checker=None, torch_dtype=torch.float16
|
||||||
)
|
)
|
||||||
|
|
||||||
>>> pipe.scheduler = UniPCMultistepScheduler.from_config(pipe.scheduler.config)
|
>>> pipe.scheduler = UniPCMultistepScheduler.from_config(pipe.scheduler.config)
|
||||||
|
|||||||
@ -80,7 +80,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-seg", torch_dtype=torch.float16)
|
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-seg", torch_dtype=torch.float16)
|
||||||
|
|
||||||
>>> pipe = StableDiffusionControlNetInpaintImg2ImgPipeline.from_pretrained(
|
>>> pipe = StableDiffusionControlNetInpaintImg2ImgPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting", controlnet=controlnet, safety_checker=None, torch_dtype=torch.float16
|
"stable-diffusion-v1-5/stable-diffusion-inpainting", controlnet=controlnet, safety_checker=None, torch_dtype=torch.float16
|
||||||
)
|
)
|
||||||
|
|
||||||
>>> pipe.scheduler = UniPCMultistepScheduler.from_config(pipe.scheduler.config)
|
>>> pipe.scheduler = UniPCMultistepScheduler.from_config(pipe.scheduler.config)
|
||||||
|
|||||||
@ -37,7 +37,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
|
|
||||||
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
||||||
>>> pipe = StableDiffusionControlNetReferencePipeline.from_pretrained(
|
>>> pipe = StableDiffusionControlNetReferencePipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-v1-5",
|
"stable-diffusion-v1-5/stable-diffusion-v1-5",
|
||||||
controlnet=controlnet,
|
controlnet=controlnet,
|
||||||
safety_checker=None,
|
safety_checker=None,
|
||||||
torch_dtype=torch.float16
|
torch_dtype=torch.float16
|
||||||
|
|||||||
@ -43,7 +43,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> import torch
|
>>> import torch
|
||||||
>>> from diffusers import StableDiffusionPipeline
|
>>> from diffusers import StableDiffusionPipeline
|
||||||
|
|
||||||
>>> pipe = DiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", custom_pipeline="stable_diffusion_ipex")
|
>>> pipe = DiffusionPipeline.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5", custom_pipeline="stable_diffusion_ipex")
|
||||||
|
|
||||||
>>> # For Float32
|
>>> # For Float32
|
||||||
>>> pipe.prepare_for_ipex(prompt, dtype=torch.float32, height=512, width=512) #value of image height/width should be consistent with the pipeline inference
|
>>> pipe.prepare_for_ipex(prompt, dtype=torch.float32, height=512, width=512) #value of image height/width should be consistent with the pipeline inference
|
||||||
@ -85,7 +85,7 @@ class StableDiffusionIPEXPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -161,8 +161,8 @@ class StableDiffusionIPEXPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -47,7 +47,7 @@ class StableDiffusionMegaPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionMegaSafetyChecker`]):
|
safety_checker ([`StableDiffusionMegaSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -46,7 +46,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> input_image = load_image("https://hf.co/datasets/huggingface/documentation-images/resolve/main/diffusers/input_image_vermeer.png")
|
>>> input_image = load_image("https://hf.co/datasets/huggingface/documentation-images/resolve/main/diffusers/input_image_vermeer.png")
|
||||||
|
|
||||||
>>> pipe = StableDiffusionReferencePipeline.from_pretrained(
|
>>> pipe = StableDiffusionReferencePipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-v1-5",
|
"stable-diffusion-v1-5/stable-diffusion-v1-5",
|
||||||
safety_checker=None,
|
safety_checker=None,
|
||||||
torch_dtype=torch.float16
|
torch_dtype=torch.float16
|
||||||
).to('cuda:0')
|
).to('cuda:0')
|
||||||
@ -112,7 +112,7 @@ class StableDiffusionReferencePipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -194,8 +194,8 @@ class StableDiffusionReferencePipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -167,7 +167,7 @@ class StableDiffusionRepaintPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -249,8 +249,8 @@ class StableDiffusionRepaintPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -678,7 +678,7 @@ class TensorRTStableDiffusionImg2ImgPipeline(DiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -766,8 +766,8 @@ class TensorRTStableDiffusionImg2ImgPipeline(DiffusionPipeline):
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -682,7 +682,7 @@ class TensorRTStableDiffusionInpaintPipeline(DiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -770,8 +770,8 @@ class TensorRTStableDiffusionInpaintPipeline(DiffusionPipeline):
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -594,7 +594,7 @@ class TensorRTStableDiffusionPipeline(DiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -682,8 +682,8 @@ class TensorRTStableDiffusionPipeline(DiffusionPipeline):
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -52,7 +52,7 @@ class TextInpainting(DiffusionPipeline, StableDiffusionMixin):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -1223,7 +1223,7 @@ class AnyTextPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
|
|||||||
@ -5,7 +5,7 @@ This script was added by @thedarkzeno .
|
|||||||
Please note that this script is not actively maintained, you can open an issue and tag @thedarkzeno or @patil-suraj though.
|
Please note that this script is not actively maintained, you can open an issue and tag @thedarkzeno or @patil-suraj though.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
export MODEL_NAME="runwayml/stable-diffusion-inpainting"
|
export MODEL_NAME="stable-diffusion-v1-5/stable-diffusion-inpainting"
|
||||||
export INSTANCE_DIR="path-to-instance-images"
|
export INSTANCE_DIR="path-to-instance-images"
|
||||||
export OUTPUT_DIR="path-to-save-model"
|
export OUTPUT_DIR="path-to-save-model"
|
||||||
|
|
||||||
@ -29,7 +29,7 @@ Prior-preservation is used to avoid overfitting and language-drift. Refer to the
|
|||||||
According to the paper, it's recommended to generate `num_epochs * num_samples` images for prior-preservation. 200-300 works well for most cases.
|
According to the paper, it's recommended to generate `num_epochs * num_samples` images for prior-preservation. 200-300 works well for most cases.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
export MODEL_NAME="runwayml/stable-diffusion-inpainting"
|
export MODEL_NAME="stable-diffusion-v1-5/stable-diffusion-inpainting"
|
||||||
export INSTANCE_DIR="path-to-instance-images"
|
export INSTANCE_DIR="path-to-instance-images"
|
||||||
export CLASS_DIR="path-to-class-images"
|
export CLASS_DIR="path-to-class-images"
|
||||||
export OUTPUT_DIR="path-to-save-model"
|
export OUTPUT_DIR="path-to-save-model"
|
||||||
@ -60,7 +60,7 @@ With the help of gradient checkpointing and the 8-bit optimizer from bitsandbyte
|
|||||||
To install `bitandbytes` please refer to this [readme](https://github.com/TimDettmers/bitsandbytes#requirements--installation).
|
To install `bitandbytes` please refer to this [readme](https://github.com/TimDettmers/bitsandbytes#requirements--installation).
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
export MODEL_NAME="runwayml/stable-diffusion-inpainting"
|
export MODEL_NAME="stable-diffusion-v1-5/stable-diffusion-inpainting"
|
||||||
export INSTANCE_DIR="path-to-instance-images"
|
export INSTANCE_DIR="path-to-instance-images"
|
||||||
export CLASS_DIR="path-to-class-images"
|
export CLASS_DIR="path-to-class-images"
|
||||||
export OUTPUT_DIR="path-to-save-model"
|
export OUTPUT_DIR="path-to-save-model"
|
||||||
@ -92,7 +92,7 @@ Pass the `--train_text_encoder` argument to the script to enable training `text_
|
|||||||
___Note: Training text encoder requires more memory, with this option the training won't fit on 16GB GPU. It needs at least 24GB VRAM.___
|
___Note: Training text encoder requires more memory, with this option the training won't fit on 16GB GPU. It needs at least 24GB VRAM.___
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
export MODEL_NAME="runwayml/stable-diffusion-inpainting"
|
export MODEL_NAME="stable-diffusion-v1-5/stable-diffusion-inpainting"
|
||||||
export INSTANCE_DIR="path-to-instance-images"
|
export INSTANCE_DIR="path-to-instance-images"
|
||||||
export CLASS_DIR="path-to-class-images"
|
export CLASS_DIR="path-to-class-images"
|
||||||
export OUTPUT_DIR="path-to-save-model"
|
export OUTPUT_DIR="path-to-save-model"
|
||||||
|
|||||||
@ -55,7 +55,7 @@ The Accelerate launch command is used to train a model using multiple GPUs and m
|
|||||||
```
|
```
|
||||||
accelerate launch --mixed_precision "fp16" \
|
accelerate launch --mixed_precision "fp16" \
|
||||||
tutorial_train_ip-adapter.py \
|
tutorial_train_ip-adapter.py \
|
||||||
--pretrained_model_name_or_path="runwayml/stable-diffusion-v1-5/" \
|
--pretrained_model_name_or_path="stable-diffusion-v1-5/stable-diffusion-v1-5/" \
|
||||||
--image_encoder_path="{image_encoder_path}" \
|
--image_encoder_path="{image_encoder_path}" \
|
||||||
--data_json_file="{data.json}" \
|
--data_json_file="{data.json}" \
|
||||||
--data_root_path="{image_path}" \
|
--data_root_path="{image_path}" \
|
||||||
@ -73,7 +73,7 @@ tutorial_train_ip-adapter.py \
|
|||||||
```
|
```
|
||||||
accelerate launch --num_processes 8 --multi_gpu --mixed_precision "fp16" \
|
accelerate launch --num_processes 8 --multi_gpu --mixed_precision "fp16" \
|
||||||
tutorial_train_ip-adapter.py \
|
tutorial_train_ip-adapter.py \
|
||||||
--pretrained_model_name_or_path="runwayml/stable-diffusion-v1-5/" \
|
--pretrained_model_name_or_path="stable-diffusion-v1-5/stable-diffusion-v1-5/" \
|
||||||
--image_encoder_path="{image_encoder_path}" \
|
--image_encoder_path="{image_encoder_path}" \
|
||||||
--data_json_file="{data.json}" \
|
--data_json_file="{data.json}" \
|
||||||
--data_root_path="{image_path}" \
|
--data_root_path="{image_path}" \
|
||||||
|
|||||||
@ -27,7 +27,7 @@ You can build multiple datasets for every subject and upload them to the 🤗 hu
|
|||||||
Before launching the training script, make sure to select the inpainting the target model, the output directory and the 🤗 datasets.
|
Before launching the training script, make sure to select the inpainting the target model, the output directory and the 🤗 datasets.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
export MODEL_NAME="runwayml/stable-diffusion-inpainting"
|
export MODEL_NAME="stable-diffusion-v1-5/stable-diffusion-inpainting"
|
||||||
export OUTPUT_DIR="path-to-save-model"
|
export OUTPUT_DIR="path-to-save-model"
|
||||||
|
|
||||||
export DATASET_1="gzguevara/mr_potato_head_masked"
|
export DATASET_1="gzguevara/mr_potato_head_masked"
|
||||||
|
|||||||
@ -177,7 +177,7 @@ class PromptDiffusionPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for more details
|
||||||
about a model's potential harms.
|
about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
|
|||||||
@ -238,7 +238,7 @@ def parse_args() -> argparse.Namespace:
|
|||||||
|
|
||||||
# EXAMPLE USAGE:
|
# EXAMPLE USAGE:
|
||||||
#
|
#
|
||||||
# python vae_roundtrip.py --use_cuda --pretrained_model_name_or_path "runwayml/stable-diffusion-v1-5" --subfolder "vae" --input_image "foo.png"
|
# python vae_roundtrip.py --use_cuda --pretrained_model_name_or_path "stable-diffusion-v1-5/stable-diffusion-v1-5" --subfolder "vae" --input_image "foo.png"
|
||||||
#
|
#
|
||||||
# python vae_roundtrip.py --use_cuda --pretrained_model_name_or_path "madebyollin/taesd" --use_tiny_nn --input_image "foo.png"
|
# python vae_roundtrip.py --use_cuda --pretrained_model_name_or_path "madebyollin/taesd" --use_tiny_nn --input_image "foo.png"
|
||||||
#
|
#
|
||||||
|
|||||||
@ -24,7 +24,8 @@ args = args.parse_args()
|
|||||||
|
|
||||||
|
|
||||||
def _extract_into_tensor(arr, timesteps, broadcast_shape):
|
def _extract_into_tensor(arr, timesteps, broadcast_shape):
|
||||||
# from: https://github.com/openai/guided-diffusion/blob/22e0df8183507e13a7813f8d38d51b072ca1e67c/guided_diffusion/gaussian_diffusion.py#L895 """
|
# from: https://github.com/openai/guided-diffusion/blob/22e0df8183507e13a7813f8d38d51b072ca1e67c/guided_diffusion/gaussian_diffusion.py#L895
|
||||||
|
# """
|
||||||
res = arr[timesteps].float()
|
res = arr[timesteps].float()
|
||||||
dims_to_append = len(broadcast_shape) - len(res.shape)
|
dims_to_append = len(broadcast_shape) - len(res.shape)
|
||||||
return res[(...,) + (None,) * dims_to_append]
|
return res[(...,) + (None,) * dims_to_append]
|
||||||
@ -507,7 +508,9 @@ def rename_state_dict(sd, embedding):
|
|||||||
|
|
||||||
|
|
||||||
# encode with stable diffusion vae
|
# encode with stable diffusion vae
|
||||||
pipe = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16)
|
pipe = StableDiffusionPipeline.from_pretrained(
|
||||||
|
"stable-diffusion-v1-5/stable-diffusion-v1-5", torch_dtype=torch.float16
|
||||||
|
)
|
||||||
pipe.vae.cuda()
|
pipe.vae.cuda()
|
||||||
|
|
||||||
# construct original decoder with jitted model
|
# construct original decoder with jitted model
|
||||||
@ -1090,7 +1093,7 @@ def new_constructor(self, **kwargs):
|
|||||||
Encoder.__init__ = new_constructor
|
Encoder.__init__ = new_constructor
|
||||||
|
|
||||||
|
|
||||||
vae = AutoencoderKL.from_pretrained("runwayml/stable-diffusion-v1-5", subfolder="vae")
|
vae = AutoencoderKL.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5", subfolder="vae")
|
||||||
consistency_vae = ConsistencyDecoderVAE(
|
consistency_vae = ConsistencyDecoderVAE(
|
||||||
encoder_args=vae.encoder.constructor_arguments,
|
encoder_args=vae.encoder.constructor_arguments,
|
||||||
decoder_args=unet.config,
|
decoder_args=unet.config,
|
||||||
@ -1117,7 +1120,7 @@ print((sample_consistency_orig - sample_consistency_new_3).abs().sum())
|
|||||||
print("running with diffusers pipeline")
|
print("running with diffusers pipeline")
|
||||||
|
|
||||||
pipe = DiffusionPipeline.from_pretrained(
|
pipe = DiffusionPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-v1-5", vae=consistency_vae, torch_dtype=torch.float16
|
"stable-diffusion-v1-5/stable-diffusion-v1-5", vae=consistency_vae, torch_dtype=torch.float16
|
||||||
)
|
)
|
||||||
pipe.to("cuda")
|
pipe.to("cuda")
|
||||||
|
|
||||||
|
|||||||
@ -128,13 +128,13 @@ class AutoModel(ConfigMixin):
|
|||||||
```py
|
```py
|
||||||
from diffusers import AutoModel
|
from diffusers import AutoModel
|
||||||
|
|
||||||
unet = AutoModel.from_pretrained("runwayml/stable-diffusion-v1-5", subfolder="unet")
|
unet = AutoModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5", subfolder="unet")
|
||||||
```
|
```
|
||||||
|
|
||||||
If you get the error message below, you need to finetune the weights for your downstream task:
|
If you get the error message below, you need to finetune the weights for your downstream task:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
Some weights of UNet2DConditionModel were not initialized from the model checkpoint at runwayml/stable-diffusion-v1-5 and are newly initialized because the shapes did not match:
|
Some weights of UNet2DConditionModel were not initialized from the model checkpoint at stable-diffusion-v1-5/stable-diffusion-v1-5 and are newly initialized because the shapes did not match:
|
||||||
- conv_in.weight: found shape torch.Size([320, 4, 3, 3]) in the checkpoint and torch.Size([320, 9, 3, 3]) in the model instantiated
|
- conv_in.weight: found shape torch.Size([320, 4, 3, 3]) in the checkpoint and torch.Size([320, 9, 3, 3]) in the model instantiated
|
||||||
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
||||||
```
|
```
|
||||||
|
|||||||
@ -113,14 +113,14 @@ class FlaxModelMixin(PushToHubMixin):
|
|||||||
>>> from diffusers import FlaxUNet2DConditionModel
|
>>> from diffusers import FlaxUNet2DConditionModel
|
||||||
|
|
||||||
>>> # load model
|
>>> # load model
|
||||||
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("runwayml/stable-diffusion-v1-5")
|
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5")
|
||||||
>>> # By default, the model parameters will be in fp32 precision, to cast these to bfloat16 precision
|
>>> # By default, the model parameters will be in fp32 precision, to cast these to bfloat16 precision
|
||||||
>>> params = model.to_bf16(params)
|
>>> params = model.to_bf16(params)
|
||||||
>>> # If you don't want to cast certain parameters (for example layer norm bias and scale)
|
>>> # If you don't want to cast certain parameters (for example layer norm bias and scale)
|
||||||
>>> # then pass the mask as follows
|
>>> # then pass the mask as follows
|
||||||
>>> from flax import traverse_util
|
>>> from flax import traverse_util
|
||||||
|
|
||||||
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("runwayml/stable-diffusion-v1-5")
|
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5")
|
||||||
>>> flat_params = traverse_util.flatten_dict(params)
|
>>> flat_params = traverse_util.flatten_dict(params)
|
||||||
>>> mask = {
|
>>> mask = {
|
||||||
... path: (path[-2] != ("LayerNorm", "bias") and path[-2:] != ("LayerNorm", "scale"))
|
... path: (path[-2] != ("LayerNorm", "bias") and path[-2:] != ("LayerNorm", "scale"))
|
||||||
@ -149,7 +149,7 @@ class FlaxModelMixin(PushToHubMixin):
|
|||||||
>>> from diffusers import FlaxUNet2DConditionModel
|
>>> from diffusers import FlaxUNet2DConditionModel
|
||||||
|
|
||||||
>>> # Download model and configuration from huggingface.co
|
>>> # Download model and configuration from huggingface.co
|
||||||
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("runwayml/stable-diffusion-v1-5")
|
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5")
|
||||||
>>> # By default, the model params will be in fp32, to illustrate the use of this method,
|
>>> # By default, the model params will be in fp32, to illustrate the use of this method,
|
||||||
>>> # we'll first cast to fp16 and back to fp32
|
>>> # we'll first cast to fp16 and back to fp32
|
||||||
>>> params = model.to_f16(params)
|
>>> params = model.to_f16(params)
|
||||||
@ -179,14 +179,14 @@ class FlaxModelMixin(PushToHubMixin):
|
|||||||
>>> from diffusers import FlaxUNet2DConditionModel
|
>>> from diffusers import FlaxUNet2DConditionModel
|
||||||
|
|
||||||
>>> # load model
|
>>> # load model
|
||||||
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("runwayml/stable-diffusion-v1-5")
|
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5")
|
||||||
>>> # By default, the model params will be in fp32, to cast these to float16
|
>>> # By default, the model params will be in fp32, to cast these to float16
|
||||||
>>> params = model.to_fp16(params)
|
>>> params = model.to_fp16(params)
|
||||||
>>> # If you want don't want to cast certain parameters (for example layer norm bias and scale)
|
>>> # If you want don't want to cast certain parameters (for example layer norm bias and scale)
|
||||||
>>> # then pass the mask as follows
|
>>> # then pass the mask as follows
|
||||||
>>> from flax import traverse_util
|
>>> from flax import traverse_util
|
||||||
|
|
||||||
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("runwayml/stable-diffusion-v1-5")
|
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5")
|
||||||
>>> flat_params = traverse_util.flatten_dict(params)
|
>>> flat_params = traverse_util.flatten_dict(params)
|
||||||
>>> mask = {
|
>>> mask = {
|
||||||
... path: (path[-2] != ("LayerNorm", "bias") and path[-2:] != ("LayerNorm", "scale"))
|
... path: (path[-2] != ("LayerNorm", "bias") and path[-2:] != ("LayerNorm", "scale"))
|
||||||
@ -216,8 +216,8 @@ class FlaxModelMixin(PushToHubMixin):
|
|||||||
pretrained_model_name_or_path (`str` or `os.PathLike`):
|
pretrained_model_name_or_path (`str` or `os.PathLike`):
|
||||||
Can be either:
|
Can be either:
|
||||||
|
|
||||||
- A string, the *model id* (for example `runwayml/stable-diffusion-v1-5`) of a pretrained model
|
- A string, the *model id* (for example `stable-diffusion-v1-5/stable-diffusion-v1-5`) of a
|
||||||
hosted on the Hub.
|
pretrained model hosted on the Hub.
|
||||||
- A path to a *directory* (for example `./my_model_directory`) containing the model weights saved
|
- A path to a *directory* (for example `./my_model_directory`) containing the model weights saved
|
||||||
using [`~FlaxModelMixin.save_pretrained`].
|
using [`~FlaxModelMixin.save_pretrained`].
|
||||||
dtype (`jax.numpy.dtype`, *optional*, defaults to `jax.numpy.float32`):
|
dtype (`jax.numpy.dtype`, *optional*, defaults to `jax.numpy.float32`):
|
||||||
@ -271,7 +271,7 @@ class FlaxModelMixin(PushToHubMixin):
|
|||||||
>>> from diffusers import FlaxUNet2DConditionModel
|
>>> from diffusers import FlaxUNet2DConditionModel
|
||||||
|
|
||||||
>>> # Download model and configuration from huggingface.co and cache.
|
>>> # Download model and configuration from huggingface.co and cache.
|
||||||
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("runwayml/stable-diffusion-v1-5")
|
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5")
|
||||||
>>> # Model was saved using *save_pretrained('./test/saved_model/')* (for example purposes, not runnable).
|
>>> # Model was saved using *save_pretrained('./test/saved_model/')* (for example purposes, not runnable).
|
||||||
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("./test/saved_model/")
|
>>> model, params = FlaxUNet2DConditionModel.from_pretrained("./test/saved_model/")
|
||||||
```
|
```
|
||||||
@ -279,7 +279,7 @@ class FlaxModelMixin(PushToHubMixin):
|
|||||||
If you get the error message below, you need to finetune the weights for your downstream task:
|
If you get the error message below, you need to finetune the weights for your downstream task:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
Some weights of UNet2DConditionModel were not initialized from the model checkpoint at runwayml/stable-diffusion-v1-5 and are newly initialized because the shapes did not match:
|
Some weights of UNet2DConditionModel were not initialized from the model checkpoint at stable-diffusion-v1-5/stable-diffusion-v1-5 and are newly initialized because the shapes did not match:
|
||||||
- conv_in.weight: found shape torch.Size([320, 4, 3, 3]) in the checkpoint and torch.Size([320, 9, 3, 3]) in the model instantiated
|
- conv_in.weight: found shape torch.Size([320, 4, 3, 3]) in the checkpoint and torch.Size([320, 9, 3, 3]) in the model instantiated
|
||||||
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
||||||
```
|
```
|
||||||
|
|||||||
@ -923,13 +923,13 @@ class ModelMixin(torch.nn.Module, PushToHubMixin):
|
|||||||
```py
|
```py
|
||||||
from diffusers import UNet2DConditionModel
|
from diffusers import UNet2DConditionModel
|
||||||
|
|
||||||
unet = UNet2DConditionModel.from_pretrained("runwayml/stable-diffusion-v1-5", subfolder="unet")
|
unet = UNet2DConditionModel.from_pretrained("stable-diffusion-v1-5/stable-diffusion-v1-5", subfolder="unet")
|
||||||
```
|
```
|
||||||
|
|
||||||
If you get the error message below, you need to finetune the weights for your downstream task:
|
If you get the error message below, you need to finetune the weights for your downstream task:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
Some weights of UNet2DConditionModel were not initialized from the model checkpoint at runwayml/stable-diffusion-v1-5 and are newly initialized because the shapes did not match:
|
Some weights of UNet2DConditionModel were not initialized from the model checkpoint at stable-diffusion-v1-5/stable-diffusion-v1-5 and are newly initialized because the shapes did not match:
|
||||||
- conv_in.weight: found shape torch.Size([320, 4, 3, 3]) in the checkpoint and torch.Size([320, 9, 3, 3]) in the model instantiated
|
- conv_in.weight: found shape torch.Size([320, 4, 3, 3]) in the checkpoint and torch.Size([320, 9, 3, 3]) in the model instantiated
|
||||||
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
||||||
```
|
```
|
||||||
@ -1800,7 +1800,7 @@ class ModelMixin(torch.nn.Module, PushToHubMixin):
|
|||||||
```py
|
```py
|
||||||
from diffusers import UNet2DConditionModel
|
from diffusers import UNet2DConditionModel
|
||||||
|
|
||||||
model_id = "runwayml/stable-diffusion-v1-5"
|
model_id = "stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
unet = UNet2DConditionModel.from_pretrained(model_id, subfolder="unet")
|
unet = UNet2DConditionModel.from_pretrained(model_id, subfolder="unet")
|
||||||
unet.num_parameters(only_trainable=True)
|
unet.num_parameters(only_trainable=True)
|
||||||
859520964
|
859520964
|
||||||
|
|||||||
@ -115,7 +115,7 @@ class StableDiffusionXLInpaintLoopBeforeDenoiser(ModularPipelineBlocks):
|
|||||||
def check_inputs(components, block_state):
|
def check_inputs(components, block_state):
|
||||||
num_channels_unet = components.num_channels_unet
|
num_channels_unet = components.num_channels_unet
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
if block_state.mask is None or block_state.masked_image_latents is None:
|
if block_state.mask is None or block_state.masked_image_latents is None:
|
||||||
raise ValueError("mask and masked_image_latents must be provided for inpainting-specific Unet")
|
raise ValueError("mask and masked_image_latents must be provided for inpainting-specific Unet")
|
||||||
num_channels_latents = block_state.latents.shape[1]
|
num_channels_latents = block_state.latents.shape[1]
|
||||||
|
|||||||
@ -159,7 +159,7 @@ init_image = download_image(img_url).resize((512, 512))
|
|||||||
mask_image = download_image(mask_url).resize((512, 512))
|
mask_image = download_image(mask_url).resize((512, 512))
|
||||||
|
|
||||||
pipe = StableDiffusionInpaintPipeline.from_pretrained(
|
pipe = StableDiffusionInpaintPipeline.from_pretrained(
|
||||||
"runwayml/stable-diffusion-inpainting",
|
"stable-diffusion-v1-5/stable-diffusion-inpainting",
|
||||||
torch_dtype=torch.float16,
|
torch_dtype=torch.float16,
|
||||||
)
|
)
|
||||||
pipe = pipe.to("cuda")
|
pipe = pipe.to("cuda")
|
||||||
|
|||||||
@ -133,8 +133,8 @@ class StableDiffusionControlNetXSPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -185,8 +185,8 @@ class AltDiffusionPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -266,8 +266,8 @@ class AltDiffusionPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -213,8 +213,8 @@ class AltDiffusionImg2ImgPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -294,8 +294,8 @@ class AltDiffusionImg2ImgPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -162,8 +162,8 @@ class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMixin, Sta
|
|||||||
instance of [`DDIMScheduler`].
|
instance of [`DDIMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -226,8 +226,8 @@ class CycleDiffusionPipeline(DiffusionPipeline, TextualInversionLoaderMixin, Sta
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely .If you're checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -62,7 +62,8 @@ class OnnxStableDiffusionInpaintPipelineLegacy(DiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
|
details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -111,7 +111,8 @@ class StableDiffusionInpaintPipelineLegacy(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
|
details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -196,8 +197,8 @@ class StableDiffusionInpaintPipelineLegacy(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -64,8 +64,8 @@ class StableDiffusionModelEditingPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
with_to_k ([`bool`]):
|
with_to_k ([`bool`]):
|
||||||
|
|||||||
@ -46,10 +46,12 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> from diffusers import DDPMParallelScheduler
|
>>> from diffusers import DDPMParallelScheduler
|
||||||
>>> from diffusers import StableDiffusionParadigmsPipeline
|
>>> from diffusers import StableDiffusionParadigmsPipeline
|
||||||
|
|
||||||
>>> scheduler = DDPMParallelScheduler.from_pretrained("runwayml/stable-diffusion-v1-5", subfolder="scheduler")
|
>>> scheduler = DDPMParallelScheduler.from_pretrained(
|
||||||
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", subfolder="scheduler"
|
||||||
|
... )
|
||||||
|
|
||||||
>>> pipe = StableDiffusionParadigmsPipeline.from_pretrained(
|
>>> pipe = StableDiffusionParadigmsPipeline.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", scheduler=scheduler, torch_dtype=torch.float16
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", scheduler=scheduler, torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
|
|
||||||
@ -95,8 +97,8 @@ class StableDiffusionParadigmsPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -303,7 +303,8 @@ class StableDiffusionPix2PixZeroPipeline(DiffusionPipeline, StableDiffusionMixin
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], [`EulerAncestralDiscreteScheduler`], or [`DDPMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], [`EulerAncestralDiscreteScheduler`], or [`DDPMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please, refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for details.
|
Please, refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
|
details.
|
||||||
feature_extractor ([`CLIPImageProcessor`]):
|
feature_extractor ([`CLIPImageProcessor`]):
|
||||||
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
Model that extracts features from generated images to be used as inputs for the `safety_checker`.
|
||||||
requires_safety_checker (bool):
|
requires_safety_checker (bool):
|
||||||
|
|||||||
@ -38,8 +38,8 @@ class VersatileDiffusionPipeline(DiffusionPipeline):
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -186,8 +186,8 @@ class LatentConsistencyModelImg2ImgPipeline(
|
|||||||
supports [`LCMScheduler`].
|
supports [`LCMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
requires_safety_checker (`bool`, *optional*, defaults to `True`):
|
requires_safety_checker (`bool`, *optional*, defaults to `True`):
|
||||||
|
|||||||
@ -165,8 +165,8 @@ class LatentConsistencyModelPipeline(
|
|||||||
supports [`LCMScheduler`].
|
supports [`LCMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
requires_safety_checker (`bool`, *optional*, defaults to `True`):
|
requires_safety_checker (`bool`, *optional*, defaults to `True`):
|
||||||
|
|||||||
@ -49,7 +49,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> from diffusers.utils import load_image
|
>>> from diffusers.utils import load_image
|
||||||
|
|
||||||
>>> pipe = LEditsPPPipelineStableDiffusion.from_pretrained(
|
>>> pipe = LEditsPPPipelineStableDiffusion.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", variant="fp16", torch_dtype=torch.float16
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", variant="fp16", torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
>>> pipe.enable_vae_tiling()
|
>>> pipe.enable_vae_tiling()
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
@ -381,8 +381,8 @@ class LEditsPPPipelineStableDiffusion(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -80,7 +80,10 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> # load control net and stable diffusion v1-5
|
>>> # load control net and stable diffusion v1-5
|
||||||
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
>>> controlnet = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny", torch_dtype=torch.float16)
|
||||||
>>> pipe = AutoPipelineForText2Image.from_pretrained(
|
>>> pipe = AutoPipelineForText2Image.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16, enable_pag=True
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5",
|
||||||
|
... controlnet=controlnet,
|
||||||
|
... torch_dtype=torch.float16,
|
||||||
|
... enable_pag=True,
|
||||||
... )
|
... )
|
||||||
|
|
||||||
>>> # speed up diffusion process with faster scheduler and memory optimization
|
>>> # speed up diffusion process with faster scheduler and memory optimization
|
||||||
@ -202,8 +205,8 @@ class StableDiffusionControlNetPAGPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
|
|||||||
@ -93,7 +93,10 @@ EXAMPLE_DOC_STRING = """
|
|||||||
... "lllyasviel/control_v11p_sd15_inpaint", torch_dtype=torch.float16
|
... "lllyasviel/control_v11p_sd15_inpaint", torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
>>> pipe = AutoPipelineForInpainting.from_pretrained(
|
>>> pipe = AutoPipelineForInpainting.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", controlnet=controlnet, torch_dtype=torch.float16, enable_pag=True
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5",
|
||||||
|
... controlnet=controlnet,
|
||||||
|
... torch_dtype=torch.float16,
|
||||||
|
... enable_pag=True,
|
||||||
... )
|
... )
|
||||||
|
|
||||||
>>> pipe.scheduler = DDIMScheduler.from_config(pipe.scheduler.config)
|
>>> pipe.scheduler = DDIMScheduler.from_config(pipe.scheduler.config)
|
||||||
@ -151,10 +154,11 @@ class StableDiffusionControlNetPAGInpaintPipeline(
|
|||||||
- [`~loaders.IPAdapterMixin.load_ip_adapter`] for loading IP Adapters
|
- [`~loaders.IPAdapterMixin.load_ip_adapter`] for loading IP Adapters
|
||||||
|
|
||||||
> [!TIP] > This pipeline can be used with checkpoints that have been specifically fine-tuned for inpainting >
|
> [!TIP] > This pipeline can be used with checkpoints that have been specifically fine-tuned for inpainting >
|
||||||
([runwayml/stable-diffusion-inpainting](https://huggingface.co/runwayml/stable-diffusion-inpainting)) as well as >
|
([stable-diffusion-v1-5/stable-diffusion-inpainting](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-inpainting))
|
||||||
default text-to-image Stable Diffusion checkpoints >
|
as well as > default text-to-image Stable Diffusion checkpoints >
|
||||||
([runwayml/stable-diffusion-v1-5](https://huggingface.co/runwayml/stable-diffusion-v1-5)). Default text-to-image >
|
([stable-diffusion-v1-5/stable-diffusion-v1-5](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5)).
|
||||||
Stable Diffusion checkpoints might be preferable for ControlNets that have been fine-tuned on those, such as >
|
Default text-to-image > Stable Diffusion checkpoints might be preferable for ControlNets that have been fine-tuned
|
||||||
|
on those, such as >
|
||||||
[lllyasviel/control_v11p_sd15_inpaint](https://huggingface.co/lllyasviel/control_v11p_sd15_inpaint).
|
[lllyasviel/control_v11p_sd15_inpaint](https://huggingface.co/lllyasviel/control_v11p_sd15_inpaint).
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
@ -175,8 +179,8 @@ class StableDiffusionControlNetPAGInpaintPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -1328,7 +1332,7 @@ class StableDiffusionControlNetPAGInpaintPipeline(
|
|||||||
|
|
||||||
# 7.1 Check that sizes of mask, masked image and latents match
|
# 7.1 Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -57,7 +57,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> from diffusers import AutoPipelineForText2Image
|
>>> from diffusers import AutoPipelineForText2Image
|
||||||
|
|
||||||
>>> pipe = AutoPipelineForText2Image.from_pretrained(
|
>>> pipe = AutoPipelineForText2Image.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16, enable_pag=True
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", torch_dtype=torch.float16, enable_pag=True
|
||||||
... )
|
... )
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
|
|
||||||
@ -190,8 +190,8 @@ class StableDiffusionPAGPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -272,8 +272,8 @@ class StableDiffusionPAGPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -61,7 +61,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> from diffusers.utils import load_image
|
>>> from diffusers.utils import load_image
|
||||||
|
|
||||||
>>> pipe = AutoPipelineForImage2Image.from_pretrained(
|
>>> pipe = AutoPipelineForImage2Image.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5",
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5",
|
||||||
... torch_dtype=torch.float16,
|
... torch_dtype=torch.float16,
|
||||||
... enable_pag=True,
|
... enable_pag=True,
|
||||||
... )
|
... )
|
||||||
@ -185,8 +185,8 @@ class StableDiffusionPAGImg2ImgPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -267,8 +267,8 @@ class StableDiffusionPAGImg2ImgPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
|
|||||||
@ -58,7 +58,7 @@ EXAMPLE_DOC_STRING = """
|
|||||||
>>> from diffusers import AutoPipelineForInpainting
|
>>> from diffusers import AutoPipelineForInpainting
|
||||||
|
|
||||||
>>> pipe = AutoPipelineForInpainting.from_pretrained(
|
>>> pipe = AutoPipelineForInpainting.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16, enable_pag=True
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", torch_dtype=torch.float16, enable_pag=True
|
||||||
... )
|
... )
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
>>> img_url = "https://raw.githubusercontent.com/CompVis/latent-diffusion/main/data/inpainting_examples/overture-creations-5sI6fQgYIuo.png"
|
>>> img_url = "https://raw.githubusercontent.com/CompVis/latent-diffusion/main/data/inpainting_examples/overture-creations-5sI6fQgYIuo.png"
|
||||||
@ -217,8 +217,8 @@ class StableDiffusionPAGInpaintPipeline(
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -299,8 +299,8 @@ class StableDiffusionPAGInpaintPipeline(
|
|||||||
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
"The configuration file of the unet has set the default `sample_size` to smaller than"
|
||||||
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
" 64 which seems highly unlikely. If your checkpoint is a fine-tuned version of any of the"
|
||||||
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
" following: \n- CompVis/stable-diffusion-v1-4 \n- CompVis/stable-diffusion-v1-3 \n-"
|
||||||
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- runwayml/stable-diffusion-v1-5"
|
" CompVis/stable-diffusion-v1-2 \n- CompVis/stable-diffusion-v1-1 \n- stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
" \n- runwayml/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
" \n- stable-diffusion-v1-5/stable-diffusion-inpainting \n you should change 'sample_size' to 64 in the"
|
||||||
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
" configuration file. Please make sure to update the config accordingly as leaving `sample_size=32`"
|
||||||
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
" in the config might lead to incorrect results in future versions. If you have downloaded this"
|
||||||
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
" checkpoint from the Hugging Face Hub, it would be very nice if you could open a Pull request for"
|
||||||
@ -1183,7 +1183,7 @@ class StableDiffusionPAGInpaintPipeline(
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -1501,7 +1501,7 @@ class StableDiffusionXLPAGInpaintPipeline(
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -179,8 +179,8 @@ class PaintByExamplePipeline(DeprecatedPipelineMixin, DiffusionPipeline, StableD
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`StableDiffusionSafetyChecker`]):
|
safety_checker ([`StableDiffusionSafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
|
|
||||||
|
|||||||
@ -48,8 +48,8 @@ class SemanticStableDiffusionPipeline(DeprecatedPipelineMixin, DiffusionPipeline
|
|||||||
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
[`DDIMScheduler`], [`LMSDiscreteScheduler`], or [`PNDMScheduler`].
|
||||||
safety_checker ([`Q16SafetyChecker`]):
|
safety_checker ([`Q16SafetyChecker`]):
|
||||||
Classification module that estimates whether generated images could be considered offensive or harmful.
|
Classification module that estimates whether generated images could be considered offensive or harmful.
|
||||||
Please refer to the [model card](https://huggingface.co/runwayml/stable-diffusion-v1-5) for more details
|
Please refer to the [model card](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) for
|
||||||
about a model's potential harms.
|
more details about a model's potential harms.
|
||||||
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
feature_extractor ([`~transformers.CLIPImageProcessor`]):
|
||||||
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
A `CLIPImageProcessor` to extract features from generated images; used as inputs to the `safety_checker`.
|
||||||
"""
|
"""
|
||||||
@ -332,7 +332,7 @@ class SemanticStableDiffusionPipeline(DeprecatedPipelineMixin, DiffusionPipeline
|
|||||||
>>> from diffusers import SemanticStableDiffusionPipeline
|
>>> from diffusers import SemanticStableDiffusionPipeline
|
||||||
|
|
||||||
>>> pipe = SemanticStableDiffusionPipeline.from_pretrained(
|
>>> pipe = SemanticStableDiffusionPipeline.from_pretrained(
|
||||||
... "runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16
|
... "stable-diffusion-v1-5/stable-diffusion-v1-5", torch_dtype=torch.float16
|
||||||
... )
|
... )
|
||||||
>>> pipe = pipe.to("cuda")
|
>>> pipe = pipe.to("cuda")
|
||||||
|
|
||||||
|
|||||||
@ -1247,7 +1247,7 @@ class StableDiffusion3InpaintPipeline(DiffusionPipeline, SD3LoraLoaderMixin, Fro
|
|||||||
|
|
||||||
# match the inpainting pipeline and will be updated with input + mask inpainting model later
|
# match the inpainting pipeline and will be updated with input + mask inpainting model later
|
||||||
if num_channels_transformer == 33:
|
if num_channels_transformer == 33:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if (
|
if (
|
||||||
|
|||||||
@ -1501,7 +1501,7 @@ class StableDiffusionXLInpaintPipeline(
|
|||||||
|
|
||||||
# 8. Check that sizes of mask, masked image and latents match
|
# 8. Check that sizes of mask, masked image and latents match
|
||||||
if num_channels_unet == 9:
|
if num_channels_unet == 9:
|
||||||
# default case for runwayml/stable-diffusion-inpainting
|
# default case for stable-diffusion-v1-5/stable-diffusion-inpainting
|
||||||
num_channels_mask = mask.shape[1]
|
num_channels_mask = mask.shape[1]
|
||||||
num_channels_masked_image = masked_image_latents.shape[1]
|
num_channels_masked_image = masked_image_latents.shape[1]
|
||||||
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
if num_channels_latents + num_channels_mask + num_channels_masked_image != self.unet.config.in_channels:
|
||||||
|
|||||||
@ -113,7 +113,8 @@ def load_or_create_model_card(
|
|||||||
|
|
||||||
Args:
|
Args:
|
||||||
repo_id_or_path (`str`):
|
repo_id_or_path (`str`):
|
||||||
The repo id (e.g., "runwayml/stable-diffusion-v1-5") or local path where to look for the model card.
|
The repo id (e.g., "stable-diffusion-v1-5/stable-diffusion-v1-5") or local path where to look for the model
|
||||||
|
card.
|
||||||
token (`str`, *optional*):
|
token (`str`, *optional*):
|
||||||
Authentication token. Will default to the stored token. See https://huggingface.co/settings/token for more
|
Authentication token. Will default to the stored token. See https://huggingface.co/settings/token for more
|
||||||
details.
|
details.
|
||||||
|
|||||||
@ -255,7 +255,7 @@ class StableDiffusionPAGInpaintPipelineFastTests(
|
|||||||
@require_torch_accelerator
|
@require_torch_accelerator
|
||||||
class StableDiffusionPAGPipelineIntegrationTests(unittest.TestCase):
|
class StableDiffusionPAGPipelineIntegrationTests(unittest.TestCase):
|
||||||
pipeline_class = StableDiffusionPAGInpaintPipeline
|
pipeline_class = StableDiffusionPAGInpaintPipeline
|
||||||
repo_id = "runwayml/stable-diffusion-v1-5"
|
repo_id = "stable-diffusion-v1-5/stable-diffusion-v1-5"
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
super().setUp()
|
super().setUp()
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user