Skip to content

Commit d9eaa21

Browse files
committed
Package updates
1 parent 9c648cf commit d9eaa21

File tree

3 files changed

+17
-18
lines changed

3 files changed

+17
-18
lines changed

README.md

Lines changed: 16 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
Wrapper for the official [Stable Diffusion](https://github.com/Stability-AI/stablediffusion) repository, to allow installing via `pip`. Please see the installation section below for more details.
22

33
# Stable Diffusion Version 2
4-
![t2i](assets/stable-samples/txt2img/768/merged-0006.png)
5-
![t2i](assets/stable-samples/txt2img/768/merged-0002.png)
6-
![t2i](assets/stable-samples/txt2img/768/merged-0005.png)
4+
![t2i](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/txt2img/768/merged-0006.png)
5+
![t2i](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/txt2img/768/merged-0002.png)
6+
![t2i](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/txt2img/768/merged-0005.png)
77

88
This repository contains [Stable Diffusion](https://github.com/CompVis/stable-diffusion) models trained from scratch and will be continuously updated with
99
new checkpoints. The following list provides an overview of all currently available models. More coming soon.
@@ -14,6 +14,8 @@ new checkpoints. The following list provides an overview of all currently availa
1414

1515
Step 1 is not necessary on Mac.
1616

17+
This will install the `ldm` package, which contains the stable diffusion code.
18+
1719
## News
1820

1921
**December 7, 2022**
@@ -32,7 +34,7 @@ Per default, the attention operation of the model is evaluated at full precision
3234
- Added a [x4 upscaling latent text-guided diffusion model](#image-upscaling-with-stable-diffusion).
3335
- New [depth-guided stable diffusion model](#depth-conditional-stable-diffusion), finetuned from _SD 2.0-base_. The model is conditioned on monocular depth estimates inferred via [MiDaS](https://github.com/isl-org/MiDaS) and can be used for structure-preserving img2img and shape-conditional synthesis.
3436

35-
![d2i](assets/stable-samples/depth2img/depth2img01.png)
37+
![d2i](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/depth2img/depth2img01.png)
3638
- A [text-guided inpainting model](#image-inpainting-with-stable-diffusion), finetuned from SD _2.0-base_.
3739

3840
We follow the [original repository](https://github.com/CompVis/stable-diffusion) and provide basic inference scripts to sample from the models.
@@ -109,13 +111,13 @@ and OpenCLIP ViT-H/14 text encoder for the diffusion model. The _SD 2-v_ model p
109111
Evaluations with different classifier-free guidance scales (1.5, 2.0, 3.0, 4.0,
110112
5.0, 6.0, 7.0, 8.0) and 50 DDIM sampling steps show the relative improvements of the checkpoints:
111113

112-
![sd evaluation results](assets/model-variants.jpg)
114+
![sd evaluation results](https://github.com/Stability-AI/stablediffusion/raw/main/assets/model-variants.jpg)
113115

114116

115117

116118
### Text-to-Image
117-
![txt2img-stable2](assets/stable-samples/txt2img/merged-0003.png)
118-
![txt2img-stable2](assets/stable-samples/txt2img/merged-0001.png)
119+
![txt2img-stable2](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/txt2img/merged-0003.png)
120+
![txt2img-stable2](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/txt2img/merged-0001.png)
119121

120122
Stable Diffusion 2 is a latent diffusion model conditioned on the penultimate text embeddings of a CLIP ViT-H/14 text encoder.
121123
We provide a [reference script for sampling](#reference-sampling-script).
@@ -147,7 +149,7 @@ non-EMA to EMA weights.
147149

148150
### Image Modification with Stable Diffusion
149151

150-
![depth2img-stable2](assets/stable-samples/depth2img/merged-0000.png)
152+
![depth2img-stable2](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/depth2img/merged-0000.png)
151153
#### Depth-Conditional Stable Diffusion
152154

153155
To augment the well-established [img2img](https://github.com/CompVis/stable-diffusion#image-modification-with-stable-diffusion) functionality of Stable Diffusion, we provide a _shape-preserving_ stable diffusion model.
@@ -166,19 +168,19 @@ streamlit run scripts/streamlit/depth2img.py configs/stable-diffusion/v2-midas-i
166168
```
167169

168170
This method can be used on the samples of the base model itself.
169-
For example, take [this sample](assets/stable-samples/depth2img/old_man.png) generated by an anonymous discord user.
171+
For example, take [this sample](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/depth2img/old_man.png) generated by an anonymous discord user.
170172
Using the [gradio](https://gradio.app) or [streamlit](https://streamlit.io/) script `depth2img.py`, the MiDaS model first infers a monocular depth estimate given this input,
171173
and the diffusion model is then conditioned on the (relative) depth output.
172174

173175
<p align="center">
174176
<b> depth2image </b><br/>
175-
<img src=assets/stable-samples/depth2img/d2i.gif>
177+
<img src=https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/depth2img/d2i.gif>
176178
</p>
177179

178-
This model is particularly useful for a photorealistic style; see the [examples](assets/stable-samples/depth2img).
180+
This model is particularly useful for a photorealistic style; see the [examples](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/depth2img).
179181
For a maximum strength of 1.0, the model removes all pixel-based information and only relies on the text prompt and the inferred monocular depth estimate.
180182

181-
![depth2img-stable3](assets/stable-samples/depth2img/merged-0005.png)
183+
![depth2img-stable3](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/depth2img/merged-0005.png)
182184

183185
#### Classic Img2Img
184186

@@ -189,7 +191,7 @@ python scripts/img2img.py --prompt "A fantasy landscape, trending on artstation"
189191
and adapt the checkpoint and config paths accordingly.
190192

191193
### Image Upscaling with Stable Diffusion
192-
![upscaling-x4](assets/stable-samples/upscaling/merged-dog.png)
194+
![upscaling-x4](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-samples/upscaling/merged-dog.png)
193195
After [downloading the weights](https://huggingface.co/stabilityai/stable-diffusion-x4-upscaler), run
194196
```
195197
python scripts/gradio/superresolution.py configs/stable-diffusion/x4-upscaling.yaml <path-to-checkpoint>
@@ -207,7 +209,7 @@ This model can be used both on real inputs and on synthesized examples. For the
207209

208210
### Image Inpainting with Stable Diffusion
209211

210-
![inpainting-stable2](assets/stable-inpainting/merged-leopards.png)
212+
![inpainting-stable2](https://github.com/Stability-AI/stablediffusion/raw/main/assets/stable-inpainting/merged-leopards.png)
211213

212214
[Download the SD 2.0-inpainting checkpoint](https://huggingface.co/stabilityai/stable-diffusion-2-inpainting) and run
213215

pyproject.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
44

55
[project]
66
name = "stable-diffusion-sdkit"
7-
version = "2.1.1"
7+
version = "2.1.2"
88
description = "High-Resolution Image Synthesis with Latent Diffusion Models. This is a wrapper around the original repo, to allow installing via pip."
99
readme = "README.md"
1010
requires-python = ">=3.8"

setup.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,9 +22,6 @@
2222
# They need to run: pip install torch torchvision --extra-index-url https://download.pytorch.org/whl/cu116
2323

2424
setup(
25-
name='stable-diffusion-sdkit',
26-
version='2.1.0',
27-
description='',
2825
packages=find_packages(),
2926
install_requires=dependencies,
3027
)

0 commit comments

Comments
 (0)