commit
bed32182f6
|
@ -27,7 +27,7 @@ More precisely, 🤗 Diffusers offers:
|
||||||
|
|
||||||
## Definitions
|
## Definitions
|
||||||
|
|
||||||
**Models**: Neural network that models **p_θ(x_t-1|x_t)** (see image below) and is trained end-to-end to *denoise* a noisy input to an image.
|
**Models**: Neural network that models $p_\theta(\mathbf{x}_{t-1}|\mathbf{x}_t)$ (see image below) and is trained end-to-end to *denoise* a noisy input to an image.
|
||||||
*Examples*: UNet, Conditioned UNet, 3D UNet, Transformer UNet
|
*Examples*: UNet, Conditioned UNet, 3D UNet, Transformer UNet
|
||||||
|
|
||||||
![model_diff_1_50](https://user-images.githubusercontent.com/23423619/171610307-dab0cd8b-75da-4d4e-9f5a-5922072e2bb5.png)
|
![model_diff_1_50](https://user-images.githubusercontent.com/23423619/171610307-dab0cd8b-75da-4d4e-9f5a-5922072e2bb5.png)
|
||||||
|
@ -44,7 +44,6 @@ The class provides functionality to compute previous image according to alpha, b
|
||||||
|
|
||||||
![imagen](https://user-images.githubusercontent.com/23423619/171609001-c3f2c1c9-f597-4a16-9843-749bf3f9431c.png)
|
![imagen](https://user-images.githubusercontent.com/23423619/171609001-c3f2c1c9-f597-4a16-9843-749bf3f9431c.png)
|
||||||
|
|
||||||
|
|
||||||
## Philosophy
|
## Philosophy
|
||||||
|
|
||||||
- Readability and clarity is prefered over highly optimized code. A strong importance is put on providing readable, intuitive and elementary code design. *E.g.*, the provided [schedulers](https://github.com/huggingface/diffusers/tree/main/src/diffusers/schedulers) are separated from the provided [models](https://github.com/huggingface/diffusers/tree/main/src/diffusers/models) and provide well-commented code that can be read alongside the original paper.
|
- Readability and clarity is prefered over highly optimized code. A strong importance is put on providing readable, intuitive and elementary code design. *E.g.*, the provided [schedulers](https://github.com/huggingface/diffusers/tree/main/src/diffusers/schedulers) are separated from the provided [models](https://github.com/huggingface/diffusers/tree/main/src/diffusers/models) and provide well-commented code that can be read alongside the original paper.
|
||||||
|
@ -59,7 +58,7 @@ The class provides functionality to compute previous image according to alpha, b
|
||||||
pip install diffusers # should install diffusers 0.0.4
|
pip install diffusers # should install diffusers 0.0.4
|
||||||
```
|
```
|
||||||
|
|
||||||
### 1. `diffusers` as a toolbox for schedulers and models.
|
### 1. `diffusers` as a toolbox for schedulers and models
|
||||||
|
|
||||||
`diffusers` is more modularized than `transformers`. The idea is that researchers and engineers can use only parts of the library easily for the own use cases.
|
`diffusers` is more modularized than `transformers`. The idea is that researchers and engineers can use only parts of the library easily for the own use cases.
|
||||||
It could become a central place for all kinds of models, schedulers, training utils and processors that one can mix and match for one's own use case.
|
It could become a central place for all kinds of models, schedulers, training utils and processors that one can mix and match for one's own use case.
|
||||||
|
@ -149,7 +148,7 @@ eta = 0.0 # <- deterministic sampling
|
||||||
for t in tqdm.tqdm(reversed(range(num_inference_steps)), total=num_inference_steps):
|
for t in tqdm.tqdm(reversed(range(num_inference_steps)), total=num_inference_steps):
|
||||||
# 1. predict noise residual
|
# 1. predict noise residual
|
||||||
orig_t = noise_scheduler.get_orig_t(t, num_inference_steps)
|
orig_t = noise_scheduler.get_orig_t(t, num_inference_steps)
|
||||||
with torch.no_grad():
|
with torch.inference_mode():
|
||||||
residual = unet(image, orig_t)
|
residual = unet(image, orig_t)
|
||||||
|
|
||||||
# 2. predict previous mean of image x_t-1
|
# 2. predict previous mean of image x_t-1
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
# Models
|
# Models
|
||||||
|
|
||||||
- Models: Neural network that models p_θ(x_t-1|x_t) (see image below) and is trained end-to-end to denoise a noisy input to an image. Examples: UNet, Conditioned UNet, 3D UNet, Transformer UNet
|
- Models: Neural network that models $p_\theta(\mathbf{x}_{t-1}|\mathbf{x}_t)$ (see image below) and is trained end-to-end to denoise a noisy input to an image. Examples: UNet, Conditioned UNet, 3D UNet, Transformer UNet
|
||||||
|
|
||||||
## API
|
## API
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue