Commit Graph

16 Commits

Author SHA1 Message Date
anton-l a73ae3e5b0 Better default for AdamW 2022-07-21 13:36:16 +02:00
anton-l 06505ba4b4 Less eval steps during training 2022-07-21 11:47:40 +02:00
anton-l 302b86bd0b Adapt training to the new UNet API 2022-07-21 11:07:21 +02:00
Anton Lozhkov 76f9b52289
Update the training examples (#102)
* New unet, gradient accumulation

* Save every n epochs

* Remove find_unused_params, hooray!

* Update examples

* Switch to DDPM completely
2022-07-20 19:51:23 +02:00
Anton Lozhkov d9316bf8bc
Fix mutable proj_out weight in the Attention layer (#73)
* Catch unused params in DDP

* Fix proj_out, add test
2022-07-04 12:36:37 +02:00
Tanishq Abraham 3abf4bc439
EMA model stepping updated to keep track of current step (#64)
ema model stepping done automatically now
2022-07-04 11:53:15 +02:00
Anton Lozhkov 8cba133f36
Add the model card template (#43)
* add a metrics logger

* fix LatentDiffusionUncondPipeline

* add VQModel in init

* add image logging to tensorboard

* switch manual templates to the modelcards package

* hide ldm example

Co-authored-by: patil-suraj <surajp815@gmail.com>
2022-06-29 15:37:23 +02:00
Patrick von Platen 932ce05d97 cancel einops 2022-06-27 15:39:41 +00:00
anton-l 1cf7933ea2 Framework-agnostic timestep broadcasting 2022-06-27 17:11:01 +02:00
anton-l 3f9e3d8ad6 add EMA during training 2022-06-27 15:23:01 +02:00
anton-l 848c86ca0a batched forward diffusion step 2022-06-22 13:38:14 +02:00
anton-l 9e31c6a749 refactor GLIDE text2im pipeline, remove classifier_free_guidance 2022-06-21 14:07:58 +02:00
anton-l 71289ba06e add lr schedule utils 2022-06-21 11:35:56 +02:00
anton-l 0417baf23d additional hub arguments 2022-06-21 11:21:10 +02:00
anton-l 9c82c32ba7 make style 2022-06-21 10:43:40 +02:00
anton-l a2117cb797 add push_to_hub 2022-06-21 10:38:34 +02:00