Commit Graph

203 Commits

Author SHA1 Message Date
Anthony Mercurio 138cb7bbed
Merge pull request #54 from harubaru/extended-mode
Adding Extended Mode Functionality
2022-11-30 22:55:30 -07:00
Anthony Mercurio 55a555850d
Remove dangling barrier 2022-11-30 22:52:48 -07:00
Anthony Mercurio 7102d313ac
Synchronize ranks for DDP 2022-11-30 19:05:02 -07:00
cafeai 1074bd6f3b Bug Fix 2022-12-01 05:10:21 +09:00
cafeai fb75cbe029 Provide Tokens for Inference 2022-12-01 04:46:01 +09:00
cafeai 981c6ca41a Cleanup 2022-12-01 04:32:10 +09:00
cafeai ee281badcd Extended Mode Updates 2022-12-01 04:31:30 +09:00
Anthony Mercurio 4572617ff9
use ddp for everything 2022-11-30 10:54:30 -07:00
Anthony Mercurio b0cec788be
Get DDP to work 2022-11-29 22:06:21 -07:00
Anthony Mercurio 8decb0bc7d
Fix distributed training 2022-11-29 18:01:17 -07:00
Anthony Mercurio 5f0a952eff
Merge pull request #48 from laksjdjf/sub
fix --ucg not being set.
2022-11-25 15:02:41 -07:00
laksjdjf 5787f7d080
Update diffusers_trainer.py 2022-11-21 13:57:26 +09:00
Anthony Mercurio 511ee9e6d2
Merge pull request #47 from chavinlo/patch-2
Move the model to device BEFORE creating the optimizer
2022-11-20 10:06:50 -05:00
Carlos Chavez f2cfe65d09
Move the movel to device BEFORE creating the optimizer
>It shouldn’t matter, as the optimizer should hold the references to the parameter (even after moving them). However, the “safer” approach would be to move the model to the device first and create the optimizer afterwards.

https://discuss.pytorch.org/t/should-i-create-optimizer-after-sending-the-model-to-gpu/133418/2
https://discuss.pytorch.org/t/effect-of-calling-model-cuda-after-constructing-an-optimizer/15165

At least in my experience with hivemind, if you initialize the optimizer and move the model afterwards, it will throw errors about finding some data in CPU and other on GPU. This shouldn't affect performance or anything I believe.
2022-11-20 00:09:35 -05:00
Anthony Mercurio 1d1f4022d2
Merge pull request #45 from lopho/patch-2
fix wandb init mode, don't log hf token
2022-11-16 16:32:50 -05:00
lopho 9916294de1
fix wandb init mode, don't log hf token
correct value for mode ('enabled' is invalid)
clear hf_token passed to wandb to avoid logging it
2022-11-16 22:28:16 +01:00
Anthony Mercurio c8eeaaf353
Merge pull request #42 from chavinlo/inference-option
Add options and local inference
2022-11-16 16:21:15 -05:00
Anthony Mercurio dc5849b235
Merge branch 'main' into inference-option 2022-11-16 16:20:57 -05:00
chavinlo a2772fc668 fixes 2022-11-16 10:55:38 -05:00
chavinlo fed3431f03 Revert "sync trainer with main branch"
This reverts commit 80e2422967.
2022-11-16 10:44:39 -05:00
Carlos Chavez 80e2422967
sync trainer with main branch 2022-11-16 10:39:20 -05:00
Anthony Mercurio da4f4b93ab
Merge pull request #43 from Maw-Fox/staging-docfix
Minor fix: Documentation Consistency
2022-11-15 13:30:27 -05:00
Maw-Fox 015eeae274
Documentation consistency. 2022-11-15 10:34:55 -07:00
Anthony Mercurio 29ffbd645e
Fix noise scheduler 2022-11-15 11:08:38 -05:00
Anthony Mercurio 5be5a487b2
Merge pull request #39 from Maw-Fox/staging-migration
Implementation of validation/resize classes
2022-11-15 10:33:47 -05:00
Maw-Fox 6c5b2e7149
Fix of fix 2022-11-15 07:15:18 -07:00
Maw-Fox 2c18d29613
Fix from upstream merge. 2022-11-15 06:42:14 -07:00
Maw-Fox b3b5523d85
Merge branch 'staging-migration' of https://github.com/maw-fox/waifu-diffusion into staging-migration 2022-11-14 20:15:15 -07:00
Carlos Chavez d600078008
Add options and local inference
Added options to:
- Disable Inference (it consumes about 2gb of VRAM even when not active)
- Disable wandb

and:
- if no hftoken is provided it just fills it with nothing so it doesn't argues
- if wandb is not enabled, save the inference outputs to a local folder along with information about it
2022-11-14 22:08:16 -05:00
Maw-Fox 773e65f324
Merge origin:main into remote:staging-migration 2022-11-14 19:59:45 -07:00
Anthony Mercurio 5c205524e5
Merge branch 'main' into staging-migration 2022-11-14 12:31:05 -05:00
Anthony Mercurio ae561d19f7
Merge pull request #40 from lopho/patch-1
Parse booleans in argument parser
2022-11-14 12:29:31 -05:00
Maw-Fox 978dd45072
Fix. 2022-11-13 08:24:40 -07:00
Maw-Fox 4943d978c1
Fix redundancies. 2022-11-13 08:22:44 -07:00
Maw-Fox 95b9407a3e
Add+config .gitignore (bring back git stage) and fix up documentation. 2022-11-12 18:48:16 -07:00
Maw-Fox 6bd6c6a4ef
Fixed/flipped help text. 2022-11-12 16:11:30 -07:00
Maw-Fox 189f621a1e
Here, let's fix this while we're at it. 2022-11-12 15:47:17 -07:00
lopho cd0910e82d
Parse booleans in argument parser
true, yes or 1 correspond to True, else False.
2022-11-12 11:14:48 +01:00
Maw-Fox d1eb3ace3f I lied. 2022-11-11 18:17:29 -07:00
Maw-Fox 6c2d5d8066 Final cleanup. 2022-11-11 18:09:09 -07:00
Maw-Fox de221ea42e Derp. ImageStore.__init__ already iterates fully :) 2022-11-11 17:59:32 -07:00
Maw-Fox 925eacf374 Cleanup 2022-11-11 17:50:23 -07:00
Maw-Fox c12cbfced3 Fixed ref typo. 2022-11-11 17:43:09 -07:00
Maw-Fox 6480336d2c Cleanup test code. 2022-11-11 17:17:50 -07:00
Maw-Fox 120d406355 Implementation of validation/resize classes. 2022-11-11 17:14:46 -07:00
Anthony Mercurio 624f0f14af
correctly set args 2022-11-11 08:25:27 -07:00
Anthony Mercurio 033c1e75d0
Merge pull request #36 from harubaru/relicense
Code overhaul & relicense to AGPL 3.0
2022-11-11 07:45:08 -07:00
harubaru 15e70690de Add aesthetic 2022-11-10 13:01:11 -07:00
harubaru 1f5b671b67 relicense 2022-11-10 12:59:53 -07:00
Anthony Mercurio bc626e80e1
Merge pull request #34 from chavinlo/patch-1
typo
2022-11-10 09:23:10 -07:00