fix balancing doc

This commit is contained in:
Victor Hall 2023-11-03 13:51:35 -04:00
parent 2512981956
commit 848b6eec71
1 changed files with 2 additions and 2 deletions

View File

@ -65,6 +65,6 @@ In this case with 200 training images and 1000 preservation images, I would sugg
## Loss scaling
Another way to attempt to balance training is to use `loss_scale.txt`. This works similarly to multiply.txt. Place a file called loss_scale.txt in the folder you want to adjust and type a decimal number in the file. A value of 1.0 would be no change. 0.5 will effectively half the learning step size for the images in that folder, and so forth. Negative values technically work, but use extreme caution as my testing shows this can really screw up your model
Another way to attempt to balance training is to use `loss_scale.txt`. This works similarly to `multiply.txt`. Place a file called loss_scale.txt in the folder you want to adjust and type a decimal number in the file. A value of 1.0 would be no change. 0.5 will effectively half the learning step size for the images in that folder, and so forth. Negative values technically work, but use extreme caution as my testing shows this can really screw up your model
This may be an alternative to using `multiply.txt``, but `multiply.txt`` 0.5 will reduce steps because it only chooses 50% of the images per epoch, while `loss_scale.txt` 0.5 will always use all the images but take smaller steps on them. Similarly `multiply.txt` with 2.0 would use the images in that folder twice per epoch whichincreases step count, where loss_scale.txt 2.0 would only use them once but take *larger* learning steps instead performing *more* steps.
This may be an alternative to using `multiply.txt`, but `multiply.txt` 0.5 will reduce steps because it only chooses 50% of the images per epoch, while `loss_scale.txt` 0.5 will always use all the images but take smaller steps on them. Similarly `multiply.txt` with 2.0 would use the images in that folder twice per epoch whichincreases step count, where loss_scale.txt 2.0 would only use them once but take *larger* learning steps instead performing *more* steps.