fix balancing doc
This commit is contained in:
parent
2512981956
commit
848b6eec71
|
@ -65,6 +65,6 @@ In this case with 200 training images and 1000 preservation images, I would sugg
|
|||
|
||||
## Loss scaling
|
||||
|
||||
Another way to attempt to balance training is to use `loss_scale.txt`. This works similarly to multiply.txt. Place a file called loss_scale.txt in the folder you want to adjust and type a decimal number in the file. A value of 1.0 would be no change. 0.5 will effectively half the learning step size for the images in that folder, and so forth. Negative values technically work, but use extreme caution as my testing shows this can really screw up your model
|
||||
Another way to attempt to balance training is to use `loss_scale.txt`. This works similarly to `multiply.txt`. Place a file called loss_scale.txt in the folder you want to adjust and type a decimal number in the file. A value of 1.0 would be no change. 0.5 will effectively half the learning step size for the images in that folder, and so forth. Negative values technically work, but use extreme caution as my testing shows this can really screw up your model
|
||||
|
||||
This may be an alternative to using `multiply.txt``, but `multiply.txt`` 0.5 will reduce steps because it only chooses 50% of the images per epoch, while `loss_scale.txt` 0.5 will always use all the images but take smaller steps on them. Similarly `multiply.txt` with 2.0 would use the images in that folder twice per epoch whichincreases step count, where loss_scale.txt 2.0 would only use them once but take *larger* learning steps instead performing *more* steps.
|
||||
This may be an alternative to using `multiply.txt`, but `multiply.txt` 0.5 will reduce steps because it only chooses 50% of the images per epoch, while `loss_scale.txt` 0.5 will always use all the images but take smaller steps on them. Similarly `multiply.txt` with 2.0 would use the images in that folder twice per epoch whichincreases step count, where loss_scale.txt 2.0 would only use them once but take *larger* learning steps instead performing *more* steps.
|
Loading…
Reference in New Issue