update docs
This commit is contained in:
parent
da60499728
commit
b3a8a53f38
|
@ -320,7 +320,7 @@ While the calculation makes sense in how it compensates for inteval and total tr
|
|||
|
||||
If you use `ema_strength_target` the actual calculated `ema_decay_rate` used will be printed in your logs, and you should pay attention to this value and use it to inform your future decisions on EMA tuning.
|
||||
|
||||
[Experimental results](https://discord.com/channels/1026983422431862825/1150790432897388556) for EMA on Discord.
|
||||
[Experimental results](https://discord.com/channels/1026983422431862825/1150790432897388556) for general use of EMA on Discord.
|
||||
|
||||
## AdaCoor optimizer
|
||||
|
||||
|
|
|
@ -1,3 +1,15 @@
|
|||
# Synthetic Captioning
|
||||
|
||||
Script now works with the following:
|
||||
|
||||
--model "THUDM/cogvlm-chat-hf"
|
||||
|
||||
--model "THUDM/cogvlm2-llama3-chat-19B"
|
||||
|
||||
--model "xtuner/llava-llama-3-8b-v1_1-transformers"
|
||||
|
||||
--model "THUDM/glm-4v-9b"
|
||||
|
||||
# CogVLM captioning
|
||||
|
||||
CogVLM ([code](https://github.com/THUDM/CogVLM)) ([model](https://huggingface.co/THUDM/cogvlm-chat-hf)) is, so far (Q1 2024), the best model for automatically generating captions.
|
||||
|
|
Loading…
Reference in New Issue