We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
2 parents 727453f + ee0ee94 commit 99c673dCopy full SHA for 99c673d
1 file changed
src/invoke_training/pipelines/flux/lora/config.py
@@ -55,7 +55,7 @@ class FluxLoraConfig(BasePipelineConfig):
55
rate.
56
"""
57
58
- transformer_learning_rate: float | None = 1e-4
+ transformer_learning_rate: float | None = 4e-4
59
"""The learning rate to use for the transformer model. If set, this overrides the optimizer's default learning
60
61
0 commit comments