File tree Expand file tree Collapse file tree 2 files changed +12
-10
lines changed Expand file tree Collapse file tree 2 files changed +12
-10
lines changed Original file line number Diff line number Diff line change @@ -138,15 +138,15 @@ def setup_optimizer(
138
138
)
139
139
140
140
lr_scheduler : LRScheduler | None = None
141
-
142
- if hyperparameters ["optimizer" ]["lr_scheduler" ] == "reduce_on_plateau" :
143
- lr_scheduler = ReduceLROnPlateau (optimizer = optim , patience = 10 ) # type: ignore
144
-
145
- elif hyperparameters ["optimizer" ]["lr_scheduler" ] == "cosine_annealing" :
146
- lr_scheduler = CosineAnnealingLR (
147
- optimizer = optim ,
148
- T_max = 10 ,
149
- )
141
+ if "lr_scheduler" in hyperparameters [ "optimizer" ]:
142
+ if hyperparameters ["optimizer" ]["lr_scheduler" ] == "reduce_on_plateau" :
143
+ lr_scheduler = ReduceLROnPlateau (optimizer = optim , patience = 10 ) # type: ignore
144
+
145
+ elif hyperparameters ["optimizer" ]["lr_scheduler" ] == "cosine_annealing" :
146
+ lr_scheduler = CosineAnnealingLR (
147
+ optimizer = optim ,
148
+ T_max = 10 ,
149
+ )
150
150
151
151
return optim , lr_scheduler
152
152
Original file line number Diff line number Diff line change @@ -152,7 +152,9 @@ class OptimizerHyperparameters(NestedParameter):
152
152
Whether to use a fused implementation of the optimizer (may be faster on CUDA).
153
153
"""
154
154
155
- lr_scheduler : Parameter [Literal ["reduce_on_plateau" ]] | None = None
155
+ lr_scheduler : Parameter [Literal ["reduce_on_plateau" , "cosine_annealing" ]] | None = field (
156
+ default = None
157
+ )
156
158
"""Learning rate scheduler."""
157
159
158
160
You can’t perform that action at this time.
0 commit comments