# @package _global_
train:
  scheduler_interval: epoch
  scheduler_monitor: ???
  scheduler:
    _target_: torch.optim.lr_scheduler.ReduceLROnPlateau
    factor: 0.2  # Decay factor when ReduceLROnPlateau is used
    patience: 20
    min_lr: 0.0  # Minimum learning rate during annealing