def get_lr_lambda(step, warm_up_steps: int, reduce_lr_steps: int): r"""Get lr_lambda for LambdaLR. E.g., .. code-block: python lr_lambda = lambda step: get_lr_lambda(step, warm_up_steps=1000, reduce_lr_steps=10000) from torch.optim.lr_scheduler import LambdaLR LambdaLR(optimizer, lr_lambda) Args: warm_up_steps: int, steps for warm up reduce_lr_steps: int, reduce learning rate by 0.9 every #reduce_lr_steps steps Returns: learning rate: float """ if step <= warm_up_steps: return step / warm_up_steps else: return 0.9 ** (step // reduce_lr_steps)