# optimizer
# optimizer = dict(type="SGD", lr=1.25e-4, momentum=0.9, weight_decay=5e-4)
optimizer = dict(type="Adam", lr=1.25e-4, weight_decay=5e-4)
# optimizer_config = dict(grad_clip=dict(max_norm=35, norm_type=2))
optimizer_config = dict(grad_clip=None)
# learning policy
lr_config = dict(policy="step",
                 warmup="linear",
                 warmup_iters=100,
                 warmup_ratio=1.25e-4,
                 step=[45, 60])
# runtime settings
total_epochs = 70
