From d83becab7c65f10b376c396e7f725be1fdc85a58 Mon Sep 17 00:00:00 2001 From: felix Date: Fri, 20 Dec 2024 10:21:07 +0100 Subject: [PATCH] update betas --- references/classification/train_pytorch_character.py | 2 +- references/classification/train_pytorch_orientation.py | 2 +- references/detection/train_pytorch.py | 2 +- references/detection/train_pytorch_ddp.py | 2 +- references/recognition/train_pytorch.py | 2 +- references/recognition/train_pytorch_ddp.py | 2 +- 6 files changed, 6 insertions(+), 6 deletions(-) diff --git a/references/classification/train_pytorch_character.py b/references/classification/train_pytorch_character.py index 2795f094c..bdca07a52 100644 --- a/references/classification/train_pytorch_character.py +++ b/references/classification/train_pytorch_character.py @@ -283,7 +283,7 @@ def main(args): optimizer = torch.optim.Adam( [p for p in model.parameters() if p.requires_grad], args.lr, - betas=(0.95, 0.99), + betas=(0.95, 0.999), eps=1e-6, weight_decay=args.weight_decay, ) diff --git a/references/classification/train_pytorch_orientation.py b/references/classification/train_pytorch_orientation.py index 939deb6f7..aff149dea 100644 --- a/references/classification/train_pytorch_orientation.py +++ b/references/classification/train_pytorch_orientation.py @@ -289,7 +289,7 @@ def main(args): optimizer = torch.optim.Adam( [p for p in model.parameters() if p.requires_grad], args.lr, - betas=(0.95, 0.99), + betas=(0.95, 0.999), eps=1e-6, weight_decay=args.weight_decay, ) diff --git a/references/detection/train_pytorch.py b/references/detection/train_pytorch.py index 5c66cd292..75288d7f1 100644 --- a/references/detection/train_pytorch.py +++ b/references/detection/train_pytorch.py @@ -335,7 +335,7 @@ def main(args): optimizer = torch.optim.Adam( [p for p in model.parameters() if p.requires_grad], args.lr, - betas=(0.95, 0.99), + betas=(0.95, 0.999), eps=1e-6, weight_decay=args.weight_decay, ) diff --git a/references/detection/train_pytorch_ddp.py b/references/detection/train_pytorch_ddp.py index 36c7a16c4..e2d9ab014 100644 --- a/references/detection/train_pytorch_ddp.py +++ b/references/detection/train_pytorch_ddp.py @@ -351,7 +351,7 @@ def main(rank: int, world_size: int, args): optimizer = torch.optim.Adam( [p for p in model.parameters() if p.requires_grad], args.lr, - betas=(0.95, 0.99), + betas=(0.95, 0.999), eps=1e-6, weight_decay=args.weight_decay, ) diff --git a/references/recognition/train_pytorch.py b/references/recognition/train_pytorch.py index 6a6a07037..f4123acc6 100644 --- a/references/recognition/train_pytorch.py +++ b/references/recognition/train_pytorch.py @@ -341,7 +341,7 @@ def main(args): optimizer = torch.optim.Adam( [p for p in model.parameters() if p.requires_grad], args.lr, - betas=(0.95, 0.99), + betas=(0.95, 0.999), eps=1e-6, weight_decay=args.weight_decay, ) diff --git a/references/recognition/train_pytorch_ddp.py b/references/recognition/train_pytorch_ddp.py index d8b2e7a8e..f62b09957 100644 --- a/references/recognition/train_pytorch_ddp.py +++ b/references/recognition/train_pytorch_ddp.py @@ -275,7 +275,7 @@ def main(rank: int, world_size: int, args): optimizer = torch.optim.Adam( [p for p in model.parameters() if p.requires_grad], args.lr, - betas=(0.95, 0.99), + betas=(0.95, 0.999), eps=1e-6, weight_decay=args.weight_decay, )