Skip to content

Commit

Permalink
update betas
Browse files Browse the repository at this point in the history
  • Loading branch information
felixdittrich92 committed Dec 20, 2024
1 parent a034ed8 commit d83beca
Show file tree
Hide file tree
Showing 6 changed files with 6 additions and 6 deletions.
2 changes: 1 addition & 1 deletion references/classification/train_pytorch_character.py
Original file line number Diff line number Diff line change
Expand Up @@ -283,7 +283,7 @@ def main(args):
optimizer = torch.optim.Adam(
[p for p in model.parameters() if p.requires_grad],
args.lr,
betas=(0.95, 0.99),
betas=(0.95, 0.999),
eps=1e-6,
weight_decay=args.weight_decay,
)
Expand Down
2 changes: 1 addition & 1 deletion references/classification/train_pytorch_orientation.py
Original file line number Diff line number Diff line change
Expand Up @@ -289,7 +289,7 @@ def main(args):
optimizer = torch.optim.Adam(
[p for p in model.parameters() if p.requires_grad],
args.lr,
betas=(0.95, 0.99),
betas=(0.95, 0.999),
eps=1e-6,
weight_decay=args.weight_decay,
)
Expand Down
2 changes: 1 addition & 1 deletion references/detection/train_pytorch.py
Original file line number Diff line number Diff line change
Expand Up @@ -335,7 +335,7 @@ def main(args):
optimizer = torch.optim.Adam(
[p for p in model.parameters() if p.requires_grad],
args.lr,
betas=(0.95, 0.99),
betas=(0.95, 0.999),
eps=1e-6,
weight_decay=args.weight_decay,
)
Expand Down
2 changes: 1 addition & 1 deletion references/detection/train_pytorch_ddp.py
Original file line number Diff line number Diff line change
Expand Up @@ -351,7 +351,7 @@ def main(rank: int, world_size: int, args):
optimizer = torch.optim.Adam(
[p for p in model.parameters() if p.requires_grad],
args.lr,
betas=(0.95, 0.99),
betas=(0.95, 0.999),
eps=1e-6,
weight_decay=args.weight_decay,
)
Expand Down
2 changes: 1 addition & 1 deletion references/recognition/train_pytorch.py
Original file line number Diff line number Diff line change
Expand Up @@ -341,7 +341,7 @@ def main(args):
optimizer = torch.optim.Adam(
[p for p in model.parameters() if p.requires_grad],
args.lr,
betas=(0.95, 0.99),
betas=(0.95, 0.999),
eps=1e-6,
weight_decay=args.weight_decay,
)
Expand Down
2 changes: 1 addition & 1 deletion references/recognition/train_pytorch_ddp.py
Original file line number Diff line number Diff line change
Expand Up @@ -275,7 +275,7 @@ def main(rank: int, world_size: int, args):
optimizer = torch.optim.Adam(
[p for p in model.parameters() if p.requires_grad],
args.lr,
betas=(0.95, 0.99),
betas=(0.95, 0.999),
eps=1e-6,
weight_decay=args.weight_decay,
)
Expand Down

0 comments on commit d83beca

Please sign in to comment.