-
Notifications
You must be signed in to change notification settings - Fork 66
/
main_BERT.py
90 lines (85 loc) · 3.43 KB
/
main_BERT.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
from tools import BERT_pretrain_run_net as pretrain
from tools import BERT_finetune_run_net as finetune
from tools import BERT_test_run_net as test_net
from utils import parser, dist_utils, misc
from utils.logger import *
from utils.config import *
import time
import os
import torch
from tensorboardX import SummaryWriter
def main():
# args
args = parser.get_args()
# CUDA
args.use_gpu = torch.cuda.is_available()
if args.use_gpu:
torch.backends.cudnn.benchmark = True
# init distributed env first, since logger depends on the dist info.
if args.launcher == 'none':
args.distributed = False
else:
args.distributed = True
dist_utils.init_dist(args.launcher)
# re-set gpu_ids with distributed training mode
_, world_size = dist_utils.get_dist_info()
args.world_size = world_size
# logger
timestamp = time.strftime('%Y%m%d_%H%M%S', time.localtime())
log_file = os.path.join(args.experiment_path, f'{timestamp}.log')
logger = get_root_logger(log_file=log_file, name=args.log_name)
# define the tensorboard writer
if not args.test:
if args.local_rank == 0:
train_writer = SummaryWriter(os.path.join(args.tfboard_path, 'train'))
val_writer = SummaryWriter(os.path.join(args.tfboard_path, 'test'))
else:
train_writer = None
val_writer = None
# config
config = get_config(args, logger = logger)
# batch size
if args.distributed:
assert config.total_bs % world_size == 0
config.dataset.train.others.bs = config.total_bs // world_size
if config.dataset.get('extra_train'):
config.dataset.extra_train.others.bs = config.total_bs // world_size * 2
config.dataset.val.others.bs = config.total_bs // world_size * 2
if config.dataset.get('test'):
config.dataset.test.others.bs = config.total_bs // world_size
else:
config.dataset.train.others.bs = config.total_bs
if config.dataset.get('extra_train'):
config.dataset.extra_train.others.bs = config.total_bs * 2
config.dataset.val.others.bs = config.total_bs * 2
if config.dataset.get('test'):
config.dataset.test.others.bs = config.total_bs
# log
log_args_to_file(args, 'args', logger = logger)
log_config_to_file(config, 'config', logger = logger)
# exit()
logger.info(f'Distributed training: {args.distributed}')
# set random seeds
if args.seed is not None:
logger.info(f'Set random seed to {args.seed}, '
f'deterministic: {args.deterministic}')
misc.set_random_seed(args.seed + args.local_rank, deterministic=args.deterministic) # seed + rank, for augmentation
if args.distributed:
assert args.local_rank == torch.distributed.get_rank()
if args.shot != -1:
config.dataset.train.others.shot = args.shot
config.dataset.train.others.way = args.way
config.dataset.train.others.fold = args.fold
config.dataset.val.others.shot = args.shot
config.dataset.val.others.way = args.way
config.dataset.val.others.fold = args.fold
# run
if args.test:
test_net(args, config)
else:
if args.finetune_model or args.scratch_model:
finetune(args, config, train_writer, val_writer)
else:
pretrain(args, config, train_writer, val_writer)
if __name__ == '__main__':
main()