def __init__(self, size): """ Args: size (int): the total number of data of the underlying dataset to sample from """ self._size = size assert size > 0 self._rank = comm.get_rank() self._world_size = comm.get_world_size() shard_size = (self._size - 1) // self._world_size + 1 begin = shard_size * self._rank end = min(shard_size * (self._rank + 1), self._size) self._local_indices = range(begin, end)
def __init__(self, size, shuffle=True, seed=None): """ Args: size (int): the total number of data of the underlying dataset to sample from shuffle (bool): whether to shuffle the indices or not seed (int): the initial seed of the shuffle. Must be the same across all workers. If None, will use a random seed shared among workers (require synchronization among all workers). """ self._size = size assert size > 0 self._shuffle = shuffle if seed is None: seed = comm.shared_random_seed() self._seed = int(seed) self._rank = comm.get_rank() self._world_size = comm.get_world_size()
def main(): args = parse_args() logger = logging.getLogger('segmentation') if not logger.isEnabledFor(logging.INFO): # setup_logger is not called setup_logger(output=config.OUTPUT_DIR, distributed_rank=args.local_rank) logger.info(pprint.pformat(args)) logger.info(config) # cudnn related setting cudnn.benchmark = config.CUDNN.BENCHMARK cudnn.deterministic = config.CUDNN.DETERMINISTIC cudnn.enabled = config.CUDNN.ENABLED gpus = list(config.GPUS) distributed = len(gpus) > 1 device = torch.device('cuda:{}'.format(args.local_rank)) if distributed: torch.cuda.set_device(args.local_rank) torch.distributed.init_process_group( backend="nccl", init_method="env://", ) # build model model = build_segmentation_model_from_cfg(config) logger.info("Model:\n{}".format(model)) logger.info("Rank of current process: {}. World size: {}".format(comm.get_rank(), comm.get_world_size())) if distributed: model = nn.SyncBatchNorm.convert_sync_batchnorm(model) model = model.to(device) if comm.get_world_size() > 1: model = DistributedDataParallel( model, device_ids=[args.local_rank], output_device=args.local_rank ) data_loader = build_train_loader_from_cfg(config) optimizer = build_optimizer(config, model) lr_scheduler = build_lr_scheduler(config, optimizer) data_loader_iter = iter(data_loader) start_iter = 0 max_iter = config.TRAIN.MAX_ITER best_param_group_id = get_lr_group_id(optimizer) # initialize model if os.path.isfile(config.MODEL.WEIGHTS): model_weights = torch.load(config.MODEL.WEIGHTS) get_module(model, distributed).load_state_dict(model_weights, strict=False) logger.info('Pre-trained model from {}'.format(config.MODEL.WEIGHTS)) elif not config.MODEL.BACKBONE.PRETRAINED: if os.path.isfile(config.MODEL.BACKBONE.WEIGHTS): pretrained_weights = torch.load(config.MODEL.BACKBONE.WEIGHTS) get_module(model, distributed).backbone.load_state_dict(pretrained_weights, strict=False) logger.info('Pre-trained backbone from {}'.format(config.MODEL.BACKBONE.WEIGHTS)) else: logger.info('No pre-trained weights for backbone, training from scratch.') # load model if config.TRAIN.RESUME: model_state_file = os.path.join(config.OUTPUT_DIR, 'checkpoint.pth.tar') if os.path.isfile(model_state_file): checkpoint = torch.load(model_state_file) start_iter = checkpoint['start_iter'] get_module(model, distributed).load_state_dict(checkpoint['state_dict']) optimizer.load_state_dict(checkpoint['optimizer']) lr_scheduler.load_state_dict(checkpoint['lr_scheduler']) logger.info('Loaded checkpoint (starting from iter {})'.format(checkpoint['start_iter'])) data_time = AverageMeter() batch_time = AverageMeter() loss_meter = AverageMeter() # Debug output. if config.DEBUG.DEBUG: debug_out_dir = os.path.join(config.OUTPUT_DIR, 'debug_train') PathManager.mkdirs(debug_out_dir) # Train loop. try: for i in range(start_iter, max_iter): # data start_time = time.time() data = next(data_loader_iter) if not distributed: data = to_cuda(data, device) _data_time = time.time() data_time.update(_data_time - start_time) image = data.pop('image') out_dict = model(image, data) loss = out_dict['loss'] torch.cuda.synchronize(device) _forward_time = time.time() if args.gpumem: gpumem = torch.cuda.memory_allocated(device) peak_usage = torch.cuda.max_memory_allocated(device) torch.cuda.reset_peak_memory_stats(device) optimizer.zero_grad() loss.backward() optimizer.step() # Get lr. lr = optimizer.param_groups[best_param_group_id]["lr"] lr_scheduler.step() _batch_time = time.time() batch_time.update(_batch_time - start_time) loss_meter.update(loss.detach().cpu().item(), image.size(0)) if args.timing: logger.info('timing - forward %f' % (_forward_time - _data_time)) logger.info('timing - both %f' % (_batch_time - _data_time)) if args.gpumem: logger.info('gpumem - %f' % gpumem) logger.info('gpumem - peak %f' % peak_usage) if i == 0 or (i + 1) % config.PRINT_FREQ == 0: msg = '[{0}/{1}] LR: {2:.7f}\t' \ 'Time: {batch_time.val:.3f}s ({batch_time.avg:.3f}s)\t' \ 'Data: {data_time.val:.3f}s ({data_time.avg:.3f}s)\t'.format( i + 1, max_iter, lr, batch_time=batch_time, data_time=data_time) msg += get_loss_info_str(get_module(model, distributed).loss_meter_dict) logger.info(msg) if i == 0 or (i + 1) % config.DEBUG.DEBUG_FREQ == 0: # TODO: Add interface for save_debug_images # if comm.is_main_process() and config.DEBUG.DEBUG: # save_debug_images( # dataset=data_loader.dataset, # batch_images=image, # batch_targets=data, # batch_outputs=out_dict, # out_dir=debug_out_dir, # iteration=i, # target_keys=config.DEBUG.TARGET_KEYS, # output_keys=config.DEBUG.OUTPUT_KEYS, # iteration_to_remove=i - config.DEBUG.KEEP_INTERVAL # ) if i>0 and (args.gpumem or args.timing): break if i == 0 or (i + 1) % config.CKPT_FREQ == 0: if comm.is_main_process(): torch.save({ 'start_iter': i + 1, 'state_dict': get_module(model, distributed).state_dict(), 'optimizer': optimizer.state_dict(), 'lr_scheduler': lr_scheduler.state_dict(), }, os.path.join(config.OUTPUT_DIR, 'checkpoint.pth.tar')) except Exception: logger.exception("Exception during training:") raise finally: if comm.is_main_process(): torch.save(get_module(model, distributed).state_dict(), os.path.join(config.OUTPUT_DIR, 'final_state.pth')) logger.info("Training finished.")
def main(): args = parse_args() logger = logging.getLogger('segmentation') if not logger.isEnabledFor(logging.INFO): # setup_logger is not called setup_logger(output=config.OUTPUT_DIR, distributed_rank=args.local_rank) # logger.info(pprint.pformat(args)) # logger.info(config) # cudnn related setting cudnn.benchmark = config.CUDNN.BENCHMARK cudnn.deterministic = config.CUDNN.DETERMINISTIC cudnn.enabled = config.CUDNN.ENABLED gpus = list(config.GPUS) distributed = len(gpus) > 1 device = torch.device('cuda:{}'.format(args.local_rank)) if distributed: torch.cuda.set_device(args.local_rank) torch.distributed.init_process_group( backend="nccl", init_method="env://", ) # build model model = build_segmentation_model_from_cfg(config) # logger.info("Model:\n{}".format(model)) logger.info("Rank of current process: {}. World size: {}".format(comm.get_rank(), comm.get_world_size())) if distributed: model = nn.SyncBatchNorm.convert_sync_batchnorm(model) model = model.to(device) if comm.get_world_size() > 1: model = DistributedDataParallel( model, device_ids=[args.local_rank], output_device=args.local_rank ) data_loader = build_train_loader_from_cfg(config) optimizer = build_optimizer(config, model) lr_scheduler = build_lr_scheduler(config, optimizer) data_loader_iter = iter(data_loader) start_iter = 0 max_iter = config.TRAIN.MAX_ITER best_param_group_id = get_lr_group_id(optimizer) # initialize model if os.path.isfile(config.MODEL.WEIGHTS): model_weights = torch.load(config.MODEL.WEIGHTS) get_module(model, distributed).load_state_dict(model_weights, strict=False) logger.info('Pre-trained model from {}'.format(config.MODEL.WEIGHTS)) elif config.MODEL.BACKBONE.PRETRAINED: if os.path.isfile(config.MODEL.BACKBONE.WEIGHTS): pretrained_weights = torch.load(config.MODEL.BACKBONE.WEIGHTS) get_module(model, distributed).backbone.load_state_dict(pretrained_weights, strict=False) logger.info('Pre-trained backbone from {}'.format(config.MODEL.BACKBONE.WEIGHTS)) else: logger.info('No pre-trained weights for backbone, training from scratch.') # load model if config.TRAIN.RESUME: model_state_file = os.path.join(config.OUTPUT_DIR, 'checkpoint.pth.tar') if os.path.isfile(model_state_file): checkpoint = torch.load(model_state_file) start_iter = checkpoint['start_iter'] get_module(model, distributed).load_state_dict(checkpoint['state_dict']) optimizer.load_state_dict(checkpoint['optimizer']) lr_scheduler.load_state_dict(checkpoint['lr_scheduler']) logger.info('Loaded checkpoint (starting from iter {})'.format(checkpoint['start_iter'])) data_time = AverageMeter() batch_time = AverageMeter() loss_meter = AverageMeter() # 显示模型的参数量 def get_parameter_number(net): total_num = sum(p.numel() for p in net.parameters()) trainable_num = sum(p.numel() for p in net.parameters() if p.requires_grad) # return {'Total': total_num/1000000, 'Trainable': trainable_num/1000000} logger.info('Total:{}M, Trainable:{}M'.format(total_num/1000000, trainable_num/1000000)) print(get_parameter_number(model)) # Debug output. if config.DEBUG.DEBUG: debug_out_dir = os.path.join(config.OUTPUT_DIR, 'debug_train') PathManager.mkdirs(debug_out_dir) # Train loop. try: for i in range(start_iter, max_iter): # data start_time = time.time() data = next(data_loader_iter) if not distributed: data = to_cuda(data, device) data_time.update(time.time() - start_time) # 取出mini-bach的数据和标签 image = data.pop('image') label = data.pop('label') # import imageio # import numpy as np # print(label.shape) # label_image = np.array(label.cpu()[0]) # print(label_image.shape) # imageio.imwrite('%s/%d_%s.png' % ('./', 1, 'debug_batch_label'), label_image.transpose(1, 2, 0)) # 向前传播 out_dict = model(image, data) # 计算代价函数 loss = out_dict['loss'] # 清零梯度准备计算 optimizer.zero_grad() # 反向传播 loss.backward() # 更新训练参数 optimizer.step() # Get lr. lr = optimizer.param_groups[best_param_group_id]["lr"] lr_scheduler.step() batch_time.update(time.time() - start_time) loss_meter.update(loss.detach().cpu().item(), image.size(0)) if i == 0 or (i + 1) % config.PRINT_FREQ == 0: msg = '[{0}/{1}] LR: {2:.7f}\t' \ 'Time: {batch_time.val:.3f}s ({batch_time.avg:.3f}s)\t' \ 'Data: {data_time.val:.3f}s ({data_time.avg:.3f}s)\t'.format( i + 1, max_iter, lr, batch_time=batch_time, data_time=data_time) msg += get_loss_info_str(get_module(model, distributed).loss_meter_dict) logger.info(msg) if i == 0 or (i + 1) % config.DEBUG.DEBUG_FREQ == 0: if comm.is_main_process() and config.DEBUG.DEBUG: save_debug_images( dataset=data_loader.dataset, label=label, batch_images=image, batch_targets=data, batch_outputs=out_dict, out_dir=debug_out_dir, iteration=i, target_keys=config.DEBUG.TARGET_KEYS, output_keys=config.DEBUG.OUTPUT_KEYS, iteration_to_remove=i - config.DEBUG.KEEP_INTERVAL ) if i == 0 or (i + 1) % config.CKPT_FREQ == 0: if comm.is_main_process(): torch.save({ 'start_iter': i + 1, 'state_dict': get_module(model, distributed).state_dict(), 'optimizer': optimizer.state_dict(), 'lr_scheduler': lr_scheduler.state_dict(), }, os.path.join(config.OUTPUT_DIR, 'checkpoint.pth.tar')) except Exception: logger.exception("Exception during training:") raise finally: if comm.is_main_process(): torch.save(get_module(model, distributed).state_dict(), os.path.join(config.OUTPUT_DIR, 'final_state.pth')) logger.info("Training finished.")