def test_return_batch_norm_params_with_notrain_when_train_is_false(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l2_regularizer {
       }
     }
     initializer {
       truncated_normal_initializer {
       }
     }
     batch_norm {
       decay: 0.7
       center: false
       scale: true
       epsilon: 0.03
       train: false
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   conv_scope_arguments = scope.values()[0]
   self.assertEqual(conv_scope_arguments['normalizer_fn'], slim.batch_norm)
   batch_norm_params = conv_scope_arguments['normalizer_params']
   self.assertAlmostEqual(batch_norm_params['decay'], 0.7)
   self.assertAlmostEqual(batch_norm_params['epsilon'], 0.03)
   self.assertFalse(batch_norm_params['center'])
   self.assertTrue(batch_norm_params['scale'])
   self.assertFalse(batch_norm_params['is_training'])
示例#2
0
def _build_ssd_feature_extractor(feature_extractor_config, is_training,
                                 reuse_weights=None):
  """Builds a ssd_meta_arch.SSDFeatureExtractor based on config.

  Args:
    feature_extractor_config: A SSDFeatureExtractor proto config from ssd.proto.
    is_training: True if this feature extractor is being built for training.
    reuse_weights: if the feature extractor should reuse weights.

  Returns:
    ssd_meta_arch.SSDFeatureExtractor based on config.

  Raises:
    ValueError: On invalid feature extractor type.
  """
  feature_type = feature_extractor_config.type
  depth_multiplier = feature_extractor_config.depth_multiplier
  min_depth = feature_extractor_config.min_depth
  pad_to_multiple = feature_extractor_config.pad_to_multiple
  batch_norm_trainable = feature_extractor_config.batch_norm_trainable
  conv_hyperparams = hyperparams_builder.build(
      feature_extractor_config.conv_hyperparams, is_training)

  if feature_type not in SSD_FEATURE_EXTRACTOR_CLASS_MAP:
    raise ValueError('Unknown ssd feature_extractor: {}'.format(feature_type))

  feature_extractor_class = SSD_FEATURE_EXTRACTOR_CLASS_MAP[feature_type]
  return feature_extractor_class(is_training, depth_multiplier, min_depth,
                                 pad_to_multiple, conv_hyperparams,
                                 batch_norm_trainable, reuse_weights)
示例#3
0
 def _build_arg_scope_with_conv_hyperparams(self):
     conv_hyperparams = hyperparams_pb2.Hyperparams()
     conv_hyperparams_text_proto = """
   regularizer {
     l2_regularizer {
     }
   }
   initializer {
     truncated_normal_initializer {
     }
   }
 """
     text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams)
     return hyperparams_builder.build(conv_hyperparams, is_training=True)
 def test_default_arg_scope_has_conv2d_transpose_op(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l1_regularizer {
       }
     }
     initializer {
       truncated_normal_initializer {
       }
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   self.assertTrue(self._get_scope_key(slim.conv2d_transpose) in scope)
 def test_explicit_fc_op_arg_scope_has_fully_connected_op(self):
   conv_hyperparams_text_proto = """
     op: FC
     regularizer {
       l1_regularizer {
       }
     }
     initializer {
       truncated_normal_initializer {
       }
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   self.assertTrue(self._get_scope_key(slim.fully_connected) in scope)
 def test_separable_conv2d_and_conv2d_and_transpose_have_same_parameters(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l1_regularizer {
       }
     }
     initializer {
       truncated_normal_initializer {
       }
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   kwargs_1, kwargs_2, kwargs_3 = scope.values()
   self.assertDictEqual(kwargs_1, kwargs_2)
   self.assertDictEqual(kwargs_1, kwargs_3)
 def test_use_relu_6_activation(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l2_regularizer {
       }
     }
     initializer {
       truncated_normal_initializer {
       }
     }
     activation: RELU_6
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   conv_scope_arguments = scope.values()[0]
   self.assertEqual(conv_scope_arguments['activation_fn'], tf.nn.relu6)
 def test_do_not_use_batch_norm_if_default(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l2_regularizer {
       }
     }
     initializer {
       truncated_normal_initializer {
       }
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   conv_scope_arguments = scope.values()[0]
   self.assertEqual(conv_scope_arguments['normalizer_fn'], None)
   self.assertEqual(conv_scope_arguments['normalizer_params'], None)
示例#9
0
 def _build_arg_scope_with_hyperparams(
         self, op_type=hyperparams_pb2.Hyperparams.FC):
     hyperparams = hyperparams_pb2.Hyperparams()
     hyperparams_text_proto = """
   activation: NONE
   regularizer {
     l2_regularizer {
     }
   }
   initializer {
     truncated_normal_initializer {
     }
   }
 """
     text_format.Merge(hyperparams_text_proto, hyperparams)
     hyperparams.op = op_type
     return hyperparams_builder.build(hyperparams, is_training=True)
 def test_variance_in_range_with_truncated_normal_initializer(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l2_regularizer {
       }
     }
     initializer {
       truncated_normal_initializer {
         mean: 0.0
         stddev: 0.8
       }
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   conv_scope_arguments = scope.values()[0]
   initializer = conv_scope_arguments['weights_initializer']
   self._assert_variance_in_range(initializer, shape=[100, 40],
                                  variance=0.49, tol=1e-1)
 def test_variance_in_range_with_variance_scaling_initializer_uniform(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l2_regularizer {
       }
     }
     initializer {
       variance_scaling_initializer {
         factor: 2.0
         mode: FAN_IN
         uniform: true
       }
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   conv_scope_arguments = scope.values()[0]
   initializer = conv_scope_arguments['weights_initializer']
   self._assert_variance_in_range(initializer, shape=[100, 40],
                                  variance=2. / 100.)
 def test_return_l1_regularized_weights(self):
   conv_hyperparams_text_proto = """
     regularizer {
       l1_regularizer {
         weight: 0.5
       }
     }
     initializer {
       truncated_normal_initializer {
       }
     }
   """
   conv_hyperparams_proto = hyperparams_pb2.Hyperparams()
   text_format.Merge(conv_hyperparams_text_proto, conv_hyperparams_proto)
   scope = hyperparams_builder.build(conv_hyperparams_proto, is_training=True)
   conv_scope_arguments = scope.values()[0]
   regularizer = conv_scope_arguments['weights_regularizer']
   weights = np.array([1., -1, 4., 2.])
   with self.test_session() as sess:
     result = sess.run(regularizer(tf.constant(weights)))
   self.assertAllClose(np.abs(weights).sum() * 0.5, result)
示例#13
0
def _build_faster_rcnn_model(frcnn_config, is_training):
  """Builds a Faster R-CNN or R-FCN detection model based on the model config.

  Builds R-FCN model if the second_stage_box_predictor in the config is of type
  `rfcn_box_predictor` else builds a Faster R-CNN model.

  Args:
    frcnn_config: A faster_rcnn.proto object containing the config for the
    desired FasterRCNNMetaArch or RFCNMetaArch.
    is_training: True if this model is being built for training purposes.

  Returns:
    FasterRCNNMetaArch based on the config.
  Raises:
    ValueError: If frcnn_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
  num_classes = frcnn_config.num_classes
  image_resizer_fn = image_resizer_builder.build(frcnn_config.image_resizer)

  feature_extractor = _build_faster_rcnn_feature_extractor(
      frcnn_config.feature_extractor, is_training)

  first_stage_only = frcnn_config.first_stage_only
  first_stage_anchor_generator = anchor_generator_builder.build(
      frcnn_config.first_stage_anchor_generator)

  first_stage_atrous_rate = frcnn_config.first_stage_atrous_rate
  first_stage_box_predictor_arg_scope = hyperparams_builder.build(
      frcnn_config.first_stage_box_predictor_conv_hyperparams, is_training)
  first_stage_box_predictor_kernel_size = (
      frcnn_config.first_stage_box_predictor_kernel_size)
  first_stage_box_predictor_depth = frcnn_config.first_stage_box_predictor_depth
  first_stage_minibatch_size = frcnn_config.first_stage_minibatch_size
  first_stage_positive_balance_fraction = (
      frcnn_config.first_stage_positive_balance_fraction)
  first_stage_nms_score_threshold = frcnn_config.first_stage_nms_score_threshold
  first_stage_nms_iou_threshold = frcnn_config.first_stage_nms_iou_threshold
  first_stage_max_proposals = frcnn_config.first_stage_max_proposals
  first_stage_loc_loss_weight = (
      frcnn_config.first_stage_localization_loss_weight)
  first_stage_obj_loss_weight = frcnn_config.first_stage_objectness_loss_weight

  initial_crop_size = frcnn_config.initial_crop_size
  maxpool_kernel_size = frcnn_config.maxpool_kernel_size
  maxpool_stride = frcnn_config.maxpool_stride

  second_stage_box_predictor = box_predictor_builder.build(
      hyperparams_builder.build,
      frcnn_config.second_stage_box_predictor,
      is_training=is_training,
      num_classes=num_classes)
  second_stage_batch_size = frcnn_config.second_stage_batch_size
  second_stage_balance_fraction = frcnn_config.second_stage_balance_fraction
  (second_stage_non_max_suppression_fn, second_stage_score_conversion_fn
  ) = post_processing_builder.build(frcnn_config.second_stage_post_processing)
  second_stage_localization_loss_weight = (
      frcnn_config.second_stage_localization_loss_weight)
  second_stage_classification_loss = (
      losses_builder.build_faster_rcnn_classification_loss(
          frcnn_config.second_stage_classification_loss))
  second_stage_classification_loss_weight = (
      frcnn_config.second_stage_classification_loss_weight)
  second_stage_mask_prediction_loss_weight = (
      frcnn_config.second_stage_mask_prediction_loss_weight)

  hard_example_miner = None
  if frcnn_config.HasField('hard_example_miner'):
    hard_example_miner = losses_builder.build_hard_example_miner(
        frcnn_config.hard_example_miner,
        second_stage_classification_loss_weight,
        second_stage_localization_loss_weight)

  common_kwargs = {
      'is_training': is_training,
      'num_classes': num_classes,
      'image_resizer_fn': image_resizer_fn,
      'feature_extractor': feature_extractor,
      'first_stage_only': first_stage_only,
      'first_stage_anchor_generator': first_stage_anchor_generator,
      'first_stage_atrous_rate': first_stage_atrous_rate,
      'first_stage_box_predictor_arg_scope':
      first_stage_box_predictor_arg_scope,
      'first_stage_box_predictor_kernel_size':
      first_stage_box_predictor_kernel_size,
      'first_stage_box_predictor_depth': first_stage_box_predictor_depth,
      'first_stage_minibatch_size': first_stage_minibatch_size,
      'first_stage_positive_balance_fraction':
      first_stage_positive_balance_fraction,
      'first_stage_nms_score_threshold': first_stage_nms_score_threshold,
      'first_stage_nms_iou_threshold': first_stage_nms_iou_threshold,
      'first_stage_max_proposals': first_stage_max_proposals,
      'first_stage_localization_loss_weight': first_stage_loc_loss_weight,
      'first_stage_objectness_loss_weight': first_stage_obj_loss_weight,
      'second_stage_batch_size': second_stage_batch_size,
      'second_stage_balance_fraction': second_stage_balance_fraction,
      'second_stage_non_max_suppression_fn':
      second_stage_non_max_suppression_fn,
      'second_stage_score_conversion_fn': second_stage_score_conversion_fn,
      'second_stage_localization_loss_weight':
      second_stage_localization_loss_weight,
      'second_stage_classification_loss':
      second_stage_classification_loss,
      'second_stage_classification_loss_weight':
      second_stage_classification_loss_weight,
      'hard_example_miner': hard_example_miner}

  if isinstance(second_stage_box_predictor, box_predictor.RfcnBoxPredictor):
    return rfcn_meta_arch.RFCNMetaArch(
        second_stage_rfcn_box_predictor=second_stage_box_predictor,
        **common_kwargs)
  else:
    return faster_rcnn_meta_arch.FasterRCNNMetaArch(
        initial_crop_size=initial_crop_size,
        maxpool_kernel_size=maxpool_kernel_size,
        maxpool_stride=maxpool_stride,
        second_stage_mask_rcnn_box_predictor=second_stage_box_predictor,
        second_stage_mask_prediction_loss_weight=(
            second_stage_mask_prediction_loss_weight),
        **common_kwargs)