def test_raise_error_on_empty_config(self):
   losses_text_proto = """
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   with self.assertRaises(ValueError):
     losses_builder.build(losses_proto)
 def test_build_reweighting_unmatched_anchors(self):
   losses_text_proto = """
     localization_loss {
       weighted_l2 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
     hard_example_miner {
     }
     classification_weight: 0.8
     localization_weight: 0.2
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   (classification_loss, localization_loss, classification_weight,
    localization_weight, hard_example_miner, _,
    _) = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(hard_example_miner, losses.HardExampleMiner))
   self.assertTrue(
       isinstance(classification_loss,
                  losses.WeightedSoftmaxClassificationLoss))
   self.assertTrue(
       isinstance(localization_loss, losses.WeightedL2LocalizationLoss))
   self.assertAlmostEqual(classification_weight, 0.8)
   self.assertAlmostEqual(localization_weight, 0.2)
 def test_build_hard_example_miner_with_non_default_values(self):
   losses_text_proto = """
     localization_loss {
       weighted_l2 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
     hard_example_miner {
       num_hard_examples: 32
       iou_threshold: 0.5
       loss_type: LOCALIZATION
       max_negatives_per_positive: 10
       min_negatives_per_image: 3
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, _, _, _, hard_example_miner, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(hard_example_miner, losses.HardExampleMiner))
   self.assertEqual(hard_example_miner._num_hard_examples, 32)
   self.assertAlmostEqual(hard_example_miner._iou_threshold, 0.5)
   self.assertEqual(hard_example_miner._max_negatives_per_positive, 10)
   self.assertEqual(hard_example_miner._min_negatives_per_image, 3)
Beispiel #4
0
def _build_ssd_model(ssd_config, is_training, add_summaries):
  """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.
    add_summaries: Whether to add tf summaries in the model.

  Returns:
    SSDMetaArch based on the config.
  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
  num_classes = ssd_config.num_classes

  # Feature extractor
  feature_extractor = _build_ssd_feature_extractor(ssd_config.feature_extractor,
                                                   is_training)

  box_coder = box_coder_builder.build(ssd_config.box_coder)
  matcher = matcher_builder.build(ssd_config.matcher)
  region_similarity_calculator = sim_calc.build(
      ssd_config.similarity_calculator)
  ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                  ssd_config.box_predictor,
                                                  is_training, num_classes)
  anchor_generator = anchor_generator_builder.build(
      ssd_config.anchor_generator)
  image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
  non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
      ssd_config.post_processing)
  (classification_loss, localization_loss, classification_weight,
   localization_weight,
   hard_example_miner) = losses_builder.build(ssd_config.loss)
  normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches

  return ssd_meta_arch.SSDMetaArch(
      is_training,
      anchor_generator,
      ssd_box_predictor,
      box_coder,
      feature_extractor,
      matcher,
      region_similarity_calculator,
      image_resizer_fn,
      non_max_suppression_fn,
      score_conversion_fn,
      classification_loss,
      localization_loss,
      classification_weight,
      localization_weight,
      normalize_loss_by_num_matches,
      hard_example_miner,
      add_summaries=add_summaries)
 def test_raise_error_when_both_focal_loss_and_hard_example_miner(self):
   losses_text_proto = """
     localization_loss {
       weighted_l2 {
       }
     }
     classification_loss {
       weighted_sigmoid_focal {
       }
     }
     hard_example_miner {
     }
     classification_weight: 0.8
     localization_weight: 0.2
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   with self.assertRaises(ValueError):
     losses_builder.build(losses_proto)
 def test_do_not_build_hard_example_miner_by_default(self):
   losses_text_proto = """
     localization_loss {
       weighted_l2 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, _, _, _, hard_example_miner, _, _ = losses_builder.build(losses_proto)
   self.assertEqual(hard_example_miner, None)
 def test_build_weighted_sigmoid_classification_loss(self):
   losses_text_proto = """
     classification_loss {
       weighted_sigmoid {
       }
     }
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   classification_loss, _, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(classification_loss,
                              losses.WeightedSigmoidClassificationLoss))
Beispiel #8
0
 def test_build_weighted_softmax_classification_loss(self):
   losses_text_proto = """
     classification_loss {
       weighted_softmax {
       }
     }
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   classification_loss, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(classification_loss,
                              losses.WeightedSoftmaxClassificationLoss))
 def test_build_weighted_iou_localization_loss(self):
   losses_text_proto = """
     localization_loss {
       weighted_iou {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, localization_loss, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(localization_loss,
                              losses.WeightedIOULocalizationLoss))
Beispiel #10
0
 def test_build_weighted_smooth_l1_localization_loss(self):
   losses_text_proto = """
     localization_loss {
       weighted_smooth_l1 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, localization_loss, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(localization_loss,
                              losses.WeightedSmoothL1LocalizationLoss))
 def test_build_weighted_smooth_l1_localization_loss_default_delta(self):
   losses_text_proto = """
     localization_loss {
       weighted_smooth_l1 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, localization_loss, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(localization_loss,
                              losses.WeightedSmoothL1LocalizationLoss))
   self.assertAlmostEqual(localization_loss._delta, 1.0)
Beispiel #12
0
 def test_build_weighted_smooth_l1_localization_loss_default_delta(self):
   losses_text_proto = """
     localization_loss {
       weighted_smooth_l1 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, localization_loss, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertIsInstance(localization_loss,
                         losses.WeightedSmoothL1LocalizationLoss)
   self.assertAlmostEqual(localization_loss._delta, 1.0)
Beispiel #13
0
 def test_build_bootstrapped_sigmoid_classification_loss(self):
   losses_text_proto = """
     classification_loss {
       bootstrapped_sigmoid {
         alpha: 0.5
       }
     }
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   classification_loss, _, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(classification_loss,
                              losses.BootstrappedSigmoidClassificationLoss))
Beispiel #14
0
 def test_build_weighted_iou_localization_loss(self):
     losses_text_proto = """
   localization_loss {
     weighted_iou {
     }
   }
   classification_loss {
     weighted_softmax {
     }
   }
 """
     losses_proto = losses_pb2.Loss()
     text_format.Merge(losses_text_proto, losses_proto)
     _, localization_loss, _, _, _, _, _ = losses_builder.build(
         losses_proto)
     self.assertIsInstance(localization_loss,
                           losses.WeightedIOULocalizationLoss)
 def test_build_weighted_softmax_classification_loss_with_logit_scale(self):
   losses_text_proto = """
     classification_loss {
       weighted_softmax {
         logit_scale: 2.0
       }
     }
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   classification_loss, _, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(classification_loss,
                              losses.WeightedSoftmaxClassificationLoss))
Beispiel #16
0
def object_center_proto_to_params(oc_config):
    """Converts CenterNet.ObjectCenter proto to parameter namedtuple."""
    loss = losses_pb2.Loss()
    # Add dummy localization loss to avoid the loss_builder throwing error.
    # TODO(yuhuic): update the loss builder to take the localization loss
    # directly.
    loss.localization_loss.weighted_l2.CopyFrom(
        losses_pb2.WeightedL2LocalizationLoss())
    loss.classification_loss.CopyFrom(oc_config.classification_loss)
    classification_loss, _, _, _, _, _, _ = (losses_builder.build(loss))
    return center_net_meta_arch.ObjectCenterParams(
        classification_loss=classification_loss,
        object_center_loss_weight=oc_config.object_center_loss_weight,
        heatmap_bias_init=oc_config.heatmap_bias_init,
        min_box_overlap_iou=oc_config.min_box_overlap_iou,
        max_box_predictions=oc_config.max_box_predictions,
        use_labeled_classes=oc_config.use_labeled_classes)
Beispiel #17
0
 def test_build_weighted_sigmoid_classification_loss(self):
     losses_text_proto = """
   classification_loss {
     weighted_sigmoid {
     }
   }
   localization_loss {
     weighted_l2 {
     }
   }
 """
     losses_proto = losses_pb2.Loss()
     text_format.Merge(losses_text_proto, losses_proto)
     classification_loss, _, _, _, _, _, _ = losses_builder.build(
         losses_proto)
     self.assertIsInstance(classification_loss,
                           losses.WeightedSigmoidClassificationLoss)
Beispiel #18
0
def _build_ssd_model(ssd_config, is_training):
    """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.

  Returns:
    SSDMetaArch based on the config.
  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
    num_classes = ssd_config.num_classes

    # Feature extractor
    feature_extractor = _build_ssd_feature_extractor(
        ssd_config.feature_extractor, is_training
        and ssd_config.feature_extractor.trainable)

    box_coder = box_coder_builder.build(ssd_config.box_coder)
    matcher = matcher_builder.build(ssd_config.matcher)
    region_similarity_calculator = sim_calc.build(
        ssd_config.similarity_calculator)
    ssd_box_predictor = box_predictor_builder.build(
        hyperparams_builder.build, ssd_config.box_predictor, is_training
        and ssd_config.box_predictor.trainable, num_classes)

    anchor_generator = anchor_generator_builder.build(
        ssd_config.anchor_generator)
    image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
    non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
        ssd_config.post_processing)
    (classification_loss, localization_loss, classification_weight,
     localization_weight,
     hard_example_miner) = losses_builder.build(ssd_config.loss)
    normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches

    return ssd_meta_arch.SSDMetaArch(
        is_training, anchor_generator, ssd_box_predictor, box_coder,
        feature_extractor, matcher, region_similarity_calculator,
        image_resizer_fn, non_max_suppression_fn, score_conversion_fn,
        classification_loss, localization_loss, classification_weight,
        localization_weight, normalize_loss_by_num_matches, hard_example_miner)
Beispiel #19
0
 def test_build_weighted_sigmoid_focal_classification_loss(self):
   losses_text_proto = """
     classification_loss {
       weighted_sigmoid_focal {
       }
     }
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   classification_loss, _, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(classification_loss,
                              losses.SigmoidFocalClassificationLoss))
   self.assertAlmostEqual(classification_loss._alpha, None)
   self.assertAlmostEqual(classification_loss._gamma, 2.0)
Beispiel #20
0
def build_man_model(model_config, is_training):

    num_classes = model_config.num_classes
    feature_extractor = _build_man_feature_extractor(model_config.feature_extractor,
                                                     is_training)

    box_coder = box_coder_builder.build(model_config.box_coder)
    matcher = matcher_builder.build(model_config.matcher)
    region_similarity_calculator = sim_calc.build(
        model_config.similarity_calculator)
    ssd_box_predictor = _build_man_box_predictor(is_training, num_classes, model_config.box_predictor)
    # ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
    #                                                 model_config.box_predictor,
    #                                                 is_training, num_classes)
    anchor_generator = _build_man_anchor_generator(model_config.anchor_generator)
    # anchor_generator = anchor_generator_builder.build(
    #     model_config.anchor_generator)
    image_resizer_fn = image_resizer_builder.build(model_config.image_resizer)
    non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
        model_config.post_processing)
    (classification_loss, localization_loss, classification_weight,
     localization_weight,
     hard_example_miner) = losses_builder.build(model_config.loss)
    normalize_loss_by_num_matches = model_config.normalize_loss_by_num_matches

    return MANMetaArch(
        is_training,
        anchor_generator,
        ssd_box_predictor,
        box_coder,
        feature_extractor,
        matcher,
        region_similarity_calculator,
        image_resizer_fn,
        non_max_suppression_fn,
        score_conversion_fn,
        classification_loss,
        localization_loss,
        classification_weight,
        localization_weight,
        normalize_loss_by_num_matches,
        hard_example_miner,
        add_summaries=False)
Beispiel #21
0
def keypoint_proto_to_params(kp_config, keypoint_map_dict):
    """Converts CenterNet.KeypointEstimation proto to parameter namedtuple."""
    label_map_item = keypoint_map_dict[kp_config.keypoint_class_name]

    classification_loss, localization_loss, _, _, _, _, _ = (
        losses_builder.build(kp_config.loss))

    keypoint_indices = [keypoint.id for keypoint in label_map_item.keypoints]
    keypoint_labels = [keypoint.label for keypoint in label_map_item.keypoints]
    keypoint_std_dev_dict = {
        label: KEYPOINT_STD_DEV_DEFAULT
        for label in keypoint_labels
    }
    if kp_config.keypoint_label_to_std:
        for label, value in kp_config.keypoint_label_to_std.items():
            keypoint_std_dev_dict[label] = value
    keypoint_std_dev = [
        keypoint_std_dev_dict[label] for label in keypoint_labels
    ]
    return center_net_meta_arch.KeypointEstimationParams(
        task_name=kp_config.task_name,
        class_id=label_map_item.id - CLASS_ID_OFFSET,
        keypoint_indices=keypoint_indices,
        classification_loss=classification_loss,
        localization_loss=localization_loss,
        keypoint_labels=keypoint_labels,
        keypoint_std_dev=keypoint_std_dev,
        task_loss_weight=kp_config.task_loss_weight,
        keypoint_regression_loss_weight=kp_config.
        keypoint_regression_loss_weight,
        keypoint_heatmap_loss_weight=kp_config.keypoint_heatmap_loss_weight,
        keypoint_offset_loss_weight=kp_config.keypoint_offset_loss_weight,
        heatmap_bias_init=kp_config.heatmap_bias_init,
        keypoint_candidate_score_threshold=(
            kp_config.keypoint_candidate_score_threshold),
        num_candidates_per_keypoint=kp_config.num_candidates_per_keypoint,
        peak_max_pool_kernel_size=kp_config.peak_max_pool_kernel_size,
        unmatched_keypoint_score=kp_config.unmatched_keypoint_score,
        box_scale=kp_config.box_scale,
        candidate_search_scale=kp_config.candidate_search_scale,
        candidate_ranking_mode=kp_config.candidate_ranking_mode,
        offset_peak_radius=kp_config.offset_peak_radius,
        per_keypoint_offset=kp_config.per_keypoint_offset)
Beispiel #22
0
 def test_build_hard_example_miner_for_localization_loss(self):
   losses_text_proto = """
     localization_loss {
       weighted_l2 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
     hard_example_miner {
       loss_type: LOCALIZATION
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, _, _, _, hard_example_miner, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(hard_example_miner, losses.HardExampleMiner))
   self.assertEqual(hard_example_miner._loss_type, 'loc')
Beispiel #23
0
 def test_build_dice_loss(self):
     losses_text_proto = """
   classification_loss {
     weighted_dice_classification_loss {
       squared_normalization: true
     }
   }
   localization_loss {
     l1_localization_loss {
     }
   }
 """
     losses_proto = losses_pb2.Loss()
     text_format.Merge(losses_text_proto, losses_proto)
     classification_loss, _, _, _, _, _, _ = losses_builder.build(
         losses_proto)
     self.assertIsInstance(classification_loss,
                           losses.WeightedDiceClassificationLoss)
     assert classification_loss._squared_normalization
Beispiel #24
0
 def test_do_not_build_hard_example_miner_by_default(self):
     losses_text_proto = """
   localization_loss {
     weighted_l2 {
     }
   }
   classification_loss {
     weighted_softmax {
     }
   }
   classification_in_image_level_loss {
     weighted_sigmoid {
     }
   }
 """
     losses_proto = losses_pb2.Loss()
     text_format.Merge(losses_text_proto, losses_proto)
     _, _, _, _, _, _, hard_example_miner = losses_builder.build(
         losses_proto)
     self.assertEqual(hard_example_miner, None)
Beispiel #25
0
 def test_build_weighted_sigmoid_focal_loss_non_default(self):
   losses_text_proto = """
     classification_loss {
       weighted_sigmoid_focal {
         alpha: 0.25
         gamma: 3.0
       }
     }
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   classification_loss, _, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertIsInstance(classification_loss,
                         losses.SigmoidFocalClassificationLoss)
   self.assertAlmostEqual(classification_loss._alpha, 0.25)
   self.assertAlmostEqual(classification_loss._gamma, 3.0)
Beispiel #26
0
def _build_east_model(east_config, is_training):
    """Builds an EAST detection model based on the model config.

  Args:
    east_config: A east.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.

  Returns:
    EASTMetaArch based on the config.
  Raises:
    ValueError: If east_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
    num_classes = east_config.num_classes

    # Feature extractor
    feature_extractor = _build_east_feature_extractor(
        east_config.feature_extractor, is_training)

    box_coder = box_coder_builder.build(east_config.box_coder)
    box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                east_config.box_predictor,
                                                is_training, num_classes)
    anchor_generator = anchor_generator_builder.build(
        east_config.anchor_generator)
    #image_resizer_fn = image_resizer_builder.build(east_config.image_resizer)
    image_resizer_fn = None
    non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
        east_config.post_processing)
    (classification_loss, localization_loss, classification_weight,
     localization_weight,
     hard_example_miner) = losses_builder.build(east_config.loss)
    normalize_loss_by_num_matches = east_config.normalize_loss_by_num_matches

    return east_meta_arch.EASTMetaArch(
        is_training, anchor_generator, box_predictor, box_coder,
        feature_extractor, image_resizer_fn, non_max_suppression_fn,
        score_conversion_fn, classification_loss, localization_loss,
        classification_weight, localization_weight,
        normalize_loss_by_num_matches)
Beispiel #27
0
 def test_build_penalty_reduced_logistic_focal_loss(self):
     losses_text_proto = """
   classification_loss {
     penalty_reduced_logistic_focal_loss {
       alpha: 2.0
       beta: 4.0
     }
   }
   localization_loss {
     l1_localization_loss {
     }
   }
 """
     losses_proto = losses_pb2.Loss()
     text_format.Merge(losses_text_proto, losses_proto)
     classification_loss, _, _, _, _, _, _ = losses_builder.build(
         losses_proto)
     self.assertIsInstance(classification_loss,
                           losses.PenaltyReducedLogisticFocalLoss)
     self.assertAlmostEqual(classification_loss._alpha, 2.0)
     self.assertAlmostEqual(classification_loss._beta, 4.0)
Beispiel #28
0
 def test_anchorwise_output(self):
   losses_text_proto = """
     localization_loss {
       weighted_smooth_l1 {
       }
     }
     classification_loss {
       weighted_softmax {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   _, localization_loss, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(localization_loss,
                              losses.WeightedSmoothL1LocalizationLoss))
   predictions = tf.constant([[[0.0, 0.0, 1.0, 1.0], [0.0, 0.0, 1.0, 1.0]]])
   targets = tf.constant([[[0.0, 0.0, 1.0, 1.0], [0.0, 0.0, 1.0, 1.0]]])
   weights = tf.constant([[1.0, 1.0]])
   loss = localization_loss(predictions, targets, weights=weights)
   self.assertEqual(loss.shape, [1, 2])
Beispiel #29
0
 def test_build_weighted_l2_localization_loss(self):
     losses_text_proto = """
   localization_loss {
     weighted_l2 {
     }
   }
   classification_loss {
     weighted_softmax {
     }
   }
   classification_in_image_level_loss {
     weighted_sigmoid {
     }
   }
 """
     losses_proto = losses_pb2.Loss()
     text_format.Merge(losses_text_proto, losses_proto)
     _, localization_loss, _, _, _, _, _ = losses_builder.build(
         losses_proto)
     self.assertTrue(
         isinstance(localization_loss, losses.WeightedL2LocalizationLoss))
Beispiel #30
0
 def test_anchorwise_output(self):
   losses_text_proto = """
     classification_loss {
       weighted_sigmoid {
         anchorwise_output: true
       }
     }
     localization_loss {
       weighted_l2 {
       }
     }
   """
   losses_proto = losses_pb2.Loss()
   text_format.Merge(losses_text_proto, losses_proto)
   classification_loss, _, _, _, _, _, _ = losses_builder.build(losses_proto)
   self.assertTrue(isinstance(classification_loss,
                              losses.WeightedSigmoidClassificationLoss))
   predictions = tf.constant([[[0.0, 1.0, 0.0], [0.0, 0.5, 0.5]]])
   targets = tf.constant([[[0.0, 1.0, 0.0], [0.0, 0.0, 1.0]]])
   weights = tf.constant([[[1.0, 1.0, 1.0], [1.0, 1.0, 1.0]]])
   loss = classification_loss(predictions, targets, weights=weights)
   self.assertEqual(loss.shape, [1, 2, 3])
Beispiel #31
0
 def test_build_weighted_softmax_classification_loss_with_logit_scale(self):
     losses_text_proto = """
   classification_loss {
     weighted_softmax {
       logit_scale: 2.0
     }
   }
   localization_loss {
     weighted_l2 {
     }
   }
   classification_in_image_level_loss {
     weighted_sigmoid {
     }
   }
 """
     losses_proto = losses_pb2.Loss()
     text_format.Merge(losses_text_proto, losses_proto)
     classification_loss, _, _, _, _, _, _ = losses_builder.build(
         losses_proto)
     self.assertTrue(
         isinstance(classification_loss,
                    losses.WeightedSoftmaxClassificationLoss))
Beispiel #32
0
def _build_lstm_model(ssd_config, lstm_config, is_training):
  """Builds an LSTM detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      LSTMSSDMetaArch.
    lstm_config: LstmModel config proto that specifies LSTM train/eval configs.
    is_training: True if this model is being built for training purposes.

  Returns:
    LSTMSSDMetaArch based on the config.
  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map), or if lstm_config.interleave_strategy is not recognized.
    ValueError: If unroll_length is not specified in the config file.
  """
  feature_extractor = _build_lstm_feature_extractor(
      ssd_config.feature_extractor, is_training, lstm_config)

  box_coder = box_coder_builder.build(ssd_config.box_coder)
  matcher = matcher_builder.build(ssd_config.matcher)
  region_similarity_calculator = sim_calc.build(
      ssd_config.similarity_calculator)

  num_classes = ssd_config.num_classes
  ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                  ssd_config.box_predictor,
                                                  is_training, num_classes)
  anchor_generator = anchor_generator_builder.build(ssd_config.anchor_generator)
  image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
  non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
      ssd_config.post_processing)
  (classification_loss, localization_loss, classification_weight,
   localization_weight, miner, _, _) = losses_builder.build(ssd_config.loss)

  normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
  encode_background_as_zeros = ssd_config.encode_background_as_zeros
  negative_class_weight = ssd_config.negative_class_weight

  # Extra configs for lstm unroll length.
  unroll_length = None
  if 'lstm' in ssd_config.feature_extractor.type:
    if is_training:
      unroll_length = lstm_config.train_unroll_length
    else:
      unroll_length = lstm_config.eval_unroll_length
  if unroll_length is None:
    raise ValueError('No unroll length found in the config file')

  target_assigner_instance = target_assigner.TargetAssigner(
      region_similarity_calculator,
      matcher,
      box_coder,
      negative_class_weight=negative_class_weight)

  lstm_model = lstm_ssd_meta_arch.LSTMSSDMetaArch(
      is_training=is_training,
      anchor_generator=anchor_generator,
      box_predictor=ssd_box_predictor,
      box_coder=box_coder,
      feature_extractor=feature_extractor,
      encode_background_as_zeros=encode_background_as_zeros,
      image_resizer_fn=image_resizer_fn,
      non_max_suppression_fn=non_max_suppression_fn,
      score_conversion_fn=score_conversion_fn,
      classification_loss=classification_loss,
      localization_loss=localization_loss,
      classification_loss_weight=classification_weight,
      localization_loss_weight=localization_weight,
      normalize_loss_by_num_matches=normalize_loss_by_num_matches,
      hard_example_miner=miner,
      unroll_length=unroll_length,
      target_assigner_instance=target_assigner_instance)

  return lstm_model
def _build_ssd_model(ssd_config, is_training):
    """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.

  Returns:
    SSDMetaArch based on the config.
  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
    num_classes = ssd_config.num_classes

    # Feature extractor
    feature_extractor = _build_ssd_feature_extractor(
        ssd_config.feature_extractor, is_training)

    box_coder = box_coder_builder.build(ssd_config.box_coder)
    matcher = matcher_builder.build(ssd_config.matcher)
    region_similarity_calculator = sim_calc.build(
        ssd_config.similarity_calculator)
    ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                    ssd_config.box_predictor,
                                                    is_training, num_classes)
    anchor_generator = anchor_generator_builder.build(
        ssd_config.anchor_generator)
    image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
    non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
        ssd_config.post_processing)
    (classification_loss, localization_loss, classification_weight,
     localization_weight,
     hard_example_miner) = losses_builder.build(ssd_config.loss)
    normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches

    common_kwargs = {
        'is_training': is_training,
        'anchor_generator': anchor_generator,
        'box_predictor': ssd_box_predictor,
        'box_coder': box_coder,
        'feature_extractor': feature_extractor,
        'matcher': matcher,
        'region_similarity_calculator': region_similarity_calculator,
        'image_resizer_fn': image_resizer_fn,
        'non_max_suppression_fn': non_max_suppression_fn,
        'score_conversion_fn': score_conversion_fn,
        'classification_loss': classification_loss,
        'localization_loss': localization_loss,
        'classification_loss_weight': classification_weight,
        'localization_loss_weight': localization_weight,
        'normalize_loss_by_num_matches': normalize_loss_by_num_matches,
        'hard_example_miner': hard_example_miner
    }

    if isinstance(anchor_generator,
                  yolo_grid_anchor_generator.YoloGridAnchorGenerator):
        return yolo_meta_arch.YOLOMetaArch(**common_kwargs)
    else:
        return ssd_meta_arch.SSDMetaArch(**common_kwargs)
Beispiel #34
0
def _build_ssd_model(ssd_config, is_training, add_summaries):
  
  num_classes = ssd_config.num_classes

  # Feature extractor
  feature_extractor = _build_ssd_feature_extractor(
      feature_extractor_config=ssd_config.feature_extractor,
      freeze_batchnorm=ssd_config.freeze_batchnorm,
      is_training=is_training)

  box_coder = box_coder_builder.build(ssd_config.box_coder)
  matcher = matcher_builder.build(ssd_config.matcher)
  region_similarity_calculator = sim_calc.build(
      ssd_config.similarity_calculator)
  encode_background_as_zeros = ssd_config.encode_background_as_zeros
  negative_class_weight = ssd_config.negative_class_weight
  anchor_generator = anchor_generator_builder.build(
      ssd_config.anchor_generator)
  if feature_extractor.is_keras_model:
    ssd_box_predictor = box_predictor_builder.build_keras(
        hyperparams_fn=hyperparams_builder.KerasLayerHyperparams,
        freeze_batchnorm=ssd_config.freeze_batchnorm,
        inplace_batchnorm_update=False,
        num_predictions_per_location_list=anchor_generator
        .num_anchors_per_location(),
        box_predictor_config=ssd_config.box_predictor,
        is_training=is_training,
        num_classes=num_classes,
        add_background_class=ssd_config.add_background_class)
  else:
    ssd_box_predictor = box_predictor_builder.build(
        hyperparams_builder.build, ssd_config.box_predictor, is_training,
        num_classes, ssd_config.add_background_class)
  image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
  non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
      ssd_config.post_processing)
  (classification_loss, localization_loss, classification_weight,
   localization_weight, hard_example_miner, random_example_sampler,
   expected_loss_weights_fn) = losses_builder.build(ssd_config.loss)
  normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
  normalize_loc_loss_by_codesize = ssd_config.normalize_loc_loss_by_codesize

  equalization_loss_config = ops.EqualizationLossConfig(
      weight=ssd_config.loss.equalization_loss.weight,
      exclude_prefixes=ssd_config.loss.equalization_loss.exclude_prefixes)

  target_assigner_instance = target_assigner.TargetAssigner(
      region_similarity_calculator,
      matcher,
      box_coder,
      negative_class_weight=negative_class_weight)

  ssd_meta_arch_fn = ssd_meta_arch.SSDMetaArch
  kwargs = {}

  return ssd_meta_arch_fn(
      is_training=is_training,
      anchor_generator=anchor_generator,
      box_predictor=ssd_box_predictor,
      box_coder=box_coder,
      feature_extractor=feature_extractor,
      encode_background_as_zeros=encode_background_as_zeros,
      image_resizer_fn=image_resizer_fn,
      non_max_suppression_fn=non_max_suppression_fn,
      score_conversion_fn=score_conversion_fn,
      classification_loss=classification_loss,
      localization_loss=localization_loss,
      classification_loss_weight=classification_weight,
      localization_loss_weight=localization_weight,
      normalize_loss_by_num_matches=normalize_loss_by_num_matches,
      hard_example_miner=hard_example_miner,
      target_assigner_instance=target_assigner_instance,
      add_summaries=add_summaries,
      normalize_loc_loss_by_codesize=normalize_loc_loss_by_codesize,
      freeze_batchnorm=ssd_config.freeze_batchnorm,
      inplace_batchnorm_update=ssd_config.inplace_batchnorm_update,
      add_background_class=ssd_config.add_background_class,
      explicit_background_class=ssd_config.explicit_background_class,
      random_example_sampler=random_example_sampler,
      expected_loss_weights_fn=expected_loss_weights_fn,
      use_confidences_as_targets=ssd_config.use_confidences_as_targets,
      implicit_example_weight=ssd_config.implicit_example_weight,
      equalization_loss_config=equalization_loss_config,
      **kwargs)
Beispiel #35
0
def _build_ssd_model(ssd_config,
                     is_training,
                     add_summaries,
                     add_background_class=True):
    """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.
    add_summaries: Whether to add tf summaries in the model.
    add_background_class: Whether to add an implicit background class to one-hot
      encodings of groundtruth labels. Set to false if using groundtruth labels
      with an explicit background class or using multiclass scores instead of
      truth in the case of distillation.
  Returns:
    SSDMetaArch based on the config.

  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
    num_classes = ssd_config.num_classes

    # Feature extractor
    feature_extractor = _build_ssd_feature_extractor(
        feature_extractor_config=ssd_config.feature_extractor,
        is_training=is_training)

    box_coder = box_coder_builder.build(ssd_config.box_coder)
    matcher = matcher_builder.build(ssd_config.matcher)
    region_similarity_calculator = sim_calc.build(
        ssd_config.similarity_calculator)
    encode_background_as_zeros = ssd_config.encode_background_as_zeros
    negative_class_weight = ssd_config.negative_class_weight
    ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                    ssd_config.box_predictor,
                                                    is_training, num_classes)
    anchor_generator = anchor_generator_builder.build(
        ssd_config.anchor_generator)
    image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
    non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
        ssd_config.post_processing)
    (classification_loss, localization_loss, classification_weight,
     localization_weight, hard_example_miner,
     random_example_sampler) = losses_builder.build(ssd_config.loss)
    normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
    normalize_loc_loss_by_codesize = ssd_config.normalize_loc_loss_by_codesize
    weight_regression_loss_by_score = (
        ssd_config.weight_regression_loss_by_score)

    target_assigner_instance = target_assigner.TargetAssigner(
        region_similarity_calculator,
        matcher,
        box_coder,
        negative_class_weight=negative_class_weight,
        weight_regression_loss_by_score=weight_regression_loss_by_score)

    expected_classification_loss_under_sampling = None
    if ssd_config.use_expected_classification_loss_under_sampling:
        expected_classification_loss_under_sampling = functools.partial(
            ops.expected_classification_loss_under_sampling,
            minimum_negative_sampling=ssd_config.minimum_negative_sampling,
            desired_negative_sampling_ratio=ssd_config.
            desired_negative_sampling_ratio)

    ssd_meta_arch_fn = ssd_meta_arch.SSDMetaArch
    # BEGIN GOOGLE-INTERNAL
    # TODO(lzc): move ssd_mask_meta_arch to third party when it has decent
    # performance relative to a comparable Mask R-CNN model (b/112561592).
    predictor_config = ssd_config.box_predictor
    predict_instance_masks = False
    if predictor_config.WhichOneof(
            'box_predictor_oneof') == 'convolutional_box_predictor':
        predict_instance_masks = (
            predictor_config.convolutional_box_predictor.HasField('mask_head'))
    elif predictor_config.WhichOneof(
            'box_predictor_oneof'
    ) == 'weight_shared_convolutional_box_predictor':
        predict_instance_masks = (
            predictor_config.weight_shared_convolutional_box_predictor.
            HasField('mask_head'))
    if predict_instance_masks:
        ssd_meta_arch_fn = ssd_mask_meta_arch.SSDMaskMetaArch
    # END GOOGLE-INTERNAL

    return ssd_meta_arch_fn(
        is_training=is_training,
        anchor_generator=anchor_generator,
        box_predictor=ssd_box_predictor,
        box_coder=box_coder,
        feature_extractor=feature_extractor,
        encode_background_as_zeros=encode_background_as_zeros,
        image_resizer_fn=image_resizer_fn,
        non_max_suppression_fn=non_max_suppression_fn,
        score_conversion_fn=score_conversion_fn,
        classification_loss=classification_loss,
        localization_loss=localization_loss,
        classification_loss_weight=classification_weight,
        localization_loss_weight=localization_weight,
        normalize_loss_by_num_matches=normalize_loss_by_num_matches,
        hard_example_miner=hard_example_miner,
        target_assigner_instance=target_assigner_instance,
        add_summaries=add_summaries,
        normalize_loc_loss_by_codesize=normalize_loc_loss_by_codesize,
        freeze_batchnorm=ssd_config.freeze_batchnorm,
        inplace_batchnorm_update=ssd_config.inplace_batchnorm_update,
        add_background_class=add_background_class,
        random_example_sampler=random_example_sampler,
        expected_classification_loss_under_sampling=
        expected_classification_loss_under_sampling)
def _build_ssd_model(ssd_config,
                     is_training,
                     add_summaries,
                     add_background_class=True):
    """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.
    add_summaries: Whether to add tf summaries in the model.
    add_background_class: Whether to add an implicit background class to one-hot
      encodings of groundtruth labels. Set to false if using groundtruth labels
      with an explicit background class or using multiclass scores instead of
      truth in the case of distillation.
  Returns:
    SSDMetaArch based on the config.

  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
    num_classes = ssd_config.num_classes

    # Feature extractor
    feature_extractor = _build_ssd_feature_extractor(
        feature_extractor_config=ssd_config.feature_extractor,
        is_training=is_training)

    box_coder = box_coder_builder.build(ssd_config.box_coder)
    matcher = matcher_builder.build(ssd_config.matcher)
    region_similarity_calculator = sim_calc.build(
        ssd_config.similarity_calculator)
    encode_background_as_zeros = ssd_config.encode_background_as_zeros
    negative_class_weight = ssd_config.negative_class_weight
    ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                    ssd_config.box_predictor,
                                                    is_training, num_classes)
    anchor_generator = anchor_generator_builder.build(
        ssd_config.anchor_generator)
    image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
    non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
        ssd_config.post_processing)
    (classification_loss, localization_loss, classification_weight,
     localization_weight, hard_example_miner,
     random_example_sampler) = losses_builder.build(ssd_config.loss)
    normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
    normalize_loc_loss_by_codesize = ssd_config.normalize_loc_loss_by_codesize

    return ssd_meta_arch.SSDMetaArch(
        is_training,
        anchor_generator,
        ssd_box_predictor,
        box_coder,
        feature_extractor,
        matcher,
        region_similarity_calculator,
        encode_background_as_zeros,
        negative_class_weight,
        image_resizer_fn,
        non_max_suppression_fn,
        score_conversion_fn,
        classification_loss,
        localization_loss,
        classification_weight,
        localization_weight,
        normalize_loss_by_num_matches,
        hard_example_miner,
        add_summaries=add_summaries,
        normalize_loc_loss_by_codesize=normalize_loc_loss_by_codesize,
        freeze_batchnorm=ssd_config.freeze_batchnorm,
        inplace_batchnorm_update=ssd_config.inplace_batchnorm_update,
        add_background_class=add_background_class,
        random_example_sampler=random_example_sampler)
Beispiel #37
0
def _build_ssd_model(ssd_config, is_training):
  """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.

  Returns:
    SSDMetaArch based on the config.
  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
  num_classes = ssd_config.num_classes  #number of clases 

  # Feature extractor
  feature_extractor = _build_ssd_feature_extractor(ssd_config.feature_extractor,      #we use ssd_mobilenet_v1 as the feature extractor 
                                                   is_training)    #set the class in ssd_mobilenr_v1_feature_extractor amd ssd_meta+arch.py 

#when taking the regression loss we are working with some transorfmation. That means our predictors will predict 4 cordinates and those codinates should be regressed with some kind embedding which was made with ground truth boxes and default boxes , then after getting those we docode them for real images 


  box_coder = box_coder_builder.build(ssd_config.box_coder) #set en encoding w.r.t ground truth boxes and achor boxes . The output creating with this object will then regressed with the predicted onece. chenck equation 2 in the ssd paper 
  matcher = matcher_builder.build(ssd_config.matcher) #matching the predicted to ground trunth- Builds a matcher object based on the matcher config
#in obove object matching is done with default boxes and ground truth boxes , that's how xij value in the paper obtained . 

  region_similarity_calculator = sim_calc.build(         #how to calculate the similarity parameter is iou .
      ssd_config.similarity_calculator)

  ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,    #This will take care of the convolutional kernal 
                                                  ssd_config.box_predictor,    
                                                  is_training, num_classes)  #this returns a box_predictor object 


  anchor_generator = anchor_generator_builder.build(         #pass an instance or object where we can create ancho boxes for differen featuremaps
      ssd_config.anchor_generator)

  image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)    #this is imortatnt  we use   fixed_shape_resizer

  non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(   #this is to work with NMS supression  output
      ssd_config.post_processing)     #score conversion function will convert logits to probabilities 

  (classification_loss, localization_loss, classification_weight,
   localization_weight,
   hard_example_miner) = losses_builder.build(ssd_config.loss)           #now the loss for hard examples  these outputs are objects 

  normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches # we devide by the matching acnhorboxes 

  return ssd_meta_arch.SSDMetaArch(        #here we initialized a object of ssd_meta_arch which will be used in trainign 
      is_training,
      anchor_generator,
      ssd_box_predictor,
      box_coder,
      feature_extractor,
      matcher,
      region_similarity_calculator,
      image_resizer_fn,
      non_max_suppression_fn,
      score_conversion_fn,
      classification_loss,
      localization_loss,
      classification_weight,
      localization_weight,
      normalize_loss_by_num_matches,
      hard_example_miner)
Beispiel #38
0
def _build_lstm_model(ssd_config, lstm_config, is_training):
  """Builds an LSTM detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      LSTMMetaArch.
    lstm_config: LstmModel config proto that specifies LSTM train/eval configs.
    is_training: True if this model is being built for training purposes.

  Returns:
    LSTMMetaArch based on the config.
  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map), or if lstm_config.interleave_strategy is not recognized.
    ValueError: If unroll_length is not specified in the config file.
  """
  feature_extractor = _build_lstm_feature_extractor(
      ssd_config.feature_extractor, is_training, lstm_config.lstm_state_depth)

  box_coder = box_coder_builder.build(ssd_config.box_coder)
  matcher = matcher_builder.build(ssd_config.matcher)
  region_similarity_calculator = sim_calc.build(
      ssd_config.similarity_calculator)

  num_classes = ssd_config.num_classes
  ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                  ssd_config.box_predictor,
                                                  is_training, num_classes)
  anchor_generator = anchor_generator_builder.build(ssd_config.anchor_generator)
  image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
  non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
      ssd_config.post_processing)
  (classification_loss, localization_loss, classification_weight,
   localization_weight, miner, _, _) = losses_builder.build(ssd_config.loss)

  normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
  encode_background_as_zeros = ssd_config.encode_background_as_zeros
  negative_class_weight = ssd_config.negative_class_weight

  # Extra configs for lstm unroll length.
  unroll_length = None
  if 'lstm' in ssd_config.feature_extractor.type:
    if is_training:
      unroll_length = lstm_config.train_unroll_length
    else:
      unroll_length = lstm_config.eval_unroll_length
  if unroll_length is None:
    raise ValueError('No unroll length found in the config file')

  target_assigner_instance = target_assigner.TargetAssigner(
      region_similarity_calculator,
      matcher,
      box_coder,
      negative_class_weight=negative_class_weight)

  lstm_model = lstm_meta_arch.LSTMMetaArch(
      is_training=is_training,
      anchor_generator=anchor_generator,
      box_predictor=ssd_box_predictor,
      box_coder=box_coder,
      feature_extractor=feature_extractor,
      encode_background_as_zeros=encode_background_as_zeros,
      image_resizer_fn=image_resizer_fn,
      non_max_suppression_fn=non_max_suppression_fn,
      score_conversion_fn=score_conversion_fn,
      classification_loss=classification_loss,
      localization_loss=localization_loss,
      classification_loss_weight=classification_weight,
      localization_loss_weight=localization_weight,
      normalize_loss_by_num_matches=normalize_loss_by_num_matches,
      hard_example_miner=miner,
      unroll_length=unroll_length,
      target_assigner_instance=target_assigner_instance)

  return lstm_model
Beispiel #39
0
def _build_ssd_model(ssd_config, is_training, add_summaries,
                     add_background_class=True):
  """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.
    add_summaries: Whether to add tf summaries in the model.
    add_background_class: Whether to add an implicit background class to one-hot
      encodings of groundtruth labels. Set to false if using groundtruth labels
      with an explicit background class or using multiclass scores instead of
      truth in the case of distillation.
  Returns:
    SSDMetaArch based on the config.

  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
  num_classes = ssd_config.num_classes

  # Feature extractor
  feature_extractor = _build_ssd_feature_extractor(
      feature_extractor_config=ssd_config.feature_extractor,
      is_training=is_training)

  box_coder = box_coder_builder.build(ssd_config.box_coder)
  matcher = matcher_builder.build(ssd_config.matcher)
  region_similarity_calculator = sim_calc.build(
      ssd_config.similarity_calculator)
  encode_background_as_zeros = ssd_config.encode_background_as_zeros
  negative_class_weight = ssd_config.negative_class_weight
  ssd_box_predictor = box_predictor_builder.build(hyperparams_builder.build,
                                                  ssd_config.box_predictor,
                                                  is_training, num_classes)
  anchor_generator = anchor_generator_builder.build(
      ssd_config.anchor_generator)
  image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
  non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
      ssd_config.post_processing)
  (classification_loss, localization_loss, classification_weight,
   localization_weight,
   hard_example_miner) = losses_builder.build(ssd_config.loss)
  normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
  normalize_loc_loss_by_codesize = ssd_config.normalize_loc_loss_by_codesize

  return ssd_meta_arch.SSDMetaArch(
      is_training,
      anchor_generator,
      ssd_box_predictor,
      box_coder,
      feature_extractor,
      matcher,
      region_similarity_calculator,
      encode_background_as_zeros,
      negative_class_weight,
      image_resizer_fn,
      non_max_suppression_fn,
      score_conversion_fn,
      classification_loss,
      localization_loss,
      classification_weight,
      localization_weight,
      normalize_loss_by_num_matches,
      hard_example_miner,
      add_summaries=add_summaries,
      normalize_loc_loss_by_codesize=normalize_loc_loss_by_codesize,
      freeze_batchnorm=ssd_config.freeze_batchnorm,
      inplace_batchnorm_update=ssd_config.inplace_batchnorm_update,
      add_background_class=add_background_class)
Beispiel #40
0
def _build_ssd_model(ssd_config, is_training, add_summaries):
  """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.
    add_summaries: Whether to add tf summaries in the model.
  Returns:
    SSDMetaArch based on the config.

  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
  num_classes = ssd_config.num_classes

  # Feature extractor
  feature_extractor = _build_ssd_feature_extractor(
      feature_extractor_config=ssd_config.feature_extractor,
      freeze_batchnorm=ssd_config.freeze_batchnorm,
      is_training=is_training)

  box_coder = box_coder_builder.build(ssd_config.box_coder)
  matcher = matcher_builder.build(ssd_config.matcher)
  region_similarity_calculator = sim_calc.build(
      ssd_config.similarity_calculator)
  encode_background_as_zeros = ssd_config.encode_background_as_zeros
  negative_class_weight = ssd_config.negative_class_weight
  anchor_generator = anchor_generator_builder.build(
      ssd_config.anchor_generator)
  if feature_extractor.is_keras_model:
    ssd_box_predictor = box_predictor_builder.build_keras(
        conv_hyperparams_fn=hyperparams_builder.KerasLayerHyperparams,
        freeze_batchnorm=ssd_config.freeze_batchnorm,
        inplace_batchnorm_update=False,
        num_predictions_per_location_list=anchor_generator
        .num_anchors_per_location(),
        box_predictor_config=ssd_config.box_predictor,
        is_training=is_training,
        num_classes=num_classes,
        add_background_class=ssd_config.add_background_class)
  else:
    ssd_box_predictor = box_predictor_builder.build(
        hyperparams_builder.build, ssd_config.box_predictor, is_training,
        num_classes, ssd_config.add_background_class)
  image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
  non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
      ssd_config.post_processing)
  (classification_loss, localization_loss, classification_weight,
   localization_weight, hard_example_miner,
   random_example_sampler) = losses_builder.build(ssd_config.loss)
  normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
  normalize_loc_loss_by_codesize = ssd_config.normalize_loc_loss_by_codesize
  weight_regression_loss_by_score = (ssd_config.weight_regression_loss_by_score)

  target_assigner_instance = target_assigner.TargetAssigner(
      region_similarity_calculator,
      matcher,
      box_coder,
      negative_class_weight=negative_class_weight,
      weight_regression_loss_by_score=weight_regression_loss_by_score)

  expected_classification_loss_under_sampling = None
  if ssd_config.use_expected_classification_loss_under_sampling:
    expected_classification_loss_under_sampling = functools.partial(
        ops.expected_classification_loss_under_sampling,
        min_num_negative_samples=ssd_config.min_num_negative_samples,
        desired_negative_sampling_ratio=ssd_config.
        desired_negative_sampling_ratio)

  ssd_meta_arch_fn = ssd_meta_arch.SSDMetaArch

  return ssd_meta_arch_fn(
      is_training=is_training,
      anchor_generator=anchor_generator,
      box_predictor=ssd_box_predictor,
      box_coder=box_coder,
      feature_extractor=feature_extractor,
      encode_background_as_zeros=encode_background_as_zeros,
      image_resizer_fn=image_resizer_fn,
      non_max_suppression_fn=non_max_suppression_fn,
      score_conversion_fn=score_conversion_fn,
      classification_loss=classification_loss,
      localization_loss=localization_loss,
      classification_loss_weight=classification_weight,
      localization_loss_weight=localization_weight,
      normalize_loss_by_num_matches=normalize_loss_by_num_matches,
      hard_example_miner=hard_example_miner,
      target_assigner_instance=target_assigner_instance,
      add_summaries=add_summaries,
      normalize_loc_loss_by_codesize=normalize_loc_loss_by_codesize,
      freeze_batchnorm=ssd_config.freeze_batchnorm,
      inplace_batchnorm_update=ssd_config.inplace_batchnorm_update,
      add_background_class=ssd_config.add_background_class,
      random_example_sampler=random_example_sampler,
      expected_classification_loss_under_sampling=
      expected_classification_loss_under_sampling)
Beispiel #41
0
def _build_ssd_model(ssd_config, is_training, add_summaries):
    """Builds an SSD detection model based on the model config.

  Args:
    ssd_config: A ssd.proto object containing the config for the desired
      SSDMetaArch.
    is_training: True if this model is being built for training purposes.
    add_summaries: Whether to add tf summaries in the model.
  Returns:
    SSDMetaArch based on the config.

  Raises:
    ValueError: If ssd_config.type is not recognized (i.e. not registered in
      model_class_map).
  """
    num_classes = ssd_config.num_classes

    # Feature extractor
    feature_extractor = _build_ssd_feature_extractor(
        feature_extractor_config=ssd_config.feature_extractor,
        freeze_batchnorm=ssd_config.freeze_batchnorm,
        is_training=is_training)

    box_coder = box_coder_builder.build(ssd_config.box_coder)
    matcher = matcher_builder.build(ssd_config.matcher)
    region_similarity_calculator = sim_calc.build(
        ssd_config.similarity_calculator)
    encode_background_as_zeros = ssd_config.encode_background_as_zeros
    negative_class_weight = ssd_config.negative_class_weight
    anchor_generator = anchor_generator_builder.build(
        ssd_config.anchor_generator)
    if feature_extractor.is_keras_model:
        ssd_box_predictor = box_predictor_builder.build_keras(
            hyperparams_fn=hyperparams_builder.KerasLayerHyperparams,
            freeze_batchnorm=ssd_config.freeze_batchnorm,
            inplace_batchnorm_update=False,
            num_predictions_per_location_list=anchor_generator.
            num_anchors_per_location(),
            box_predictor_config=ssd_config.box_predictor,
            is_training=is_training,
            num_classes=num_classes,
            add_background_class=ssd_config.add_background_class)
    else:
        ssd_box_predictor = box_predictor_builder.build(
            hyperparams_builder.build, ssd_config.box_predictor, is_training,
            num_classes, ssd_config.add_background_class)
    image_resizer_fn = image_resizer_builder.build(ssd_config.image_resizer)
    non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
        ssd_config.post_processing)
    (classification_loss, localization_loss, classification_weight,
     localization_weight, hard_example_miner, random_example_sampler,
     expected_loss_weights_fn) = losses_builder.build(ssd_config.loss)
    normalize_loss_by_num_matches = ssd_config.normalize_loss_by_num_matches
    normalize_loc_loss_by_codesize = ssd_config.normalize_loc_loss_by_codesize

    equalization_loss_config = ops.EqualizationLossConfig(
        weight=ssd_config.loss.equalization_loss.weight,
        exclude_prefixes=ssd_config.loss.equalization_loss.exclude_prefixes)

    target_assigner_instance = target_assigner.TargetAssigner(
        region_similarity_calculator,
        matcher,
        box_coder,
        negative_class_weight=negative_class_weight)

    ssd_meta_arch_fn = ssd_meta_arch.SSDMetaArch
    kwargs = {}

    return ssd_meta_arch_fn(
        is_training=is_training,
        anchor_generator=anchor_generator,
        box_predictor=ssd_box_predictor,
        box_coder=box_coder,
        feature_extractor=feature_extractor,
        encode_background_as_zeros=encode_background_as_zeros,
        image_resizer_fn=image_resizer_fn,
        non_max_suppression_fn=non_max_suppression_fn,
        score_conversion_fn=score_conversion_fn,
        classification_loss=classification_loss,
        localization_loss=localization_loss,
        classification_loss_weight=classification_weight,
        localization_loss_weight=localization_weight,
        normalize_loss_by_num_matches=normalize_loss_by_num_matches,
        hard_example_miner=hard_example_miner,
        target_assigner_instance=target_assigner_instance,
        add_summaries=add_summaries,
        normalize_loc_loss_by_codesize=normalize_loc_loss_by_codesize,
        freeze_batchnorm=ssd_config.freeze_batchnorm,
        inplace_batchnorm_update=ssd_config.inplace_batchnorm_update,
        add_background_class=ssd_config.add_background_class,
        explicit_background_class=ssd_config.explicit_background_class,
        random_example_sampler=random_example_sampler,
        expected_loss_weights_fn=expected_loss_weights_fn,
        use_confidences_as_targets=ssd_config.use_confidences_as_targets,
        implicit_example_weight=ssd_config.implicit_example_weight,
        equalization_loss_config=equalization_loss_config,
        return_raw_detections_during_predict=(
            ssd_config.return_raw_detections_during_predict),
        **kwargs)
    tf.pad(one_hot_encoding, [[0, 0], [1, 0]], mode='CONSTANT')
    for one_hot_encoding in tf.unstack(groundtruth_class)
]
groundtruth_boxlists = [
    box_list.BoxList(boxes) for boxes in tf.unstack(groundtruth_box)
]

# construct models
box_coder = box_coder_builder.build(model.ssd.box_coder)
matcher = matcher_builder.build(model.ssd.matcher)
region_similarity_calculator = sim_calc_builder.build(
    model.ssd.similarity_calculator)
anchor_generator = anchor_generator_builder.build(model.ssd.anchor_generator)
(classification_loss, localization_loss, classification_weight,
 localization_weight,
 hard_example_miner) = losses_builder.build(model.ssd.loss)
image_resizer_fn = image_resizer_builder.build(model.ssd.image_resizer)
non_max_suppression_fn, score_conversion_fn = post_processing_builder.build(
    model.ssd.post_processing)
(classification_loss, localization_loss, classification_weight,
 localization_weight,
 hard_example_miner) = losses_builder.build(model.ssd.loss)
normalize_loss_by_num_matches = model.ssd.normalize_loss_by_num_matches
matcher = matcher_builder.build(model.ssd.matcher)
unmatched_cls_target = tf.constant([1] + num_classes * [0], tf.float32)
_target_assigner = target_assigner.TargetAssigner(
    region_similarity_calculator,
    matcher,
    box_coder,
    positive_class_weight=1.0,
    negative_class_weight=1.0,