示例#1
0
    def testEstimatorModelPredict(self):
        example_path = self._get_output_data_dir('examples')
        self._prepare_predict_examples(example_path)
        model_path = self._get_output_data_dir('model')
        self._build_predict_model(model_path)
        prediction_log_path = self._get_output_data_dir('predictions')
        self._run_inference_with_beam(
            example_path,
            model_spec_pb2.InferenceSpecType(
                saved_model_spec=model_spec_pb2.SavedModelSpec(
                    model_path=model_path)), prediction_log_path)

        results = self._get_results(prediction_log_path)
        self.assertLen(results, 2)
        self.assertEqual(
            results[0].predict_log.request.inputs[
                run_inference._DEFAULT_INPUT_KEY].string_val[0],
            self._predict_examples[0].SerializeToString())
        self.assertEqual(results[0].predict_log.response.outputs['y'].dtype,
                         tf.float32)
        self.assertLen(
            results[0].predict_log.response.outputs['y'].tensor_shape.dim, 2)
        self.assertEqual(
            results[0].predict_log.response.outputs['y'].tensor_shape.dim[0].
            size, 1)
        self.assertEqual(
            results[0].predict_log.response.outputs['y'].tensor_shape.dim[1].
            size, 1)
def run(args, pipeline_args):
    """
    Run inference pipeline using data generated from streaming pipeline.
    """
    pipeline_options = PipelineOptions(pipeline_args,
                                       save_main_session=True,
                                       streaming=False)

    with beam.Pipeline(options=pipeline_options) as pipeline:
        _ = (pipeline
             | 'ReadTFExample' >> beam.io.tfrecordio.ReadFromTFRecord(
                 file_pattern=args.tfrecord_folder)
             | 'ParseExamples' >> beam.Map(tf.train.Example.FromString)
             | RunInference(
                 model_spec_pb2.InferenceSpecType(
                     saved_model_spec=model_spec_pb2.SavedModelSpec(
                         signature_name=['serving_default'],
                         model_path=args.saved_model_location)))
             | beam.ParDo(
                 process_encdec_inf_rtn.ProcessReturn(
                     config={
                         'tf_transform_graph_dir': args.tf_transform_graph_dir,
                         'model_config': config.MODEL_CONFIG
                     }))
             | beam.ParDo(
                 process_encdec_inf_rtn.CheckAnomalous(threshold=0.15))
             | beam.ParDo(print))
示例#3
0
    def testClassifyModel(self):
        example_path = self._get_output_data_dir('examples')
        self._prepare_multihead_examples(example_path)
        model_path = self._get_output_data_dir('model')
        self._build_multihead_model(model_path)
        prediction_log_path = self._get_output_data_dir('predictions')
        self._run_inference_with_beam(
            example_path,
            model_spec_pb2.InferenceSpecType(
                saved_model_spec=model_spec_pb2.SavedModelSpec(
                    model_path=model_path, signature_name=['classify_sum'])),
            prediction_log_path)

        results = self._get_results(prediction_log_path)
        self.assertLen(results, 2)
        classify_log = results[0].classify_log
        self.assertLen(classify_log.request.input.example_list.examples, 1)
        self.assertEqual(classify_log.request.input.example_list.examples[0],
                         self._multihead_examples[0])
        self.assertLen(classify_log.response.result.classifications, 1)
        self.assertLen(classify_log.response.result.classifications[0].classes,
                       1)
        self.assertAlmostEqual(
            classify_log.response.result.classifications[0].classes[0].score,
            1.0)
示例#4
0
 def testMultiInferenceModel(self):
   example_path = self._get_output_data_dir('examples')
   self._prepare_multihead_examples(example_path)
   model_path = self._get_output_data_dir('model')
   self._build_multihead_model(model_path)
   prediction_log_path = self._get_output_data_dir('predictions')
   self._run_inference_with_beam(
       example_path,
       model_spec_pb2.InferenceSpecType(
           saved_model_spec=model_spec_pb2.SavedModelSpec(
               model_path=model_path,
               signature_name=['regress_diff', 'classify_sum'])),
       prediction_log_path)
   results = self._get_results(prediction_log_path)
   self.assertLen(results, 2)
   multi_inference_log = results[0].multi_inference_log
   self.assertLen(multi_inference_log.request.input.example_list.examples, 1)
   self.assertEqual(multi_inference_log.request.input.example_list.examples[0],
                    self._multihead_examples[0])
   self.assertLen(multi_inference_log.response.results, 2)
   signature_names = []
   for result in multi_inference_log.response.results:
     signature_names.append(result.model_spec.signature_name)
   self.assertIn('regress_diff', signature_names)
   self.assertIn('classify_sum', signature_names)
   result = multi_inference_log.response.results[0]
   self.assertEqual(result.model_spec.signature_name, 'regress_diff')
   self.assertLen(result.regression_result.regressions, 1)
   self.assertAlmostEqual(result.regression_result.regressions[0].value, 0.6)
   result = multi_inference_log.response.results[1]
   self.assertEqual(result.model_spec.signature_name, 'classify_sum')
   self.assertLen(result.classification_result.classifications, 1)
   self.assertLen(result.classification_result.classifications[0].classes, 1)
   self.assertAlmostEqual(
       result.classification_result.classifications[0].classes[0].score, 1.0)
示例#5
0
def run(args, pipeline_args):
    """
    Run inference pipeline using data generated from streaming pipeline.
    """
    pipeline_options = PipelineOptions(pipeline_args,
                                       save_main_session=True,
                                       streaming=True)

    with beam.Pipeline(options=pipeline_options) as pipeline:
        _ = (
            pipeline
            | 'ReadTFExample' >> beam.io.gcp.pubsub.ReadStringsFromPubSub(
                subscription=args.pubsub_subscription)
            |
            'ParseExamples' >> beam.Map(lambda x: Parse(x, tf.train.Example()))
            | RunInference(
                model_spec_pb2.InferenceSpecType(
                    saved_model_spec=model_spec_pb2.SavedModelSpec(
                        signature_name=['serving_default'],
                        model_path=args.saved_model_location)))
            | beam.ParDo(
                process_encdec_inf_rtn.ProcessReturn(
                    config={
                        'tf_transform_graph_dir': args.tf_transform_graph_dir,
                        'model_config': config.MODEL_CONFIG
                    }))
            | beam.ParDo(process_encdec_inf_rtn.CheckAnomalous(threshold=0.7))
            | beam.ParDo(print))
示例#6
0
 def testModelPathInvalid(self):
   example_path = self._get_output_data_dir('examples')
   self._prepare_predict_examples(example_path)
   prediction_log_path = self._get_output_data_dir('predictions')
   with self.assertRaisesRegexp(IOError, 'SavedModel file does not exist.*'):
     self._run_inference_with_beam(
         example_path,
         model_spec_pb2.InferenceSpecType(
             saved_model_spec=model_spec_pb2.SavedModelSpec(
                 model_path=self._get_output_data_dir())), prediction_log_path)
示例#7
0
    def _run_model_inference(self, model_path: Text,
                             example_uris: Mapping[Text,
                                                   Text], output_path: Text,
                             model_spec: bulk_inferrer_pb2.ModelSpec) -> None:
        """Runs model inference on given example data.

    Args:
      model_path: Path to model.
      example_uris: Mapping of example split name to example uri.
      output_path: Path to output generated prediction logs.
      model_spec: bulk_inferrer_pb2.ModelSpec instance.

    Returns:
      None
    """

        try:
            from tfx_bsl.public.beam import run_inference
            from tfx_bsl.public.proto import model_spec_pb2
        except ImportError:
            # TODO(b/151468119): Remove this branch after next release.
            run_inference = importlib.import_module(
                'tfx_bsl.beam.run_inference')
            model_spec_pb2 = importlib.import_module(
                'tfx_bsl.proto.model_spec_pb2')
        saved_model_spec = model_spec_pb2.SavedModelSpec(
            model_path=model_path,
            tag=model_spec.tag,
            signature_name=model_spec.model_signature_name)
        # TODO(b/151468119): Remove this branch after next release.
        if getattr(model_spec_pb2, 'InferenceEndpoint', False):
            inference_endpoint = getattr(model_spec_pb2, 'InferenceEndpoint')()
        else:
            inference_endpoint = model_spec_pb2.InferenceSpecType()
        inference_endpoint.saved_model_spec.CopyFrom(saved_model_spec)
        with self._make_beam_pipeline() as pipeline:
            data_list = []
            for split, example_uri in example_uris.items():
                data = (
                    pipeline
                    | 'ReadData[{}]'.format(split) >> beam.io.ReadFromTFRecord(
                        file_pattern=io_utils.all_files_pattern(example_uri)))
                data_list.append(data)
            _ = ([data for data in data_list]
                 | 'FlattenExamples' >> beam.Flatten(pipeline=pipeline)
                 | 'ParseExamples' >> beam.Map(tf.train.Example.FromString)
                 | 'RunInference' >>
                 run_inference.RunInference(inference_endpoint)
                 | 'WritePredictionLogs' >> beam.io.WriteToTFRecord(
                     output_path,
                     file_name_suffix='.gz',
                     coder=beam.coders.ProtoCoder(
                         prediction_log_pb2.PredictionLog)))
        logging.info('Inference result written to %s.', output_path)
示例#8
0
 def testInfersElementType(self, input_element, output_type):
   # TODO(zwestrick): Skip building the model, which is not actually used, or
   # stop using parameterized tests if performance becomes an issue.
   model_path = self._get_output_data_dir('model')
   self._build_predict_model(model_path)
   spec = model_spec_pb2.InferenceSpecType(
       saved_model_spec=model_spec_pb2.SavedModelSpec(model_path=model_path))
   inference_transform = run_inference.RunInferenceImpl(spec)
   with beam.Pipeline() as p:
     inference = (p | beam.Create([input_element]) | inference_transform)
     self.assertEqual(inference.element_type, output_type)
示例#9
0
 def _get_inference_spec(
     self, model_path: Text,
     exec_properties: Dict[Text, Any]) -> model_spec_pb2.InferenceSpecType:
   model_spec = bulk_inferrer_pb2.ModelSpec()
   proto_utils.json_to_proto(exec_properties['model_spec'], model_spec)
   saved_model_spec = model_spec_pb2.SavedModelSpec(
       model_path=model_path,
       tag=model_spec.tag,
       signature_name=model_spec.model_signature_name)
   result = model_spec_pb2.InferenceSpecType()
   result.saved_model_spec.CopyFrom(saved_model_spec)
   return result
示例#10
0
 def _get_inference_spec(
     self, model_path: str,
     exec_properties: Dict[str, Any]) -> model_spec_pb2.InferenceSpecType:
   model_spec = bulk_inferrer_pb2.ModelSpec()
   proto_utils.json_to_proto(
       exec_properties[standard_component_specs.MODEL_SPEC_KEY], model_spec)
   saved_model_spec = model_spec_pb2.SavedModelSpec(
       model_path=model_path,
       tag=model_spec.tag,
       signature_name=model_spec.model_signature_name)
   result = model_spec_pb2.InferenceSpecType()
   result.saved_model_spec.CopyFrom(saved_model_spec)
   return result
示例#11
0
    def testKerasModelPredict(self):
        inputs = tf.keras.Input(shape=(1, ), name='input1')
        output1 = tf.keras.layers.Dense(1,
                                        activation=tf.nn.sigmoid,
                                        name='output1')(inputs)
        output2 = tf.keras.layers.Dense(1,
                                        activation=tf.nn.sigmoid,
                                        name='output2')(inputs)
        inference_model = tf.keras.models.Model(inputs, [output1, output2])

        class TestKerasModel(tf.keras.Model):
            def __init__(self, inference_model):
                super(TestKerasModel, self).__init__(name='test_keras_model')
                self.inference_model = inference_model

            @tf.function(input_signature=[
                tf.TensorSpec(shape=[None], dtype=tf.string, name='inputs')
            ])
            def call(self, serialized_example):
                features = {
                    'input1':
                    tf.compat.v1.io.FixedLenFeature([1],
                                                    dtype=tf.float32,
                                                    default_value=0)
                }
                input_tensor_dict = tf.io.parse_example(
                    serialized_example, features)
                return inference_model(input_tensor_dict['input1'])

        model = TestKerasModel(inference_model)
        model.compile(optimizer=tf.keras.optimizers.Adam(lr=.001),
                      loss=tf.keras.losses.binary_crossentropy,
                      metrics=['accuracy'])

        model_path = self._get_output_data_dir('model')
        tf.compat.v1.keras.experimental.export_saved_model(model,
                                                           model_path,
                                                           serving_only=True)

        example_path = self._get_output_data_dir('examples')
        self._prepare_predict_examples(example_path)
        prediction_log_path = self._get_output_data_dir('predictions')
        self._run_inference_with_beam(
            example_path,
            model_spec_pb2.InferenceSpecType(
                saved_model_spec=model_spec_pb2.SavedModelSpec(
                    model_path=model_path)), prediction_log_path)

        results = self._get_results(prediction_log_path)
        self.assertLen(results, 2)
示例#12
0
    def testTelemetry(self):
        example_path = self._get_output_data_dir('examples')
        self._prepare_multihead_examples(example_path)
        model_path = self._get_output_data_dir('model')
        self._build_multihead_model(model_path)
        inference_spec_type = model_spec_pb2.InferenceSpecType(
            saved_model_spec=model_spec_pb2.SavedModelSpec(
                model_path=model_path, signature_name=['classify_sum']))
        pipeline = beam.Pipeline()
        _ = (pipeline
             | 'ReadExamples' >> beam.io.ReadFromTFRecord(example_path)
             | 'MaybeDecode' >> beam.Map(lambda x: x if _randbool() else tf.
                                         train.Example.FromString(x))
             | 'RunInference' >>
             run_inference.RunInferenceImpl(inference_spec_type))
        run_result = pipeline.run()
        run_result.wait_until_finish()

        num_inferences = run_result.metrics().query(
            MetricsFilter().with_name('num_inferences'))
        self.assertTrue(num_inferences['counters'])
        self.assertEqual(num_inferences['counters'][0].result, 2)
        num_instances = run_result.metrics().query(
            MetricsFilter().with_name('num_instances'))
        self.assertTrue(num_instances['counters'])
        self.assertEqual(num_instances['counters'][0].result, 2)
        inference_request_batch_size = run_result.metrics().query(
            MetricsFilter().with_name('inference_request_batch_size'))
        self.assertTrue(inference_request_batch_size['distributions'])
        self.assertEqual(
            inference_request_batch_size['distributions'][0].result.sum, 2)
        inference_request_batch_byte_size = run_result.metrics().query(
            MetricsFilter().with_name('inference_request_batch_byte_size'))
        self.assertTrue(inference_request_batch_byte_size['distributions'])
        self.assertEqual(
            inference_request_batch_byte_size['distributions'][0].result.sum,
            sum(element.ByteSize() for element in self._multihead_examples))
        inference_batch_latency_micro_secs = run_result.metrics().query(
            MetricsFilter().with_name('inference_batch_latency_micro_secs'))
        self.assertTrue(inference_batch_latency_micro_secs['distributions'])
        self.assertGreaterEqual(
            inference_batch_latency_micro_secs['distributions'][0].result.sum,
            0)
        load_model_latency_milli_secs = run_result.metrics().query(
            MetricsFilter().with_name('load_model_latency_milli_secs'))
        self.assertTrue(load_model_latency_milli_secs['distributions'])
        self.assertGreaterEqual(
            load_model_latency_milli_secs['distributions'][0].result.sum, 0)
示例#13
0
  def test_basic(self):
    examples = [
        text_format.Parse(
            """
              features {
                feature { key: "x" value { float_list { value: 0 }}}
              }
              """, tf.train.Example()),
        text_format.Parse(
            """
              features {
                feature { key: "x" value { float_list { value: 1 }}}
              }
              """, tf.train.Example())
    ]
    model_paths = [self._get_output_data_dir(m) for m in ('model1', 'model2')]
    for model_path in model_paths:
      self._build_predict_model(model_path)
    specs = [
        model_spec_pb2.InferenceSpecType(
            saved_model_spec=model_spec_pb2.SavedModelSpec(model_path=p))
        for p in model_paths
    ]
    with self._make_beam_pipeline() as pipeline:
      predictions = (
          pipeline
          | beam.Create(examples)
          | run_inference.RunInferencePerModelImpl(specs)
          | beam.MapTuple(
              lambda _, p2: p2.predict_log.response.outputs['y'].float_val[0]))
      assert_that(predictions, equal_to([0.0, 2.0]))

      predictions_table = (
          pipeline
          |
          'CreateTable' >> beam.Create([(i, e) for i, e in enumerate(examples)])
          | 'RunInferencePerModelTable' >>
          run_inference.RunInferencePerModelImpl(specs)
          | beam.MapTuple(lambda k, v:  # pylint: disable=g-long-lambda
                          (k, v[1].predict_log.response.outputs['y'].float_val[
                              0])))
      assert_that(
          predictions_table,
          equal_to([(0, 0.0), (1, 2.0)]),
          label='AssertTable')
示例#14
0
  def testRegressModel(self, keyed_input: bool, decode_examples: bool):
    example_path = self._get_output_data_dir('examples')
    self._prepare_multihead_examples(example_path)
    model_path = self._get_output_data_dir('model')
    self._build_multihead_model(model_path)
    prediction_log_path = self._get_output_data_dir('predictions')
    self._run_inference_with_beam(
        example_path,
        model_spec_pb2.InferenceSpecType(
            saved_model_spec=model_spec_pb2.SavedModelSpec(
                model_path=model_path, signature_name=['regress_diff'])),
        prediction_log_path, keyed_input, decode_examples)

    results = self._get_results(prediction_log_path)
    self.assertLen(results, 2)
    regress_log = results[0].regress_log
    self.assertLen(regress_log.request.input.example_list.examples, 1)
    self.assertEqual(regress_log.request.input.example_list.examples[0],
                     self._multihead_examples[0])
    self.assertLen(regress_log.response.result.regressions, 1)
    self.assertAlmostEqual(regress_log.response.result.regressions[0].value,
                           0.6)
def run(args, pipeline_args):
    """
    Run inference pipeline using data generated from streaming pipeline.
    """
    pipeline_options = PipelineOptions(pipeline_args,
                                       save_main_session=True,
                                       streaming=False)

    with beam.Pipeline(options=pipeline_options) as pipeline:
        _ = (pipeline
             | 'ReadTFExample' >> beam.io.tfrecordio.ReadFromTFRecord(
                 file_pattern=args.tfrecord_folder)
             | 'ParseExamples' >> beam.Map(tf.train.Example.FromString)
             | beam.ParDo(filter_examples.FilterExamples())
             | RunInference(
                 model_spec_pb2.InferenceSpecType(
                     saved_model_spec=model_spec_pb2.SavedModelSpec(
                         signature_name=['serving_default'],
                         model_path=args.saved_model_location)))
             | beam.ParDo(process_inference_return.ProcessReturn())
             | beam.ParDo(process_inference_return.CheckAnomalous())
             | beam.ParDo(print))
示例#16
0
    def Do(self, input_dict: Dict[Text, List[types.Artifact]],
           output_dict: Dict[Text, List[types.Artifact]],
           exec_properties: Dict[Text, Any]) -> None:
        """Runs batch inference on a given model with given input examples.

        Args:
          input_dict: Input dict from input key to a list of Artifacts.
            - examples: examples for inference.
            - model: exported model.
            - model_blessing: model blessing result, optional.
          output_dict: Output dict from output key to a list of Artifacts.
            - output: bulk inference results.
          exec_properties: A dict of execution properties.
            - model_spec: JSON string of bulk_inferrer_pb2.ModelSpec instance.
            - data_spec: JSON string of bulk_inferrer_pb2.DataSpec instance.

        Returns:
          None
        """
        self._log_startup(input_dict, output_dict, exec_properties)

        source = exec_properties[StepKeys.SOURCE]
        args = exec_properties[StepKeys.ARGS]
        c = source_utils.load_source_path_class(source)
        inferrer_step: BaseInferrer = c(**args)

        output_examples = artifact_utils.get_single_instance(
            output_dict[PREDICTIONS])

        if EXAMPLES not in input_dict:
            raise ValueError('\'examples\' is missing in input dict.')
        if MODEL not in input_dict:
            raise ValueError('Input models are not valid, model '
                             'need to be specified.')
        if MODEL_BLESSING in input_dict:
            model_blessing = artifact_utils.get_single_instance(
                input_dict['model_blessing'])
            if not model_utils.is_model_blessed(model_blessing):
                logging.info('Model on %s was not blessed', model_blessing.uri)
                return
        else:
            logging.info(
                'Model blessing is not provided, exported model will be '
                'used.')

        model = artifact_utils.get_single_instance(input_dict[MODEL])
        model_path = path_utils.serving_model_path(model.uri)
        logging.info('Use exported model from %s.', model_path)

        output_example_spec = bulk_inferrer_pb2.OutputExampleSpec(
            output_columns_spec=[
                bulk_inferrer_pb2.OutputColumnsSpec(
                    predict_output=bulk_inferrer_pb2.PredictOutput(
                        output_columns=[
                            bulk_inferrer_pb2.PredictOutputCol(
                                output_key=x,
                                output_column=f'{x}_label',
                            ) for x in inferrer_step.get_labels()
                        ]))
            ])

        model_spec = bulk_inferrer_pb2.ModelSpec()
        saved_model_spec = model_spec_pb2.SavedModelSpec(
            model_path=model_path,
            tag=model_spec.tag,
            signature_name=model_spec.model_signature_name)
        inference_spec = model_spec_pb2.InferenceSpecType()
        inference_spec.saved_model_spec.CopyFrom(saved_model_spec)

        self._run_model_inference(output_example_spec, input_dict[EXAMPLES],
                                  output_examples, inference_spec,
                                  inferrer_step)