Path: blob/master/site/en-snapshot/tfx/guide/bulkinferrer.md
56931 views
The BulkInferrer TFX Pipeline Component
The BulkInferrer TFX component performs batch inference on unlabeled data. The generated InferenceResult(tensorflow_serving.apis.prediction_log_pb2.PredictionLog) contains the original features and the prediction results.
BulkInferrer consumes:
A trained model in SavedModel format.
Unlabelled tf.Examples that contain features.
(Optional) Validation result from Evaluator component.
BulkInferrer emits:
Using the BulkInferrer Component
A BulkInferrer TFX component is used to perform batch inference on unlabeled tf.Examples. It is typically deployed after an Evaluator component to perform inference with a validated model, or after a Trainer component to directly perform inference on exported model.
It currently performs in-memory model inference and remote inference. Remote inference requires the model to be hosted on Cloud AI Platform.
Typical code looks like this:
More details are available in the BulkInferrer API reference.