In the TFX pipelines, how do we make use of BulkInferrer
?
It is quite standard to connect the BulkInferrer
with a trained model
or pushed_model
. However, what if I don't want to train the model again, instead I would love to use a previously trained model
or pushed_model
and use the BulkInferrer
for batch inference(kinda like serving with BulkInferrer
). Is it possible to do that?
If not, what is the purpose of BulkInferrer
component, just to do a one-time prediction or validation after the whole training?
Any suggestions or comments will be appreciated, thanks.
BulkInferrer is the newest component added in the TFX library to support the batch inference of unlabelled data.
Robert Crowe nicely positioned BulkInferrer in the inference pipeline, during the recent BEAM summit
Here is a list of use cases of why would someone use the BulkInferrer, trying to approach it in the case of ml-pipelines rather than data-pipelines:
To do this in your ML pipeline without retraining your model, you can include BulkInferrer indeed at the end of the pipeline and reuse the results from previous runs if the inputs and configuration has not changed. This is achieved by both Argo and Tekton workflow managers on Kubeflow pipelines, as they implement TFX, see step caching.