Batch prediction components

TheBatchPredictionJob resource lets you run an asynchronousprediction request. Request batch predictions directly from themodelresource. You don't need to deploy the model to anendpoint. For data typesthat support both batch and online predictions you can use batch predictions.This is useful when you don't require an immediate response and want to processaccumulated data by using a single request.

To make a batch prediction, specify an input source and an output locationfor Vertex AI to store predictions results. The inputs and outputsdepend on themodel type that you're working with. For example, batchpredictions for the AutoML image model type require an inputJSON Lines file and the name of a Cloud Storage bucket to store the output.For more information about batch prediction, seeGet batch predictions.

You can use theModelBatchPredictOp component to access this resource through Vertex AI Pipelines.

API reference

Tutorials

Version history and release notes

To learn more about the version history and changes to the Google Cloud Pipeline Components SDK, see theGoogle Cloud Pipeline Components SDK Release Notes.

Technical support contacts

If you have any questions, reach out tokubeflow-pipelines-components@google.com.

Except as otherwise noted, the content of this page is licensed under theCreative Commons Attribution 4.0 License, and code samples are licensed under theApache 2.0 License. For details, see theGoogle Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2026-02-18 UTC.