The Dataproc Serverless components let you run Apache Spark batch workloads from a pipeline within Vertex AI Pipelines. Dataproc Serverless runs the batch workloads on a managed compute infrastructure, autoscaling resources as needed.
Learn more about Dataproc Serverless and supported Spark workloads.
In Dataproc Serverless, a Batch
resource represents a batch workload.
The Google Cloud Pipeline Components SDK includes the following operators to
create Batch
resources and monitor their execution:
API reference
For component reference, see the Google Cloud Pipeline Components SDK reference for Dataproc Serverless components .
For Dataproc Serverless resource reference, see the following API reference page:
Batch
resource
Tutorials
Version history and release notes
To learn more about the version history and changes to the Google Cloud Pipeline Components SDK, see the Google Cloud Pipeline Components SDK Release Notes.
Technical support contacts
If you have any questions, reach out to [email protected].