Marketplace
BrowsePublish
Marketplace
You are viewing an outdated version of provider-gcp.Go to Latest
upbound/provider-gcp@v0.26.0
WorkflowTemplate
dataproc.gcp.upbound.io
WorkflowTemplate
upbound/provider-gcp@v0.26.0dataproc.gcp.upbound.io

WorkflowTemplate is the Schema for the WorkflowTemplates API. A Workflow Template is a reusable workflow configuration.

Type

CRD

Group

dataproc.gcp.upbound.io

Version

v1beta1

apiVersion: dataproc.gcp.upbound.io/v1beta1

kind: WorkflowTemplate

API Documentation
apiVersion
string
kind
string
metadata
object
spec
object
object

WorkflowTemplateSpec defines the desired state of WorkflowTemplate

forProvider
requiredobject
requiredobject

No description provided.

jobs
requiredarray
requiredarray

Required. The Directed Acyclic Graph of Jobs to submit.

array

Optional. Job is a Hadoop job.

array

Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

args
array
array

Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.

array

Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.

array

Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

array

Optional. The runtime log config for job execution.

mainClass
string
hiveJob
array
array

Optional. Job is a Hive job.

array

Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

array

A list of queries.

queries
requiredarray
requiredarray

Required. The queries to execute. You do not need to end a query expression with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of a Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": } }

labels
object
pigJob
array
array

Optional. Job is a Pig job.

array

Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

array

Optional. The runtime log config for job execution.

array

A list of queries.

queries
requiredarray
requiredarray

Required. The queries to execute. You do not need to end a query expression with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of a Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": } }

array

Optional. The optional list of prerequisite job step_ids. If not specified, the job will start at the beginning of workflow.

array

Optional. Job is a Presto job.

array

Optional. Presto client tags to attach to this query

array

Optional. The runtime log config for job execution.

array

A list of queries.

queries
requiredarray
requiredarray

Required. The queries to execute. You do not need to end a query expression with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of a Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": } }

array

Optional. Job is a PySpark job.

array

Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

args
array
array

Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.

array

Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.

array

Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

array

Optional. The runtime log config for job execution.

mainPythonFileUri
requiredstring
array

Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.

array

Optional. Job scheduling configuration.

array

Optional. Job is a Spark job.

array

Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

args
array
array

Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.

array

Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.

array

Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

array

Optional. The runtime log config for job execution.

mainClass
string
array

Optional. Job is a SparkR job.

array

Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.

args
array
array

Optional. The arguments to pass to the driver. Do not include arguments, such as --conf, that can be set as job properties, since a collision may occur that causes an incorrect job submission.

array

Optional. HCFS URIs of files to be placed in the working directory of each executor. Useful for naively parallel tasks.

array

Optional. The runtime log config for job execution.

mainRFileUri
requiredstring
array

Optional. Job is a SparkSql job.

array

Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.

array

Optional. The runtime log config for job execution.

array

A list of queries.

queries
requiredarray
requiredarray

Required. The queries to execute. You do not need to end a query expression with a semicolon. Multiple queries can be specified in one string by separating each with a semicolon. Here is an example of a Dataproc API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { "queryList": { "queries": } }

stepId
requiredstring
labels
object
location
requiredstring
array

Optional. Template parameters whose values are substituted into the template. Values for parameters must be provided when the template is instantiated.

fields
requiredarray
requiredarray

Required. Paths to all fields that the parameter replaces. A field is allowed to appear in at most one parameter's list of field paths. A field path is similar in syntax to a .sparkJob.args

name
requiredstring
array

Optional. Validation rules to be applied to this parameter's value.

regex
array
array

Validation based on regular expressions.

regexes
requiredarray
requiredarray

Required. RE2 regular expressions used to validate the parameter's value. The value must match the regex in its entirety (substring matches are not sufficient).

values
array
array

Required. List of allowed values for the parameter.

values
requiredarray
requiredarray

Required. List of allowed values for the parameter.

placement
requiredarray
requiredarray

Required. WorkflowTemplate scheduling information.

array

Optional. A selector that chooses target cluster for jobs based on metadata. The selector is evaluated at the time each job is submitted.

clusterLabels
requiredobject
zone
string
array

A cluster that is managed by the workflow.

clusterName
requiredstring
config
requiredarray
requiredarray

Required. The cluster configuration.

array

Optional. Autoscaling config for the policy associated with the cluster. Cluster does not autoscale if this field is unset.

policy
string
array

Optional. Encryption settings for the cluster.

array

Optional. Port/endpoint configuration for this cluster

array

Optional. The shared Compute Engine config settings for all instances in a cluster.

metadata
object
network
string
array

Optional. Node Group Affinity for sole-tenant clusters.

nodeGroup
requiredstring
array

Optional. Reservation Affinity for consuming Zonal reservation.

key
string
values
array
array

Required. List of allowed values for the parameter.

array

Optional. The URIs of service account scopes to be included in Compute Engine instances. The following base set of scopes is always included: * https://www.googleapis.com/auth/cloud.useraccounts.readonly * https://www.googleapis.com/auth/devstorage.read_write * https://www.googleapis.com/auth/logging.write If no scopes are specified, the following defaults are also provided: * https://www.googleapis.com/auth/bigquery * https://www.googleapis.com/auth/bigtable.admin.table * https://www.googleapis.com/auth/bigtable.data * https://www.googleapis.com/auth/devstorage.full_control

array

Optional. Shielded Instance Config for clusters using Compute Engine Shielded VMs. Structure defined below.

tags
array
array

The Compute Engine tags to add to all instances (see (https://cloud.google.com/compute/docs/label-or-tag-resources#tags)).

zone
string
array

Optional. Commands to execute on each node after config is completed. By default, executables are run on master and all worker nodes. You can test a node's role metadata to run an executable on a master or worker node, as shown below using curl (you can also use wget): ROLE=$(curl -H Metadata-Flavor:Google http://metadata/computeMetadata/v1/instance/attributes/dataproc-role) if ; then ... master specific actions ... else ... worker specific actions ... fi

array

Optional. Lifecycle setting for the cluster.

array

Optional. The Compute Engine config settings for additional worker instances in a cluster.

array

Optional. The Compute Engine accelerator configuration for these instances.

array

Optional. Disk option config settings.

image
string
array

Optional. The Compute Engine config settings for additional worker instances in a cluster.

array

Optional. The Compute Engine accelerator configuration for these instances.

array

Optional. Disk option config settings.

image
string
array

Optional. Security settings for the cluster.

array

Optional. The config settings for software inside the cluster.

array

No description provided.

array

Optional. The Compute Engine config settings for additional worker instances in a cluster.

array

Optional. The Compute Engine accelerator configuration for these instances.

array

Optional. Disk option config settings.

image
string
labels
object
project
string
version
number
object

ProviderConfigReference specifies how the provider that will be used to create, observe, update, and delete this managed resource should be configured.

name
requiredstring
policy
object
object

Policies for referencing.

resolve
string
object

ProviderReference specifies the provider that will be used to create, observe, update, and delete this managed resource. Deprecated: Please use ProviderConfigReference, i.e. providerConfigRef

name
requiredstring
policy
object
object

Policies for referencing.

resolve
string
object

PublishConnectionDetailsTo specifies the connection secret config which contains a name, metadata and a reference to secret store config to which any connection details for this managed resource should be written. Connection details frequently include the endpoint, username, and password required to connect to the managed resource.

configRef
object
object

SecretStoreConfigRef specifies which secret store config should be used for this ConnectionSecret.

name
requiredstring
policy
object
object

Policies for referencing.

resolve
string
metadata
object
object

Metadata is the metadata for connection secret.

labels
object
type
string
name
requiredstring
object

WriteConnectionSecretToReference specifies the namespace and name of a Secret to which any connection details for this managed resource should be written. Connection details frequently include the endpoint, username, and password required to connect to the managed resource. This field is planned to be replaced in a future release in favor of PublishConnectionDetailsTo. Currently, both could be set independently and connection details would be published to both without affecting each other.

name
requiredstring
namespace
requiredstring
status
object
object

WorkflowTemplateStatus defines the observed state of WorkflowTemplate.

object

No description provided.

id
string
array

Required. WorkflowTemplate scheduling information.

array

A cluster that is managed by the workflow.

config
array
array

Required. The cluster configuration.

array

Optional. Port/endpoint configuration for this cluster

httpPorts
object
array

Optional. Lifecycle setting for the cluster.

array

Optional. The Compute Engine config settings for additional worker instances in a cluster.

array

Output only. The list of instance names. Dataproc derives the names from cluster_name, num_instances, and the instance group.

array

Output only. The config for Compute Engine Instance Group Manager that manages this group. This is only used for preemptible instance groups.

array

Optional. The Compute Engine config settings for additional worker instances in a cluster.

array

Output only. The list of instance names. Dataproc derives the names from cluster_name, num_instances, and the instance group.

array

Output only. The config for Compute Engine Instance Group Manager that manages this group. This is only used for preemptible instance groups.

array

Optional. The Compute Engine config settings for additional worker instances in a cluster.

array

Output only. The list of instance names. Dataproc derives the names from cluster_name, num_instances, and the instance group.

array

Output only. The config for Compute Engine Instance Group Manager that manages this group. This is only used for preemptible instance groups.

array

Conditions of the resource.

lastTransitionTime
requiredstring
message
string
reason
requiredstring
status
requiredstring
type
requiredstring
Marketplace

Discover the building blocks for your internal cloud platform.

© 2022 Upbound, Inc.

SolutionsProvidersConfigurations
LearnDocumentationTry for Free
MorePrivacy PolicyTerms & Conditions
Marketplace

© 2022 Upbound, Inc.

Marketplace

Discover the building blocksfor your internal cloud platform.