- Resource: Task
- TriggerSpec
- Type
- ExecutionStatus
- SparkTaskConfig
- InfrastructureSpec
- BatchComputeResources
- ContainerImageRuntime
- VpcNetwork
- NotebookTaskConfig
- Methods
Resource: Task
A task represents a user-visible job.
| JSON representation | 
|---|
| { "name" : string , "uid" : string , "createTime" : string , "updateTime" : string , "description" : string , "displayName" : string , "state" : enum ( | 
name 
 string 
Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{locationId}/lakes/{lakeId}/ tasks/{taskId}.
uid 
 string 
Output only. System generated globally unique ID for the task. This ID will be different if the task is deleted and re-created with the same name.
createTime 
 string (  
 Timestamp 
 
format)
Output only. The time when the task was created.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z" 
, "2014-10-02T15:01:23.045123456Z" 
or "2014-10-02T15:01:23+05:30" 
.
updateTime 
 string (  
 Timestamp 
 
format)
Output only. The time when the task was last updated.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z" 
, "2014-10-02T15:01:23.045123456Z" 
or "2014-10-02T15:01:23+05:30" 
.
description 
 string 
Optional. Description of the task.
displayName 
 string 
Optional. User friendly display name.
state 
 enum (  
 State 
 
)
Output only. Current state of the task.
labels 
 map (key: string, value: string) 
Optional. User-defined labels for the task.
An object containing a list of "key": value 
pairs. Example: { "name": "wrench", "mass": "1.3kg", "count": "3" } 
.
triggerSpec 
 object (  
 TriggerSpec 
 
)
Required. Spec related to how often and when a task should be triggered.
executionSpec 
 object (  
 ExecutionSpec 
 
)
Required. Spec related to how a task is executed.
executionStatus 
 object (  
 ExecutionStatus 
 
)
Output only. Status of the latest task executions.
config 
. Task template specific user-specified config. config 
can be only one of the following:spark 
 object (  
 SparkTaskConfig 
 
)
Config related to running custom Spark tasks.
notebook 
 object (  
 NotebookTaskConfig 
 
)
Config related to running scheduled Notebooks.
TriggerSpec
Task scheduling and trigger settings.
| JSON representation | 
|---|
| { "type" : enum ( | 
type 
 enum (  
 Type 
 
)
Required. Immutable. Trigger type of the user-specified Task.
startTime 
 string (  
 Timestamp 
 
format)
Optional. The first run of the task will be after this time. If not specified, the task will run shortly after being submitted if ON_DEMAND and based on the schedule if RECURRING.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z" 
, "2014-10-02T15:01:23.045123456Z" 
or "2014-10-02T15:01:23+05:30" 
.
disabled 
 boolean 
Optional. Prevent the task from executing. This does not cancel already running tasks. It is intended to temporarily disable RECURRING tasks.
maxRetries 
 integer 
Optional. Number of retry attempts before aborting. Set to zero to never attempt to retry a failed task.
trigger 
. Trigger only applies for RECURRING tasks. trigger 
can be only one of the following:schedule 
 string 
Optional. Cron schedule ( https://en.wikipedia.org/wiki/Cron 
) for running tasks periodically. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, CRON_TZ=America/New_York 1 * * * * 
, or TZ=America/New_York 1 * * * * 
. This field is required for RECURRING tasks.
Type
Determines how often and when the job will run.
| Enums | |
|---|---|
| TYPE_UNSPECIFIED | Unspecified trigger type. | 
| ON_DEMAND | The task runs one-time shortly after Task Creation. | 
| RECURRING | The task is scheduled to run periodically. | 
ExecutionStatus
Status of the task execution (e.g. Jobs).
| JSON representation | 
|---|
|  { 
 "updateTime" 
 : 
 string 
 , 
 "latestJob" 
 : 
 { 
 object (  | 
| Fields | |
|---|---|
| updateTime |   Output only. Last update time of the status. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples:  | 
| latestJob |   Output only. latest job execution | 
SparkTaskConfig
User-specified config for running a Spark task.
| JSON representation | 
|---|
| { "fileUris" : [ string ] , "archiveUris" : [ string ] , "infrastructureSpec" : { object ( | 
fileUris[] 
 string 
Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
archiveUris[] 
 string 
Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
infrastructureSpec 
 object (  
 InfrastructureSpec 
 
)
Optional. Infrastructure specification for the execution.
driver 
. Required. The specification of the main method to call to drive the job. Specify either the jar file that contains the main class or the main class name. driver 
can be only one of the following:mainJarFileUri 
 string 
The Cloud Storage URI of the jar file that contains the main class. The execution args are passed in as a sequence of named process arguments ( --key=value 
).
mainClass 
 string 
The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris 
. The execution args are passed in as a sequence of named process arguments ( --key=value 
).
pythonScriptFile 
 string 
The Gcloud Storage URI of the main Python file to use as the driver. Must be a .py file. The execution args are passed in as a sequence of named process arguments ( --key=value 
).
sqlScriptFile 
 string 
A reference to a query file. This should be the Cloud Storage URI of the query file. The execution args are used to declare a set of script variables ( set key="value"; 
).
sqlScript 
 string 
The query text. The execution args are used to declare a set of script variables ( set key="value"; 
).
InfrastructureSpec
Configuration for the underlying infrastructure used to run workloads.
| JSON representation | 
|---|
| { // Union field | 
resources 
. Hardware config. resources 
can be only one of the following:batch 
 object (  
 BatchComputeResources 
 
)
Compute resources needed for a Task when using Dataproc Serverless.
runtime 
. Software config. runtime 
can be only one of the following:containerImage 
 object (  
 ContainerImageRuntime 
 
)
Container Image Runtime Configuration.
network 
. Networking config. network 
can be only one of the following:vpcNetwork 
 object (  
 VpcNetwork 
 
)
Vpc network.
BatchComputeResources
Batch compute resources associated with the task.
| JSON representation | 
|---|
| { "executorsCount" : integer , "maxExecutorsCount" : integer } | 
| Fields | |
|---|---|
| executorsCount |   Optional. Total number of job executors. Executor Count should be between 2 and 100. [Default=2] | 
| maxExecutorsCount |   Optional. Max configurable executors. If maxExecutorsCount > executorsCount, then auto-scaling is enabled. Max Executor Count should be between 2 and 1000. [Default=1000] | 
ContainerImageRuntime
Container Image Runtime Configuration used with Batch execution.
| JSON representation | 
|---|
| { "image" : string , "javaJars" : [ string ] , "pythonPackages" : [ string ] , "properties" : { string : string , ... } } | 
| Fields | |
|---|---|
| image |   Optional. Container image to use. | 
| javaJars[] |   Optional. A list of Java JARS to add to the classpath. Valid input includes Cloud Storage URIs to Jar binaries. For example, gs://bucket-name/my/path/to/file.jar | 
| pythonPackages[] |   Optional. A list of python packages to be installed. Valid formats include Cloud Storage URI to a PIP installable library. For example, gs://bucket-name/my/path/to/lib.tar.gz | 
| properties |   Optional. Override to common configuration of open source components installed on the Dataproc cluster. The properties to set on daemon config files. Property keys are specified in  An object containing a list of  | 
VpcNetwork
Cloud VPC Network used to run the infrastructure.
| JSON representation | 
|---|
| { "networkTags" : [ string ] , // Union field | 
network_name 
. The Cloud VPC network identifier. network_name 
can be only one of the following:network 
 string 
Optional. The Cloud VPC network in which the job is run. By default, the Cloud VPC network named Default within the project is used.
subNetwork 
 string 
Optional. The Cloud VPC sub-network in which the job is run.
NotebookTaskConfig
Config for running scheduled notebooks.
| JSON representation | 
|---|
|  { 
 "notebook" 
 : 
 string 
 , 
 "infrastructureSpec" 
 : 
 { 
 object (  | 
| Fields | |
|---|---|
| notebook |   Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables (  | 
| infrastructureSpec |   Optional. Infrastructure specification for the execution. | 
| fileUris[] |   Optional. Cloud Storage URIs of files to be placed in the working directory of each executor. | 
| archiveUris[] |   Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip. | 
| Methods | |
|---|---|
|   | Creates a task resource within a lake. | 
|   | Delete the task resource. | 
|   | Get task resource. | 
|   | Gets the access control policy for a resource. | 
|   | Lists tasks under the given lake. | 
|   | Update the task resource. | 
|   | Run an on demand execution of a Task. | 
|   | Sets the access control policy on the specified resource. | 
|   | Returns permissions that a caller has on the specified resource. | 

