- Resource: Task
- TriggerSpec
- Type
- ExecutionStatus
- SparkTaskConfig
- InfrastructureSpec
- BatchComputeResources
- ContainerImageRuntime
- VpcNetwork
- NotebookTaskConfig
- Methods
Resource: Task
A task represents a user-visible job.
JSON representation |
---|
{ "name" : string , "uid" : string , "createTime" : string , "updateTime" : string , "description" : string , "displayName" : string , "state" : enum ( |
name
string
Output only. The relative resource name of the task, of the form: projects/{project_number}/locations/{locationId}/lakes/{lakeId}/ tasks/{taskId}.
uid
string
Output only. System generated globally unique ID for the task. This ID will be different if the task is deleted and re-created with the same name.
createTime
string (
Timestamp
format)
Output only. The time when the task was created.
Uses RFC 3339, where generated output will always be Z-normalized and uses 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
updateTime
string (
Timestamp
format)
Output only. The time when the task was last updated.
Uses RFC 3339, where generated output will always be Z-normalized and uses 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
description
string
Optional. Description of the task.
displayName
string
Optional. User friendly display name.
state
enum (
State
)
Output only. Current state of the task.
labels
map (key: string, value: string)
Optional. User-defined labels for the task.
An object containing a list of "key": value
pairs. Example: { "name": "wrench", "mass": "1.3kg", "count": "3" }
.
triggerSpec
object (
TriggerSpec
)
Required. Spec related to how often and when a task should be triggered.
executionSpec
object (
ExecutionSpec
)
Required. Spec related to how a task is executed.
executionStatus
object (
ExecutionStatus
)
Output only. Status of the latest task executions.
config
. Task template specific user-specified config. config
can be only one of the following:spark
object (
SparkTaskConfig
)
Config related to running custom Spark tasks.
notebook
object (
NotebookTaskConfig
)
Config related to running scheduled Notebooks.
TriggerSpec
Task scheduling and trigger settings.
JSON representation |
---|
{ "type" : enum ( |
type
enum (
Type
)
Required. Immutable. Trigger type of the user-specified Task.
startTime
string (
Timestamp
format)
Optional. The first run of the task will be after this time. If not specified, the task will run shortly after being submitted if ON_DEMAND and based on the schedule if RECURRING.
Uses RFC 3339, where generated output will always be Z-normalized and uses 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
disabled
boolean
Optional. Prevent the task from executing. This does not cancel already running tasks. It is intended to temporarily disable RECURRING tasks.
maxRetries
integer
Optional. Number of retry attempts before aborting. Set to zero to never attempt to retry a failed task.
trigger
. Trigger only applies for RECURRING tasks. trigger
can be only one of the following:schedule
string
Optional. Cron schedule ( https://en.wikipedia.org/wiki/Cron
) for running tasks periodically. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, CRON_TZ=America/New_York 1 * * * *
, or TZ=America/New_York 1 * * * *
. This field is required for RECURRING tasks.
Type
Determines how often and when the job will run.
Enums | |
---|---|
TYPE_UNSPECIFIED
|
Unspecified trigger type. |
ON_DEMAND
|
The task runs one-time shortly after Task Creation. |
RECURRING
|
The task is scheduled to run periodically. |
ExecutionStatus
Status of the task execution (e.g. Jobs).
JSON representation |
---|
{
"updateTime"
:
string
,
"latestJob"
:
{
object (
|
Fields | |
---|---|
updateTime
|
Output only. Last update time of the status. Uses RFC 3339, where generated output will always be Z-normalized and uses 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
latestJob
|
Output only. latest job execution |
SparkTaskConfig
User-specified config for running a Spark task.
JSON representation |
---|
{ "fileUris" : [ string ] , "archiveUris" : [ string ] , "infrastructureSpec" : { object ( |
fileUris[]
string
Optional. Cloud Storage URIs of files to be placed in the working directory of each executor.
archiveUris[]
string
Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
infrastructureSpec
object (
InfrastructureSpec
)
Optional. Infrastructure specification for the execution.
driver
. Required. The specification of the main method to call to drive the job. Specify either the jar file that contains the main class or the main class name. driver
can be only one of the following:mainJarFileUri
string
The Cloud Storage URI of the jar file that contains the main class. The execution args are passed in as a sequence of named process arguments ( --key=value
).
mainClass
string
The name of the driver's main class. The jar file that contains the class must be in the default CLASSPATH or specified in jar_file_uris
. The execution args are passed in as a sequence of named process arguments ( --key=value
).
pythonScriptFile
string
The Gcloud Storage URI of the main Python file to use as the driver. Must be a .py file. The execution args are passed in as a sequence of named process arguments ( --key=value
).
sqlScriptFile
string
A reference to a query file. This should be the Cloud Storage URI of the query file. The execution args are used to declare a set of script variables ( set key="value";
).
sqlScript
string
The query text. The execution args are used to declare a set of script variables ( set key="value";
).
InfrastructureSpec
Configuration for the underlying infrastructure used to run workloads.
JSON representation |
---|
{ // Union field |
resources
. Hardware config. resources
can be only one of the following:batch
object (
BatchComputeResources
)
Compute resources needed for a Task when using Dataproc Serverless.
runtime
. Software config. runtime
can be only one of the following:containerImage
object (
ContainerImageRuntime
)
Container Image Runtime Configuration.
network
. Networking config. network
can be only one of the following:vpcNetwork
object (
VpcNetwork
)
Vpc network.
BatchComputeResources
Batch compute resources associated with the task.
JSON representation |
---|
{ "executorsCount" : integer , "maxExecutorsCount" : integer } |
Fields | |
---|---|
executorsCount
|
Optional. Total number of job executors. Executor Count should be between 2 and 100. [Default=2] |
maxExecutorsCount
|
Optional. Max configurable executors. If maxExecutorsCount > executorsCount, then auto-scaling is enabled. Max Executor Count should be between 2 and 1000. [Default=1000] |
ContainerImageRuntime
Container Image Runtime Configuration used with Batch execution.
JSON representation |
---|
{ "image" : string , "javaJars" : [ string ] , "pythonPackages" : [ string ] , "properties" : { string : string , ... } } |
Fields | |
---|---|
image
|
Optional. Container image to use. |
javaJars[]
|
Optional. A list of Java JARS to add to the classpath. Valid input includes Cloud Storage URIs to Jar binaries. For example, gs://bucket-name/my/path/to/file.jar |
pythonPackages[]
|
Optional. A list of python packages to be installed. Valid formats include Cloud Storage URI to a PIP installable library. For example, gs://bucket-name/my/path/to/lib.tar.gz |
properties
|
Optional. Override to common configuration of open source components installed on the Dataproc cluster. The properties to set on daemon config files. Property keys are specified in An object containing a list of |
VpcNetwork
Cloud VPC Network used to run the infrastructure.
JSON representation |
---|
{ "networkTags" : [ string ] , // Union field |
network_name
. The Cloud VPC network identifier. network_name
can be only one of the following:network
string
Optional. The Cloud VPC network in which the job is run. By default, the Cloud VPC network named Default within the project is used.
subNetwork
string
Optional. The Cloud VPC sub-network in which the job is run.
NotebookTaskConfig
Config for running scheduled notebooks.
JSON representation |
---|
{
"notebook"
:
string
,
"infrastructureSpec"
:
{
object (
|
Fields | |
---|---|
notebook
|
Required. Path to input notebook. This can be the Cloud Storage URI of the notebook file or the path to a Notebook Content. The execution args are accessible as environment variables ( |
infrastructureSpec
|
Optional. Infrastructure specification for the execution. |
fileUris[]
|
Optional. Cloud Storage URIs of files to be placed in the working directory of each executor. |
archiveUris[]
|
Optional. Cloud Storage URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip. |
Methods |
|
---|---|
|
Creates a task resource within a lake. |
|
Delete the task resource. |
|
Get task resource. |
|
Gets the access control policy for a resource. |
|
Lists tasks under the given lake. |
|
Update the task resource. |
|
Run an on demand execution of a Task. |
|
Sets the access control policy on the specified resource. |
|
Returns permissions that a caller has on the specified resource. |