Tool: get_notebook_execution_schedule
Use this tool to get the details and status of a specific notebook execution schedule.
The following sample demonstrate how to use curl
to invoke the get_notebook_execution_schedule
MCP tool.
| Curl Request |
|---|
curl --location 'https://aiplatform.googleapis.com/mcp/generate' \ --header 'content-type: application/json' \ --header 'accept: application/json, text/event-stream' \ --data '{ "method": "tools/call", "params": { "name": "get_notebook_execution_schedule", "arguments": { // provide these details according to the tool' s MCP specification } } , "jsonrpc" : "2.0" , "id" : 1 } ' |
Input Schema
Request message for ScheduleService.GetSchedule
.
GetScheduleRequest
| JSON representation |
|---|
{ "name" : string } |
| Fields | |
|---|---|
name
|
Required. The name of the Schedule resource. Format: |
Output Schema
An instance of a Schedule periodically schedules runs to make API calls based on user specified time specification and API request type.
Schedule
| JSON representation |
|---|
{ "name" : string , "displayName" : string , "startTime" : string , "endTime" : string , "maxRunCount" : string , "startedRunCount" : string , "state" : enum ( |
name
string
Immutable. The resource name of the Schedule.
displayName
string
Required. User provided name of the Schedule. The name can be up to 128 characters long and can consist of any UTF-8 characters.
startTime
string (
Timestamp
format)
Optional. Timestamp after which the first run can be scheduled. Default to Schedule create time if not specified.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
endTime
string (
Timestamp
format)
Optional. Timestamp after which no new runs can be scheduled. If specified, The schedule will be completed when either end_time is reached or when scheduled_run_count >= max_run_count. If not specified, new runs will keep getting scheduled until this Schedule is paused or deleted. Already scheduled runs will be allowed to complete. Unset if not specified.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
maxRunCount
string ( int64
format)
Optional. Maximum run count of the schedule. If specified, The schedule will be completed when either started_run_count >= max_run_count or when end_time is reached. If not specified, new runs will keep getting scheduled until this Schedule is paused or deleted. Already scheduled runs will be allowed to complete. Unset if not specified.
startedRunCount
string ( int64
format)
Output only. The number of runs started by this schedule.
state
enum (
State
)
Output only. The state of this Schedule.
createTime
string (
Timestamp
format)
Output only. Timestamp when this Schedule was created.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
updateTime
string (
Timestamp
format)
Output only. Timestamp when this Schedule was updated.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
nextRunTime
string (
Timestamp
format)
Output only. Timestamp when this Schedule should schedule the next run. Having a next_run_time in the past means the runs are being started behind schedule.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
lastPauseTime
string (
Timestamp
format)
Output only. Timestamp when this Schedule was last paused. Unset if never paused.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
lastResumeTime
string (
Timestamp
format)
Output only. Timestamp when this Schedule was last resumed. Unset if never resumed from pause.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
maxConcurrentRunCount
string ( int64
format)
Required. Maximum number of runs that can be started concurrently for this Schedule. This is the limit for starting the scheduled requests and not the execution of the operations/jobs created by the requests (if applicable).
maxConcurrentActiveRunCount
string ( int64
format)
Optional. Specifies the maximum number of active runs that can be executed concurrently for this Schedule. This limits the number of runs that can be in a non-terminal state at the same time. Currently, this field is only supported for requests of type CreatePipelineJobRequest.
allowQueueing
boolean
Optional. Whether new scheduled runs can be queued when max_concurrent_runs limit is reached. If set to true, new runs will be queued instead of skipped. Default to false.
catchUp
boolean
Output only. Whether to backfill missed runs when the schedule is resumed from PAUSED state. If set to true, all missed runs will be scheduled. New runs will be scheduled after the backfill is complete. Default to false.
lastScheduledRunResponse
object (
RunResponse
)
Output only. Response of the last scheduled run. This is the response for starting the scheduled requests and not the execution of the operations/jobs created by the requests (if applicable). Unset if no run has been scheduled yet.
time_specification
. Required. The time specification to launch scheduled runs. time_specification
can be only one of the following:cron
string
Cron schedule ( https://en.wikipedia.org/wiki/Cron ) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *".
request
. Required. The API request template to launch the scheduled runs. User-specified ID is not supported in the request template. request
can be only one of the following:createPipelineJobRequest
object (
CreatePipelineJobRequest
)
Request for PipelineService.CreatePipelineJob
. CreatePipelineJobRequest.parent field is required (format: projects/{project}/locations/{location}).
createNotebookExecutionJobRequest
object (
CreateNotebookExecutionJobRequest
)
Request for NotebookService.CreateNotebookExecutionJob
.
CreatePipelineJobRequest
| JSON representation |
|---|
{
"parent"
:
string
,
"pipelineJob"
:
{
object (
|
| Fields | |
|---|---|
parent
|
Required. The resource name of the Location to create the PipelineJob in. Format: |
pipelineJob
|
Required. The PipelineJob to create. |
pipelineJobId
|
The ID to use for the PipelineJob, which will become the final component of the PipelineJob name. If not provided, an ID will be automatically generated. This value should be less than 128 characters, and valid characters are |
PipelineJob
| JSON representation |
|---|
{ "name" : string , "displayName" : string , "createTime" : string , "startTime" : string , "endTime" : string , "updateTime" : string , "pipelineSpec" : { object } , "state" : enum ( |
| Fields | |
|---|---|
name
|
Output only. The resource name of the PipelineJob. |
displayName
|
The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters. |
createTime
|
Output only. Pipeline creation time. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
startTime
|
Output only. Pipeline start time. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
endTime
|
Output only. Pipeline end time. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
updateTime
|
Output only. Timestamp when this PipelineJob was most recently updated. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
pipelineSpec
|
A compiled definition of a pipeline, represented as a |
state
|
Output only. The detailed state of the job. |
jobDetail
|
Output only. The details of pipeline run. Not available in the list view. |
error
|
Output only. The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED. |
labels
|
The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. - An object containing a list of |
runtimeConfig
|
Runtime config of the pipeline. |
encryptionSpec
|
Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key. |
serviceAccount
|
The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the |
network
|
The full name of the Compute Engine network
to which the Pipeline Job's workload should be peered. For example, Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network. |
reservedIpRanges[]
|
A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range']. |
pscInterfaceConfig
|
Optional. Configuration for PSC-I for PipelineJob. |
templateUri
|
A template uri from where the |
templateMetadata
|
Output only. Pipeline template metadata. Will fill up fields if |
scheduleName
|
Output only. The schedule resource name. Only returned if the Pipeline is created by Schedule API. |
preflightValidations
|
Optional. Whether to do component level validations before job creation. |
Timestamp
| JSON representation |
|---|
{ "seconds" : string , "nanos" : integer } |
| Fields | |
|---|---|
seconds
|
Represents seconds of UTC time since Unix epoch 1970-01-01T00:00:00Z. Must be between -62135596800 and 253402300799 inclusive (which corresponds to 0001-01-01T00:00:00Z to 9999-12-31T23:59:59Z). |
nanos
|
Non-negative fractions of a second at nanosecond resolution. This field is the nanosecond portion of the duration, not an alternative to seconds. Negative second values with fractions must still have non-negative nanos values that count forward in time. Must be between 0 and 999,999,999 inclusive. |
Struct
| JSON representation |
|---|
{ "fields" : { string : value , ... } } |
| Fields | |
|---|---|
fields
|
Unordered map of dynamically typed values. An object containing a list of |
FieldsEntry
| JSON representation |
|---|
{ "key" : string , "value" : value } |
| Fields | |
|---|---|
key
|
|
value
|
|
Value
| JSON representation |
|---|
{ // Union field |
kind
. The kind of value. kind
can be only one of the following:nullValue
null
Represents a JSON null
.
numberValue
number
Represents a JSON number. Must not be NaN
, Infinity
or -Infinity
, since those are not supported in JSON. This also cannot represent large Int64 values, since JSON format generally does not support them in its number type.
stringValue
string
Represents a JSON string.
boolValue
boolean
Represents a JSON boolean ( true
or false
literal in JSON).
structValue
object (
Struct
format)
Represents a JSON object.
listValue
array (
ListValue
format)
Represents a JSON array.
ListValue
| JSON representation |
|---|
{ "values" : [ value ] } |
| Fields | |
|---|---|
values[]
|
Repeated field of dynamically typed values. |
PipelineJobDetail
| JSON representation |
|---|
{ "pipelineContext" : { object ( |
| Fields | |
|---|---|
pipelineContext
|
Output only. The context of the pipeline. |
pipelineRunContext
|
Output only. The context of the current pipeline run. |
taskDetails[]
|
Output only. The runtime details of the tasks under the pipeline. |
Context
| JSON representation |
|---|
{ "name" : string , "displayName" : string , "etag" : string , "labels" : { string : string , ... } , "createTime" : string , "updateTime" : string , "parentContexts" : [ string ] , "schemaTitle" : string , "schemaVersion" : string , "metadata" : { object } , "description" : string } |
| Fields | |
|---|---|
name
|
Immutable. The resource name of the Context. |
displayName
|
User provided display name of the Context. May be up to 128 Unicode characters. |
etag
|
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels
|
The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded). An object containing a list of |
createTime
|
Output only. Timestamp when this Context was created. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
updateTime
|
Output only. Timestamp when this Context was last updated. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
parentContexts[]
|
Output only. A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts. |
schemaTitle
|
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
schemaVersion
|
The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
metadata
|
Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB. |
description
|
Description of the Context |
LabelsEntry
| JSON representation |
|---|
{ "key" : string , "value" : string } |
| Fields | |
|---|---|
key
|
|
value
|
|
PipelineTaskDetail
| JSON representation |
|---|
{ "taskId" : string , "parentTaskId" : string , "taskName" : string , "createTime" : string , "startTime" : string , "endTime" : string , "executorDetail" : { object ( |
| Fields | |
|---|---|
taskId
|
Output only. The system generated ID of the task. |
parentTaskId
|
Output only. The id of the parent task if the task is within a component scope. Empty if the task is at the root level. |
taskName
|
Output only. The user specified name of the task that is defined in |
createTime
|
Output only. Task create time. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
startTime
|
Output only. Task start time. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
endTime
|
Output only. Task end time. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
executorDetail
|
Output only. The detailed execution info. |
state
|
Output only. State of the task. |
execution
|
Output only. The execution metadata of the task. |
error
|
Output only. The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED. |
pipelineTaskStatus[]
|
Output only. A list of task status. This field keeps a record of task status evolving over time. |
inputs
|
Output only. The runtime input artifacts of the task. An object containing a list of |
outputs
|
Output only. The runtime output artifacts of the task. An object containing a list of |
taskUniqueName
|
Output only. The unique name of a task. This field is used by rerun pipeline job. Console UI and Vertex AI SDK will support triggering pipeline job reruns. The name is constructed by concatenating all the parent tasks name with the task name. For example, if a task named "child_task" has a parent task named "parent_task_1" and parent task 1 has a parent task named "parent_task_2", the task unique name will be "parent_task_2.parent_task_1.child_task". |
PipelineTaskExecutorDetail
| JSON representation |
|---|
{ // Union field |
Union field details
.
details
can be only one of the following:
containerDetail
object (
ContainerDetail
)
Output only. The detailed info for a container executor.
customJobDetail
object (
CustomJobDetail
)
Output only. The detailed info for a custom job executor.
ContainerDetail
| JSON representation |
|---|
{ "mainJob" : string , "preCachingCheckJob" : string , "failedMainJobs" : [ string ] , "failedPreCachingCheckJobs" : [ string ] } |
| Fields | |
|---|---|
mainJob
|
Output only. The name of the |
preCachingCheckJob
|
Output only. The name of the |
failedMainJobs[]
|
Output only. The names of the previously failed |
failedPreCachingCheckJobs[]
|
Output only. The names of the previously failed |
CustomJobDetail
| JSON representation |
|---|
{ "job" : string , "failedJobs" : [ string ] } |
| Fields | |
|---|---|
job
|
Output only. The name of the |
failedJobs[]
|
Output only. The names of the previously failed |
Execution
| JSON representation |
|---|
{
"name"
:
string
,
"displayName"
:
string
,
"state"
:
enum (
|
| Fields | |
|---|---|
name
|
Output only. The resource name of the Execution. |
displayName
|
User provided display name of the Execution. May be up to 128 Unicode characters. |
state
|
The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions. |
etag
|
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels
|
The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded). An object containing a list of |
createTime
|
Output only. Timestamp when this Execution was created. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
updateTime
|
Output only. Timestamp when this Execution was last updated. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
schemaTitle
|
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
schemaVersion
|
The version of the schema in Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
metadata
|
Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB. |
description
|
Description of the Execution |
LabelsEntry
| JSON representation |
|---|
{ "key" : string , "value" : string } |
| Fields | |
|---|---|
key
|
|
value
|
|
Status
| JSON representation |
|---|
{ "code" : integer , "message" : string , "details" : [ { "@type" : string , field1 : ... , ... } ] } |
| Fields | |
|---|---|
code
|
The status code, which should be an enum value of |
message
|
A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the |
details[]
|
A list of messages that carry the error details. There is a common set of message types for APIs to use. An object containing fields of an arbitrary type. An additional field |
Any
| JSON representation |
|---|
{ "typeUrl" : string , "value" : string } |
| Fields | |
|---|---|
typeUrl
|
Identifies the type of the serialized Protobuf message with a URI reference consisting of a prefix ending in a slash and the fully-qualified type name. Example: type.googleapis.com/google.protobuf.StringValue This string must contain at least one The prefix is arbitrary and Protobuf implementations are expected to simply strip off everything up to and including the last All type URL strings must be legal URI references with the additional restriction (for the text format) that the content of the reference must consist only of alphanumeric characters, percent-encoded escapes, and characters in the following set (not including the outer backticks): In the original design of |
value
|
Holds a Protobuf serialization of the type described by type_url. A base64-encoded string. |
PipelineTaskStatus
| JSON representation |
|---|
{ "updateTime" : string , "state" : enum ( |
| Fields | |
|---|---|
updateTime
|
Output only. Update time of this status. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
state
|
Output only. The state of the task. |
error
|
Output only. The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried. |
InputsEntry
| JSON representation |
|---|
{
"key"
:
string
,
"value"
:
{
object (
|
| Fields | |
|---|---|
key
|
|
value
|
|
ArtifactList
| JSON representation |
|---|
{
"artifacts"
:
[
{
object (
|
| Fields | |
|---|---|
artifacts[]
|
Output only. A list of artifact metadata. |
Artifact
| JSON representation |
|---|
{
"name"
:
string
,
"displayName"
:
string
,
"uri"
:
string
,
"etag"
:
string
,
"labels"
:
{
string
:
string
,
...
}
,
"createTime"
:
string
,
"updateTime"
:
string
,
"state"
:
enum (
|
| Fields | |
|---|---|
name
|
Output only. The resource name of the Artifact. |
displayName
|
User provided display name of the Artifact. May be up to 128 Unicode characters. |
uri
|
The uniform resource identifier of the artifact file. May be empty if there is no actual artifact file. |
etag
|
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. |
labels
|
The labels with user-defined metadata to organize your Artifacts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Artifact (System labels are excluded). An object containing a list of |
createTime
|
Output only. Timestamp when this Artifact was created. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
updateTime
|
Output only. Timestamp when this Artifact was last updated. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
state
|
The state of this Artifact. This is a property of the Artifact, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines), and the system does not prescribe or check the validity of state transitions. |
schemaTitle
|
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
schemaVersion
|
The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store. |
metadata
|
Properties of the Artifact. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB. |
description
|
Description of the Artifact |
LabelsEntry
| JSON representation |
|---|
{ "key" : string , "value" : string } |
| Fields | |
|---|---|
key
|
|
value
|
|
OutputsEntry
| JSON representation |
|---|
{
"key"
:
string
,
"value"
:
{
object (
|
| Fields | |
|---|---|
key
|
|
value
|
|
LabelsEntry
| JSON representation |
|---|
{ "key" : string , "value" : string } |
| Fields | |
|---|---|
key
|
|
value
|
|
RuntimeConfig
| JSON representation |
|---|
{ "parameters" : { string : { object ( |
| Fields | |
|---|---|
parameters
|
Deprecated. Use An object containing a list of |
gcsOutputDirectory
|
Required. A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern |
parameterValues
|
The runtime parameters of the PipelineJob. The parameters will be passed into An object containing a list of |
failurePolicy
|
Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion. |
inputArtifacts
|
The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact. An object containing a list of |
ParametersEntry
| JSON representation |
|---|
{
"key"
:
string
,
"value"
:
{
object (
|
| Fields | |
|---|---|
key
|
|
value
|
|
Value
| JSON representation |
|---|
{ // Union field |
Union field value
.
value
can be only one of the following:
intValue
string ( int64
format)
An integer value.
doubleValue
number
A double value.
stringValue
string
A string value.
ParameterValuesEntry
| JSON representation |
|---|
{ "key" : string , "value" : value } |
| Fields | |
|---|---|
key
|
|
value
|
|
InputArtifactsEntry
| JSON representation |
|---|
{
"key"
:
string
,
"value"
:
{
object (
|
| Fields | |
|---|---|
key
|
|
value
|
|
InputArtifact
| JSON representation |
|---|
{ // Union field |
Union field kind
.
kind
can be only one of the following:
artifactId
string
Artifact resource id from MLMD. Which is the last portion of an artifact resource name: projects/{project}/locations/{location}/metadataStores/default/artifacts/{artifact_id}
. The artifact must stay within the same project, location and default metadatastore as the pipeline.
EncryptionSpec
| JSON representation |
|---|
{ "kmsKeyName" : string } |
| Fields | |
|---|---|
kmsKeyName
|
Required. Resource name of the Cloud KMS key used to protect the resource. The Cloud KMS key must be in the same region as the resource. It must have the format |
PscInterfaceConfig
| JSON representation |
|---|
{
"networkAttachment"
:
string
,
"dnsPeeringConfigs"
:
[
{
object (
|
| Fields | |
|---|---|
networkAttachment
|
Optional. The name of the Compute Engine network attachment to attach to the resource within the region and user project. To specify this field, you must have already created a network attachment . This field is only used for resources using PSC-I. |
dnsPeeringConfigs[]
|
Optional. DNS peering configurations. When specified, Vertex AI will attempt to configure DNS peering zones in the tenant project VPC to resolve the specified domains using the target network's Cloud DNS. The user must grant the dns.peer role to the Vertex AI Service Agent on the target project. |
DnsPeeringConfig
| JSON representation |
|---|
{ "domain" : string , "targetProject" : string , "targetNetwork" : string } |
| Fields | |
|---|---|
domain
|
Required. The DNS name suffix of the zone being peered to, e.g., "my-internal-domain.corp.". Must end with a dot. |
targetProject
|
Required. The project ID hosting the Cloud DNS managed zone that contains the 'domain'. The Vertex AI Service Agent requires the dns.peer role on this project. |
targetNetwork
|
Required. The VPC network name in the target_project where the DNS zone specified by 'domain' is visible. |
PipelineTemplateMetadata
| JSON representation |
|---|
{ "version" : string } |
| Fields | |
|---|---|
version
|
The version_name in artifact registry. Will always be presented in output if the Format is "sha256:abcdef123456...". |
CreateNotebookExecutionJobRequest
| JSON representation |
|---|
{
"parent"
:
string
,
"notebookExecutionJob"
:
{
object (
|
| Fields | |
|---|---|
parent
|
Required. The resource name of the Location to create the NotebookExecutionJob. Format: |
notebookExecutionJob
|
Required. The NotebookExecutionJob to create. |
notebookExecutionJobId
|
Optional. User specified ID for the NotebookExecutionJob. |
NotebookExecutionJob
| JSON representation |
|---|
{ "name" : string , "displayName" : string , "executionTimeout" : string , "scheduleResourceName" : string , "jobState" : enum ( |
name
string
Output only. The resource name of this NotebookExecutionJob. Format: projects/{project_id}/locations/{location}/notebookExecutionJobs/{job_id}
displayName
string
The display name of the NotebookExecutionJob. The name can be up to 128 characters long and can consist of any UTF-8 characters.
executionTimeout
string (
Duration
format)
Max running time of the execution job in seconds (default 86400s / 24 hrs).
A duration in seconds with up to nine fractional digits, ending with ' s
'. Example: "3.5s"
.
scheduleResourceName
string
The Schedule resource name if this job is triggered by one. Format: projects/{project_id}/locations/{location}/schedules/{schedule_id}
jobState
enum (
JobState
)
Output only. The state of the NotebookExecutionJob.
status
object (
Status
)
Output only. Populated when the NotebookExecutionJob is completed. When there is an error during notebook execution, the error details are populated.
createTime
string (
Timestamp
format)
Output only. Timestamp when this NotebookExecutionJob was created.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
updateTime
string (
Timestamp
format)
Output only. Timestamp when this NotebookExecutionJob was most recently updated.
Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: "2014-10-02T15:01:23Z"
, "2014-10-02T15:01:23.045123456Z"
or "2014-10-02T15:01:23+05:30"
.
labels
map (key: string, value: string)
The labels with user-defined metadata to organize NotebookExecutionJobs.
Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed.
See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
An object containing a list of "key": value
pairs. Example: { "name": "wrench", "mass": "1.3kg", "count": "3" }
.
kernelName
string
The name of the kernel to use during notebook execution. If unset, the default kernel is used.
encryptionSpec
object (
EncryptionSpec
)
Customer-managed encryption key spec for the notebook execution job. This field is auto-populated if the NotebookRuntimeTemplate
has an encryption spec.
notebook_source
. The input notebook. notebook_source
can be only one of the following:dataformRepositorySource
object (
DataformRepositorySource
)
The Dataform Repository pointing to a single file notebook repository.
gcsNotebookSource
object (
GcsNotebookSource
)
The Cloud Storage url pointing to the ipynb file. Format: gs://bucket/notebook_file.ipynb
directNotebookSource
object (
DirectNotebookSource
)
The contents of an input notebook file.
environment_spec
. The compute config to use for an execution job. environment_spec
can be only one of the following:notebookRuntimeTemplateResourceName
string
The NotebookRuntimeTemplate to source compute configuration from.
customEnvironmentSpec
object (
CustomEnvironmentSpec
)
The custom compute configuration for an execution job.
execution_sink
. The location to store the notebook execution result. execution_sink
can be only one of the following:gcsOutputUri
string
The Cloud Storage location to upload the result to. Format: gs://bucket-name
execution_identity
. The identity to run the execution as. execution_identity
can be only one of the following:executionUser
string
The user email to run the execution as. Only supported by Colab runtimes.
serviceAccount
string
The service account to run the execution as.
runtime_environment
. Runtime environment for the notebook execution job. If unspecified, the default runtime of Colab is used. runtime_environment
can be only one of the following:workbenchRuntime
object (
WorkbenchRuntime
)
The Workbench runtime configuration to use for the notebook execution.
DataformRepositorySource
| JSON representation |
|---|
{ "dataformRepositoryResourceName" : string , "commitSha" : string } |
| Fields | |
|---|---|
dataformRepositoryResourceName
|
The resource name of the Dataform Repository. Format: |
commitSha
|
The commit SHA to read repository with. If unset, the file will be read at HEAD. |
GcsNotebookSource
| JSON representation |
|---|
{ "uri" : string , "generation" : string } |
| Fields | |
|---|---|
uri
|
The Cloud Storage uri pointing to the ipynb file. Format: |
generation
|
The version of the Cloud Storage object to read. If unset, the current version of the object is read. See https://cloud.google.com/storage/docs/metadata#generation-number . |
DirectNotebookSource
| JSON representation |
|---|
{ "content" : string } |
| Fields | |
|---|---|
content
|
The base64-encoded contents of the input notebook file. A base64-encoded string. |
CustomEnvironmentSpec
| JSON representation |
|---|
{ "machineSpec" : { object ( |
| Fields | |
|---|---|
machineSpec
|
The specification of a single machine for the execution job. |
persistentDiskSpec
|
The specification of a persistent disk to attach for the execution job. |
networkSpec
|
The network configuration to use for the execution job. |
MachineSpec
| JSON representation |
|---|
{ "machineType" : string , "acceleratorType" : enum ( |
| Fields | |
|---|---|
machineType
|
Immutable. The type of the machine. See the list of machine types supported for prediction See the list of machine types supported for custom training . For |
acceleratorType
|
Immutable. The type of accelerator(s) that may be attached to the machine as per |
acceleratorCount
|
The number of accelerators to attach to the machine. For accelerator optimized machine types ( https://cloud.google.com/compute/docs/accelerator-optimized-machines) , One may set the accelerator_count from 1 to N for machine with N GPUs. If accelerator_count is less than or equal to N / 2, Vertex will co-schedule the replicas of the model into the same VM to save cost. For example, if the machine type is a3-highgpu-8g, which has 8 H100 GPUs, one can set accelerator_count to 1 to 8. If accelerator_count is 1, 2, 3, or 4, Vertex will co-schedule 8, 4, 2, or 2 replicas of the model into the same VM to save cost. When co-scheduling, CPU, memory and storage on the VM will be distributed to replicas on the VM. For example, one can expect a co-scheduled replica requesting 2 GPUs out of a 8-GPU VM will receive 25% of the CPU, memory and storage of the VM. Note that the feature is not compatible with [multihost_gpu_node_count][]. When multihost_gpu_node_count is set, the co-scheduling will not be enabled. |
gpuPartitionSize
|
Optional. Immutable. The Nvidia GPU partition size. When specified, the requested accelerators will be partitioned into smaller GPU partitions. For example, if the request is for 8 units of NVIDIA A100 GPUs, and gpu_partition_size="1g.10gb", the service will create 8 * 7 = 56 partitioned MIG instances. The partition size must be a value supported by the requested accelerator. Refer to Nvidia GPU Partitioning for the available partition sizes. If set, the accelerator_count should be set to 1. |
tpuTopology
|
Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). |
reservationAffinity
|
Optional. Immutable. Configuration controlling how this resource pool consumes reservation. |
ReservationAffinity
| JSON representation |
|---|
{
"reservationAffinityType"
:
enum (
|
| Fields | |
|---|---|
reservationAffinityType
|
Required. Specifies the reservation affinity type. |
key
|
Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use |
values[]
|
Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation or reservation block. |
PersistentDiskSpec
| JSON representation |
|---|
{ "diskType" : string , "diskSizeGb" : string } |
| Fields | |
|---|---|
diskType
|
Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) |
diskSizeGb
|
Size in GB of the disk (default is 100GB). |
NetworkSpec
| JSON representation |
|---|
{ "enableInternetAccess" : boolean , "network" : string , "subnetwork" : string } |
| Fields | |
|---|---|
enableInternetAccess
|
Whether to enable public internet access. Default false. |
network
|
The full name of the Google Compute Engine network |
subnetwork
|
The name of the subnet that this instance is in. Format: |
Duration
| JSON representation |
|---|
{ "seconds" : string , "nanos" : integer } |
| Fields | |
|---|---|
seconds
|
Signed seconds of the span of time. Must be from -315,576,000,000 to +315,576,000,000 inclusive. Note: these bounds are computed from: 60 sec/min * 60 min/hr * 24 hr/day * 365.25 days/year * 10000 years |
nanos
|
Signed fractions of a second at nanosecond resolution of the span of time. Durations less than one second are represented with a 0 |
LabelsEntry
| JSON representation |
|---|
{ "key" : string , "value" : string } |
| Fields | |
|---|---|
key
|
|
value
|
|
RunResponse
| JSON representation |
|---|
{ "scheduledRunTime" : string , "runResponse" : string } |
| Fields | |
|---|---|
scheduledRunTime
|
The scheduled run time based on the user-specified schedule. Uses RFC 3339, where generated output will always be Z-normalized and use 0, 3, 6 or 9 fractional digits. Offsets other than "Z" are also accepted. Examples: |
runResponse
|
The response of the scheduled run. |
Tool Annotations
Destructive Hint: ❌ | Idempotent Hint: ❌ | Read Only Hint: ✅ | Open World Hint: ❌

