This document explains how to access, query, and interpret Dataplex Universal Catalog logs using Cloud Logging. Accessing Dataplex Universal Catalog job and service logs helps you troubleshoot issues and monitor data management activities, including AI-powered data discovery and data quality scanning. By centralizing logs in Logging, you can analyze job performance, set up alerts for failures or anomalies, and route logs to other Google Cloud services like BigQuery for long-term retention and analysis.
To understand the costs, see Google Cloud Observability pricing .
For more information about logging retention, see Logs retention periods .
To disable all logs or exclude logs from Logging, see Exclusion filters .
To route logs from Logging to Cloud Storage, BigQuery, or Pub/Sub, see Routing and storage overview .
Use cases
Dataplex Universal Catalog logging supports use cases across different industries:
- Troubleshoot data pipeline failures:When a Dataplex Universal Catalog task
for data processing fails,
processlogs provide detailed error messages that help data engineers identify and resolve issues in their Spark jobs or custom tasks. - Monitor data quality:A financial services company can monitor
data_quality_scan_rule_resultlogs to track data quality trends over time, get alerts on quality degradations for critical data assets, and provide auditors with evidence of data quality checks for regulatory compliance. - Track metadata enrichment:A retail company using metadata import jobs
to enrich their catalog can use
metadata_joblogs to verify that imports are completing successfully and that all metadata items are being processed correctly. - Audit data discovery:Organizations can use
discoverylogs to monitor how and when new data sources are being discovered and registered within Dataplex Universal Catalog, providing an audit trail for data onboarding processes.
How Dataplex Universal Catalog logging works
Dataplex Universal Catalog sends logs for service operations and job executions to
Cloud Logging. Each log entry contains details about the operation or job,
such as its status, start and end time, associated resources (like a data scan
or task), and outcome. Different types of operations, such as data scanning,
discovery, metadata import, and data processing, generate different log types,
which you can query using logName
in Logging.
You can access and analyze these logs using the Logs Explorer in the Google Cloud console or by routing them to other destinations like Cloud Storage buckets or BigQuery tables for further analysis.
Access Dataplex Universal Catalog service logs in Logging
Dataplex Universal Catalog publishes the following service logs to Cloud Logging.
| Log type | Log name | logName
query |
Log description |
|---|---|---|---|
|
Data scan event logs
|
dataplex.googleapis.com/data_scan
|
logName=(projects/ $PROJECT_ID
/logs/dataplex.googleapis.com%2Fdata_scan)
|
Event logs for data scan jobs indicating job state, results, and statistics |
|
Data quality scan rule result logs
|
dataplex.googleapis.com/data_quality_scan_rule_result
|
logName=(projects/ $PROJECT_ID
/logs/dataplex.googleapis.com%2Fdata_quality_scan_rule_result)
|
Results of data quality scan rules in a data quality job |
|
Discovery logs
|
dataplex.googleapis.com/discovery
|
logName=(projects/ $PROJECT_ID
/logs/dataplex.googleapis.com%2Fdiscovery)
|
Discovery progress and updates over assets in a zone |
|
Metadata job logs
|
dataplex.googleapis.com/metadata_job
|
logName=(projects/ $PROJECT_ID
/logs/dataplex.googleapis.com%2Fmetadata_job)
|
Logs about metadata import jobs and import items in the metadata import file |
|
Process logs
|
dataplex.googleapis.com/process
|
logName=(projects/ $PROJECT_ID
/logs/dataplex.googleapis.com%2Fprocess)
|
Job runs resulting from data processing tasks |
Replace the following:
- PROJECT_ID : the ID of your project
To access Logging, you can use the Logs Explorer
in the
Google Cloud console, the gcloud logging
commands
, or
the Logging API
.
Query data scan event logs
When you use Dataplex Universal Catalog to create and run a data scan , a data scan event log is produced in Logging for the resulting job.
Console
-
In the Google Cloud console, go to the Logs explorerpage.
-
In the Logs Explorerview, find the Querytab.
-
Click the Resourcemenu.
-
Select Cloud Dataplex DataScan. Click Apply.
-
Click the Log namemenu.
-
In the Search log namesfield, enter
dataplex.googleapis.com%2Fdata_scan. Select data_scanand click Apply. -
Optional: Filter the logs to a specific data scan ID or location by adding the following filters in the log query:
resource.labels.location=" LOCATION " resource.labels.datascan_id=" DATA_SCAN_ID "
-
Click Run query.
gcloud
To read your data scan event log entries, use the gcloud logging read
command
with the following query:
gcloud logging read \ 'resource.type="dataplex.googleapis.com/DataScan" AND logName=projects/ PROJECT_ID /logs/dataplex.googleapis.com%2Fdata_scan AND resource.labels.location= LOCATION AND resource.labels.datascan_id= DATA_SCAN_ID ' --limit 10
REST
To list log entries, use the entries.list
method
.
Query data quality scan rule result logs
When you use Dataplex Universal Catalog to create and run a data quality scan , a data quality scan rule result log is produced in Logging for the resulting job.
Console
-
In the Google Cloud console, go to the Logs explorerpage.
-
In the Logs Explorerview, find the Querytab.
-
Click the Resourcemenu.
-
Select Cloud Dataplex DataScan. Click Apply.
-
Click the Log namemenu.
-
In the Search log namesfield, enter
dataplex.googleapis.com%2Fdata_quality_scan_rule_result. Select data_quality_scan_rule_resultand click Apply. -
Optional: Filter the logs to a specific data scan ID or location by adding the following filters in the log query:
resource.labels.location=" LOCATION " resource.labels.datascan_id=" DATA_SCAN_ID "
-
Click Run query.
gcloud
To read your data quality scan rule result log entries, use the gcloud logging read
command
with the following query:
gcloud logging read \ 'resource.type="dataplex.googleapis.com/DataScan" AND logName=projects/ PROJECT_ID /logs/dataplex.googleapis.com%2Fdata_quality_scan_rule_result AND resource.labels.location= LOCATION AND resource.labels.datascan_id= DATA_SCAN_ID ' --limit 10
REST
To list log entries, use the entries.list
method
.
Query discovery logs
When you use Dataplex Universal Catalog to discover data in assets, a discovery log is produced in Logging.
Console
-
In the Google Cloud console, go to the Logs explorerpage.
-
In the Logs Explorerview, find the Querytab.
-
Click the Resourcemenu.
-
Select Cloud Dataplex Zone. Click Apply.
-
Click the Log namemenu.
-
In the Search log namesfield, enter
dataplex.googleapis.com%2Fdiscovery. Select discoveryand click Apply. -
Optional: Filter the logs to a specific asset by adding the following filters in the log query:
resource.labels.location=" LOCATION " resource.labels.lake_id=" LAKE_ID " resource.labels.zone_id=" ZONE_ID " jsonPayload.assetId=" ASSET_ID "
-
Click Run query.
gcloud
To read your discovery log entries, use the gcloud logging read
command
with the following query:
gcloud logging read \ 'resource.type="dataplex.googleapis.com/Zone" AND logName=projects/ PROJECT_ID /logs/dataplex.googleapis.com%2Fdiscovery AND resource.labels.location= LOCATION AND resource.labels.lake_id= LAKE_ID AND resource.labels.zone_id= ZONE_ID AND jsonPayload.assetId= ASSET_ID ' --limit 10
REST
To list log entries, use the entries.list
method
.
Query metadata job logs
When you run a metadata import job , metadata job logs are produced in Logging.
Console
-
In the Google Cloud console, go to the Logs explorerpage.
-
In the Logs Explorerview, find the Querytab.
-
Click the Resourcemenu.
-
Select Cloud Dataplex Metadata Job.
-
Optional: To filter the logs to a specific location or metadata job ID, select a location or job ID.
-
Click Apply.
-
Click the Log namemenu.
-
Type
dataplex.googleapis.com%2Fmetadata_joband then select metadata_job. -
Click Apply.
gcloud
To read your metadata job log entries, use the gcloud logging read
command
with the following query:
gcloud logging read \ 'resource.type="dataplex.googleapis.com/MetadataJob" AND logName=projects/ PROJECT_ID /logs/dataplex.googleapis.com%2Fmetadata_job AND resource.labels.location= LOCATION AND resource.labels.metadata_job_id= METADATA_JOB_ID --limit 10
REST
To list log entries, use the entries.list
method
.
Query process logs
When you use Dataplex Universal Catalog to schedule and run tasks , a process log is produced in Logging for the resulting job.
Console
-
In the Google Cloud console, go to the Logs explorerpage.
-
In the Logs Explorerview, find the Querytab.
-
Click the Resourcemenu.
-
Select Cloud Dataplex Task. Click Apply.
-
Click the Log namemenu.
-
In the Search log namesfield, enter
dataplex.googleapis.com%2Fprocess. Select processand click Apply. -
Optional: Filter the logs to a specific task by adding the following filters in the log query:
resource.labels.location=" LOCATION " resource.labels.lake_id=" LAKE_ID " resource.labels.task_id=" TASK_ID "
-
Click Run query.
gcloud
To read your process log entries, use the gcloud logging read
command
with the following query:
gcloud logging read \ 'resource.type="dataplex.googleapis.com/Task" AND logName=projects/ PROJECT_ID /logs/dataplex.googleapis.com%2Fprocess AND resource.labels.location= LOCATION AND resource.labels.lake_id= LAKE_ID AND resource.labels.task_id= TASK_ID ' --limit 10
REST
To list log entries, use the entries.list
method
.
What's next
- Learn more about Cloud Logging .
- Learn about Dataplex Universal Catalog monitoring .

