Collect WatchGuard EDR logs
This document explains how to ingest WatchGuard EDR logs to Google Security Operations using Google Cloud Storage V2.
WatchGuard EDR (formerly Panda Adaptive Defense) is a cloud-managed endpoint detection and response platform providing advanced threat protection, behavioral analysis, and threat hunting. The WatchGuard Cloud API provides programmatic access to security event data including detections, indicators of attack, and threat intelligence logs.
Before you begin
Make sure that you have the following prerequisites:
- A Google SecOps instance
- A GCP project with Cloud Storage API enabled
- Permissions to create and manage GCS buckets
- Permissions to manage IAM policies on GCS buckets
- Permissions to create Cloud Run services, Pub/Sub topics, and Cloud Scheduler jobs
- Privileged access to the WatchGuard Cloud console with administrator permissions
- A WatchGuard Cloud API key or OAuth2 credentials
Create Google Cloud Storage bucket
- Go to the Google Cloud Console .
- Select your project or create a new one.
- In the navigation menu, go to Cloud Storage > Buckets.
- Click Create bucket.
-
Provide the following configuration details:
Setting Value Name your bucket Enter a globally unique name (for example, watchguard-edr-logs)Location type Choose based on your needs (Region, Dual-region, Multi-region) Location Select the location (for example, us-central1)Storage class Standard (recommended for frequently accessed logs) Access control Uniform (recommended) Protection tools Optional: Enable object versioning or retention policy -
Click Create.
Collect WatchGuard EDR API credentials
Obtain API credentials
- Sign in to the WatchGuard Cloud console as an administrator.
- Go to Administration > Managed Access.
- Click API Accessor navigate to the API key management section.
- Click Generate API Key.
- Enter a name for the API key (for example,
Google SecOps Integration). -
Copy and save the following details in a secure location:
- API Key ID: The API access key
- API Secret: The API secret key
- Account ID: Your WatchGuard Cloud account ID
Determine API base URL
The WatchGuard Cloud API base URL depends on your data center region:
| Region | API Base URL |
|---|---|
| US | https://api.usa.cloud.watchguard.com
|
| EU | https://api.eu.cloud.watchguard.com
|
Test API access
-
Test your credentials before proceeding with the integration:
# Replace with your actual credentials WG_API_KEY = "your-api-key-id" WG_API_SECRET = "your-api-secret" WG_ACCOUNT_ID = "your-account-id" WG_BASE_URL = "https://api.usa.cloud.watchguard.com" # Get access token TOKEN = $( curl -s -X POST " ${ WG_BASE_URL } /oauth/token" \ -H "Content-Type: application/x-www-form-urlencoded" \ -d "grant_type=client_credentials&client_id= ${ WG_API_KEY } & client_secret= ${ WG_API_SECRET } & scope=api-access" \ | jq -r '.access_token' ) # Test API access - list indicators curl -s -X GET " ${ WG_BASE_URL } /rest/aether/v1/accounts/ ${ WG_ACCOUNT_ID } /indicators? $top =1" \ -H "Authorization: Bearer ${ TOKEN } "
Create service account for Cloud Run function
The Cloud Run function needs a service account with permissions to write to GCS bucket and be invoked by Pub/Sub.
Create service account
- In the GCP Console, go to IAM & Admin > Service Accounts.
- Click Create Service Account.
- Provide the following configuration details:
- Service account name: Enter
watchguard-edr-logs-collector-sa - Service account description: Enter
Service account for Cloud Run function to collect WatchGuard EDR logs
- Service account name: Enter
- Click Create and Continue.
- In the Grant this service account access to projectsection, add the following roles:
- Click Select a role.
- Search for and select Storage Object Admin.
- Click + Add another role.
- Search for and select Cloud Run Invoker.
- Click + Add another role.
- Search for and select Cloud Functions Invoker.
- Click Continue.
- Click Done.
These roles are required for:
- Storage Object Admin: Write logs to GCS bucket and manage state files
- Cloud Run Invoker: Allow Pub/Sub to invoke the function
- Cloud Functions Invoker: Allow function invocation
Grant IAM permissions on GCS bucket
Grant the service account write permissions on the GCS bucket:
- Go to Cloud Storage > Buckets.
- Click on your bucket name (for example,
watchguard-edr-logs). - Go to the Permissionstab.
- Click Grant access.
- Provide the following configuration details:
- Add principals: Enter the service account email (for example,
watchguard-edr-logs-collector-sa@PROJECT_ID.iam.gserviceaccount.com) - Assign roles: Select Storage Object Admin
- Add principals: Enter the service account email (for example,
- Click Save.
Create Pub/Sub topic
Create a Pub/Sub topic that Cloud Scheduler will publish to and the Cloud Run function will subscribe to.
- In the GCP Console, go to Pub/Sub > Topics.
- Click Create topic.
- Provide the following configuration details:
- Topic ID: Enter
watchguard-edr-logs-trigger - Leave other settings as default
- Topic ID: Enter
- Click Create.
Create Cloud Run function to collect logs
The Cloud Run function will be triggered by Pub/Sub messages from Cloud Scheduler to fetch logs from the WatchGuard Cloud API and write them to GCS.
- In the GCP Console, go to Cloud Run.
- Click Create service.
- Select Function(use an inline editor to create a function).
-
In the Configuresection, provide the following configuration details:
Setting Value Service name watchguard-edr-logs-collectorRegion Select region matching your GCS bucket (for example, us-central1)Runtime Select Python 3.12or later -
In the Trigger (optional)section:
- Click + Add trigger.
- Select Cloud Pub/Sub.
- In Select a Cloud Pub/Sub topic, choose the topic
watchguard-edr-logs-trigger. - Click Save.
-
In the Authenticationsection:
- Select Require authentication.
- Check Identity and Access Management (IAM).
-
Scroll down and expand Containers, Networking, Security.
-
Go to the Securitytab:
- Service account: Select the service account
watchguard-edr-logs-collector-sa.
- Service account: Select the service account
-
Go to the Containerstab:
- Click Variables & Secrets.
- Click + Add variablefor each environment variable:
Variable Name Example Value Description GCS_BUCKETwatchguard-edr-logsGCS bucket name GCS_PREFIXwatchguardPrefix for log files STATE_KEYwatchguard/state.jsonState file path WG_API_KEYyour-api-key-idWatchGuard Cloud API key ID WG_API_SECRETyour-api-secretWatchGuard Cloud API secret WG_ACCOUNT_IDyour-account-idWatchGuard Cloud account ID WG_API_BASEhttps://api.usa.cloud.watchguard.comWatchGuard Cloud API base URL MAX_RECORDS5000Max records per run PAGE_SIZE1000Records per page LOOKBACK_HOURS24Initial lookback period -
In the Variables & Secretssection, scroll down to Requests:
- Request timeout: Enter
600seconds (10 minutes)
- Request timeout: Enter
-
Go to the Settingstab:
- In the Resourcessection:
- Memory: Select 512 MiBor higher
- CPU: Select 1
- In the Resourcessection:
-
In the Revision scalingsection:
- Minimum number of instances: Enter
0 - Maximum number of instances: Enter
100(or adjust based on expected load)
- Minimum number of instances: Enter
-
Click Create.
-
Wait for the service to be created (1-2 minutes).
-
After the service is created, the inline code editorwill open automatically.
Add function code
- Enter mainin the Entry pointfield.
-
In the inline code editor, create two files:
-
main.py:
import functions_framework from google.cloud import storage import json import os import urllib3 from datetime import datetime , timezone , timedelta import time # Initialize HTTP client with timeouts http = urllib3 . PoolManager ( timeout = urllib3 . Timeout ( connect = 5.0 , read = 30.0 ), retries = False , ) # Initialize Storage client storage_client = storage . Client () # Environment variables GCS_BUCKET = os . environ . get ( 'GCS_BUCKET' ) GCS_PREFIX = os . environ . get ( 'GCS_PREFIX' , 'watchguard' ) STATE_KEY = os . environ . get ( 'STATE_KEY' , 'watchguard/state.json' ) WG_API_KEY = os . environ . get ( 'WG_API_KEY' ) WG_API_SECRET = os . environ . get ( 'WG_API_SECRET' ) WG_ACCOUNT_ID = os . environ . get ( 'WG_ACCOUNT_ID' ) WG_API_BASE = os . environ . get ( 'WG_API_BASE' , 'https://api.usa.cloud.watchguard.com' ) MAX_RECORDS = int ( os . environ . get ( 'MAX_RECORDS' , '5000' )) PAGE_SIZE = int ( os . environ . get ( 'PAGE_SIZE' , '1000' )) LOOKBACK_HOURS = int ( os . environ . get ( 'LOOKBACK_HOURS' , '24' )) def parse_datetime ( value : str ) - > datetime : """Parse ISO datetime string to datetime object.""" if value . endswith ( "Z" ): value = value [: - 1 ] + "+00:00" return datetime . fromisoformat ( value ) def get_access_token (): """ Obtain OAuth2 access token using client credentials grant. """ api_base = WG_API_BASE . rstrip ( '/' ) token_url = f " { api_base } /oauth/token" headers = { 'Content-Type' : 'application/x-www-form-urlencoded' , 'Accept' : 'application/json' } body = ( f "grant_type=client_credentials" f "&client_id= { WG_API_KEY } " f "&client_secret= { WG_API_SECRET } " f "&scope=api-access" ) backoff = 1.0 for attempt in range ( 3 ): response = http . request ( 'POST' , token_url , body = body , headers = headers ) if response . status == 429 : retry_after = int ( response . headers . get ( 'Retry-After' , str ( int ( backoff )))) print ( f "Rate limited (429) on token request. Retrying after { retry_after } s..." ) time . sleep ( retry_after ) backoff = min ( backoff * 2 , 30.0 ) continue if response . status != 200 : raise RuntimeError ( f "Failed to get access token: { response . status } - { response . data . decode ( 'utf-8' ) } " ) data = json . loads ( response . data . decode ( 'utf-8' )) return data [ 'access_token' ] raise RuntimeError ( "Failed to get access token after 3 retries" ) @functions_framework . cloud_event def main ( cloud_event ): """ Cloud Run function triggered by Pub/Sub to fetch WatchGuard EDR security event logs and write to GCS. Args: cloud_event: CloudEvent object containing Pub/Sub message """ if not all ([ GCS_BUCKET , WG_API_KEY , WG_API_SECRET , WG_ACCOUNT_ID ]): print ( 'Error: Missing required environment variables' ) return try : bucket = storage_client . bucket ( GCS_BUCKET ) # Load state state = load_state ( bucket , STATE_KEY ) # Determine time window now = datetime . now ( timezone . utc ) last_time = None if isinstance ( state , dict ) and state . get ( "last_event_time" ): try : last_time = parse_datetime ( state [ "last_event_time" ]) # Overlap by 2 minutes to catch any delayed events last_time = last_time - timedelta ( minutes = 2 ) except Exception as e : print ( f "Warning: Could not parse last_event_time: { e } " ) if last_time is None : last_time = now - timedelta ( hours = LOOKBACK_HOURS ) print ( f "Fetching logs from { last_time . isoformat () } to { now . isoformat () } " ) # Get access token token = get_access_token () # Fetch logs from multiple endpoints all_records = [] newest_event_time = None for endpoint_type in [ 'indicators' , 'threats' ]: records , newest_time = fetch_logs ( token = token , endpoint_type = endpoint_type , start_time = last_time , end_time = now , page_size = PAGE_SIZE , max_records = MAX_RECORDS , ) all_records . extend ( records ) if newest_time : if newest_event_time is None or parse_datetime ( newest_time ) > parse_datetime ( newest_event_time ): newest_event_time = newest_time if not all_records : print ( "No new log records found." ) save_state ( bucket , STATE_KEY , now . isoformat ()) return # Write to GCS as NDJSON timestamp = now . strftime ( '%Y%m %d _%H%M%S' ) object_key = f " { GCS_PREFIX } /logs_ { timestamp } .ndjson" blob = bucket . blob ( object_key ) ndjson = ' \n ' . join ([ json . dumps ( record , ensure_ascii = False ) for record in all_records ]) + ' \n ' blob . upload_from_string ( ndjson , content_type = 'application/x-ndjson' ) print ( f "Wrote { len ( all_records ) } records to gs:// { GCS_BUCKET } / { object_key } " ) # Update state with newest event time if newest_event_time : save_state ( bucket , STATE_KEY , newest_event_time ) else : save_state ( bucket , STATE_KEY , now . isoformat ()) print ( f "Successfully processed { len ( all_records ) } records" ) except Exception as e : print ( f 'Error processing logs: { str ( e ) } ' ) raise def load_state ( bucket , key ): """Load state from GCS.""" try : blob = bucket . blob ( key ) if blob . exists (): state_data = blob . download_as_text () return json . loads ( state_data ) except Exception as e : print ( f "Warning: Could not load state: { e } " ) return {} def save_state ( bucket , key , last_event_time_iso : str ): """Save the last event timestamp to GCS state file.""" try : state = { 'last_event_time' : last_event_time_iso } blob = bucket . blob ( key ) blob . upload_from_string ( json . dumps ( state , indent = 2 ), content_type = 'application/json' ) print ( f "Saved state: last_event_time= { last_event_time_iso } " ) except Exception as e : print ( f "Warning: Could not save state: { e } " ) def fetch_logs ( token : str , endpoint_type : str , start_time : datetime , end_time : datetime , page_size : int , max_records : int ): """ Fetch security event logs from WatchGuard Cloud API with OData-style pagination and rate limiting. Args: token: OAuth2 access token endpoint_type: API endpoint type (indicators, threats) start_time: Start time for log query end_time: End time for log query page_size: Number of records per page max_records: Maximum total records to fetch Returns: Tuple of (records list, newest_event_time ISO string) """ api_base = WG_API_BASE . rstrip ( '/' ) endpoint = f " { api_base } /rest/aether/v1/accounts/ { WG_ACCOUNT_ID } / { endpoint_type } " headers = { 'Authorization' : f 'Bearer { token } ' , 'Accept' : 'application/json' , 'User-Agent' : 'GoogleSecOps-WatchGuardEDRCollector/1.0' } records = [] newest_time = None page_num = 0 skip = 0 backoff = 1.0 start_iso = start_time . strftime ( '%Y-%m- %d T%H:%M:%SZ' ) end_iso = end_time . strftime ( '%Y-%m- %d T%H:%M:%SZ' ) while True : page_num += 1 if len ( records ) > = max_records : print ( f "Reached max_records limit ( { max_records } ) for { endpoint_type } " ) break url = f " { endpoint } ?$top= { min ( page_size , max_records - len ( records )) } & $skip= { skip } & $filter=date ge { start_iso } and date le { end_iso } & $orderby=date asc" try : response = http . request ( 'GET' , url , headers = headers ) # Handle rate limiting with exponential backoff if response . status == 429 : retry_after = int ( response . headers . get ( 'Retry-After' , str ( int ( backoff )))) print ( f "Rate limited (429). Retrying after { retry_after } s..." ) time . sleep ( retry_after ) backoff = min ( backoff * 2 , 30.0 ) continue backoff = 1.0 if response . status != 200 : print ( f "HTTP Error: { response . status } " ) response_text = response . data . decode ( 'utf-8' ) print ( f "Response body: { response_text } " ) return records , newest_time data = json . loads ( response . data . decode ( 'utf-8' )) page_results = data . get ( 'value' , data . get ( 'data' , [])) if not page_results : print ( f "No more results (empty page) for { endpoint_type } " ) break print ( f " { endpoint_type } page { page_num } : Retrieved { len ( page_results ) } events" ) # Add endpoint type for identification for event in page_results : event [ '_wg_log_type' ] = endpoint_type records . extend ( page_results ) # Track newest event time for event in page_results : try : event_ts = event . get ( 'date' ) or event . get ( 'timestamp' ) or event . get ( 'createdAt' ) if event_ts : event_time = str ( event_ts ) if newest_time is None or parse_datetime ( event_time ) > parse_datetime ( newest_time ): newest_time = event_time except Exception as e : print ( f "Warning: Could not parse event time: { e } " ) # Check for more results if len ( page_results ) < page_size : print ( f "No more pages for { endpoint_type } (last page not full)" ) break skip += len ( page_results ) except Exception as e : print ( f "Error fetching { endpoint_type } logs: { e } " ) return records , newest_time print ( f "Retrieved { len ( records ) } total { endpoint_type } records from { page_num } pages" ) return records , newest_time -
requirements.txt:
functions-framework==3.* google-cloud-storage==2.* urllib3>=2.0.0
-
-
Click Deployto save and deploy the function.
-
Wait for deployment to complete (2-3 minutes).
Create Cloud Scheduler job
Cloud Scheduler will publish messages to the Pub/Sub topic at regular intervals, triggering the Cloud Run function.
- In the GCP Console, go to Cloud Scheduler.
- Click Create Job.
-
Provide the following configuration details:
Setting Value Name watchguard-edr-logs-collector-hourlyRegion Select same region as Cloud Run function Frequency 0 * * * *(every hour, on the hour)Timezone Select timezone (UTC recommended) Target type Pub/Sub Topic Select the topic watchguard-edr-logs-triggerMessage body {}(empty JSON object) -
Click Create.
Schedule frequency options
Choose frequency based on log volume and latency requirements:
| Frequency | Cron Expression | Use Case |
|---|---|---|
|
Every 5 minutes
|
*/5 * * * *
|
High-volume, low-latency |
|
Every 15 minutes
|
*/15 * * * *
|
Medium volume |
|
Every hour
|
0 * * * *
|
Standard (recommended) |
|
Every 6 hours
|
0 */6 * * *
|
Low volume, batch processing |
|
Daily
|
0 0 * * *
|
Historical data collection |
Test the integration
- In the Cloud Schedulerconsole, find your job.
- Click Force runto trigger the job manually.
- Wait a few seconds.
- Go to Cloud Run > Services.
- Click on
watchguard-edr-logs-collector. - Click the Logstab.
-
Verify the function executed successfully. Look for:
Fetching logs from YYYY-MM-DDTHH:MM:SS+00:00 to YYYY-MM-DDTHH:MM:SS+00:00 indicators page 1: Retrieved X events threats page 1: Retrieved X events Wrote X records to gs://watchguard-edr-logs/watchguard/logs_YYYYMMDD_HHMMSS.ndjson Successfully processed X records -
Go to Cloud Storage > Buckets.
-
Click on your bucket name (
watchguard-edr-logs). -
Navigate to the
watchguard/folder. -
Verify that a new
.ndjsonfile was created with the current timestamp.
If you see errors in the logs:
- HTTP 401: Check API credentials in environment variables
- HTTP 403: Verify API key has required permissions in WatchGuard Cloud console
- HTTP 429: Rate limiting - function will automatically retry with backoff
- Missing environment variables: Check all required variables are set
Configure a feed in Google SecOps to ingest WatchGuard EDR logs
- Go to SIEM Settings > Feeds.
- Click Add New Feed.
- Click Configure a single feed.
- In the Feed namefield, enter a name for the feed (for example,
WatchGuard EDR Logs). - Select Google Cloud Storage V2as the Source type.
- Select WatchGuard EDRas the Log type.
-
Click Get Service Account. A unique service account email will be displayed, for example:
chronicle-12345678@chronicle-gcp-prod.iam.gserviceaccount.com -
Copy this email address.
-
Click Next.
-
Specify values for the following input parameters:
-
Storage bucket URL: Enter the GCS bucket URI with the prefix path:
gs://watchguard-edr-logs/watchguard/- Replace:
-
watchguard-edr-logs: Your GCS bucket name. -
watchguard: Optional prefix/folder path where logs are stored (leave empty for root).
-
- Replace:
-
Source deletion option: Select the deletion option according to your preference:
- Never: Never deletes any files after transfers (recommended for testing).
- Delete transferred files: Deletes files after successful transfer.
-
Delete transferred files and empty directories: Deletes files and empty directories after successful transfer.
-
Maximum File Age: Include files modified in the last number of days (default is 180 days)
-
Asset namespace: The asset namespace
-
Ingestion labels: The label to be applied to the events from this feed
-
-
Click Next.
-
Review your new feed configuration in the Finalizescreen, and then click Submit.
Grant IAM permissions to the Google SecOps service account
The Google SecOps service account needs Storage Object Viewerrole on your GCS bucket.
- Go to Cloud Storage > Buckets.
- Click on your bucket name.
- Go to the Permissionstab.
- Click Grant access.
- Provide the following configuration details:
- Add principals: Paste the Google SecOps service account email
- Assign roles: Select Storage Object Viewer
- Click Save.
UDM mapping table
| Log Field | UDM Mapping | Logic |
|---|---|---|
| |
about.asset.asset_id | Set to device_vendor.device_product:deviceExternalId |
| |
about.file.full_path | Value from filePath if not empty, else from _hash if file_is_not_hash, else from fileHash if file_is_not_hash |
| |
about.file.size | Value from fsize if > 0 |
| |
about.hostname | Value from dvchost |
| |
about.ip | Merged from dvc array after IP validation |
| |
about.mac | Value from dvcmac if valid MAC, else from dvc_mac if slot not present |
| |
about.nat_ip | Value from deviceTranslatedAddress |
| |
about.process.command_line | Value from Subject if not present, else Emne, else Path |
| |
about.process.pid | Value from dvcpid |
| |
about.resource.attribute.permissions | Value from filePermission |
| |
about.resource.attribute.labels | Value from resource_Type_label |
| |
additional.fields | Merged from various additional_* labels like additional_eventId, additional_devicePayloadId, etc. |
| |
metadata.collected_timestamp | Value from alertDateTime if ISO format, else from AlertDate if ISO format, else from Received if ISO format, else from Generated if ISO format |
| |
metadata.description | Value from msg |
| |
metadata.event_timestamp | Value from Date if ISO format |
| |
metadata.event_type | Set to "PROCESS_UNCATEGORIZED" if file_full_path not empty, else "SCAN_UNCATEGORIZED" if event_name in LogSpyware or LogPredictiveMachineLearning, else "STATUS_UPDATE" if has_principal true, else "GENERIC_EVENT" |
| |
metadata.product_event_type | Value from huntingRule if not empty, else from ThreatType, else from device_event_class_id - event_name, else from device_event_class_id, else from event_name |
| |
metadata.product_log_id | Value from pandaAlertId if not empty, else from externalId |
| |
metadata.product_name | Set to "ALERTS" for JSON, else from device_product |
| |
metadata.product_version | Value from device_version |
| |
metadata.url_back_to_product | Value from directLink |
| |
metadata.vendor_name | Set to "WATCHGUARD" for JSON, else from device_vendor |
| |
network.application_protocol | Value from app_protocol_output if not empty |
| |
network.direction | Set to "INBOUND" if deviceDirection == 0, else "OUTBOUND" if == 1 |
| |
network.http.method | Value from requestMethod |
| |
network.http.user_agent | Value from requestClientApplication |
| |
network.ip_protocol | Value from ip_protocol_out if not empty |
| |
network.received_bytes | Value from in if > 0 and integer |
| |
network.sent_bytes | Value from out if > 0 and integer |
| |
principal.administrative_domain | Value from sntdom if not empty, else from Domain, else from Domene |
| |
principal.application | Value from sourceServiceName |
| |
principal.asset.asset_id | Value from MUID if not empty |
| |
principal.asset.hostname | Value from machineName if not empty, else from HostName, else from SourceMachineName, else from MachineName |
| |
principal.asset.ip | Merged from HostIp if not empty, else SourceIP, else MachineIP |
| |
principal.asset.product_object_id | Value from ClientId |
| |
principal.group.group_display_name | Value from Group_name if not empty, else Gruppenavn |
| |
principal.hostname | Value from machineName if not empty, else from temp_dhost if not empty, else from shost if IP validation fails, else from Device_name, else Enhetsnavn |
| |
principal.ip | Value from principal_ip if valid IP, else from src if valid IP |
| |
principal.mac | Value from smac if valid MAC |
| |
principal.nat_ip | Value from sourceTranslatedAddress if valid IP |
| |
principal.nat_port | Value from sourceTranslatedPort if > 0 |
| |
principal.port | Value from spt if integer and not 0 |
| |
principal.process.command_line | Value from sproc |
| |
principal.process.pid | Value from spid |
| |
principal.user.attribute.roles | Value from spriv |
| |
principal.user.user_display_name | Value from suser if not starts with {, else from SourceUserName, else from CustomerName if not empty |
| |
principal.user.userid | Value from contents.0.LoggedUser if not empty and sets event_type to USER_UNCATEGORIZED, else from temp_duid, else from User, else Bruker |
| |
security_result.action | Set to "ALLOW" if act in accept/notified or outcome REDIRECTED_USER_MAY_PROCEED or categoryOutcome Success or cs2 Allow, else "BLOCK" if act deny/blocked or outcome BLOCKED or categoryOutcome Failure or cs2 Denied, else "FAIL" if outcome Failure |
| |
security_result.action_details | Value from act, else from Action_Taken |
| |
security_result.attack_details.tactics | Merged from tactics_data if not empty |
| |
security_result.attack_details.techniques | Merged from technique_data if not empty |
| |
security_result.category_details | Value from cat |
| |
security_result.description | Value from msg_data_2 if not empty, else from THRuleName, else from Type, else Scan_Type |
| |
security_result.detection_fields | Merged from operation_label, operasjon_label, permission_label, tillatelse_label, infection_channel_label, spyware_Grayware_Type_label, threat_probability_label |
| |
security_result.rule_name | Value from mwProfile |
| |
security_result.severity | Set to "INFORMATIONAL" if severity == 1, "LOW" if 2, "MEDIUM" if 3, "HIGH" if 4, "CRITICAL" if 5; else LOW if severity in 0/1/2, MEDIUM if 3/4/5/INFO, HIGH if 6/7/SEVERE, CRITICAL if 8/9/10/VERY-HIGH |
| |
security_result.summary | Value from appcategory, else Result |
| |
security_result.threat_name | Value from Spyware if not empty, else Virus_Malware_Name, else Unknown_Threat |
| |
src.file.full_path | Value from oldFilePath |
| |
src.file.size | Value from oldFileSize if > 0 |
| |
target.administrative_domain | Value from dntdom |
| |
target.application | Value from destinationServiceName |
| |
target.file.full_path | Value from ItemPath |
| |
target.file.md5 | Value from ItemHash if lowercase succeeds |
| |
target.hostname | Value from temp_dhost if not empty |
| |
target.ip | Value from dst_ip if valid IP |
| |
target.mac | Value from dmac if valid MAC |
| |
target.nat_ip | Value from destination_translated_address if valid IP |
| |
target.nat_port | Value from destinationTranslatedPort if integer |
| |
target.port | Value from dpt if integer and in range |
| |
target.process.command_line | Value from dproc |
| |
target.process.file.full_path | Value from contents.0.ChildPath if not empty, else from file_full_path |
| |
target.process.file.md5 | Value from contents.0.ChildMd5 if lowercase succeeds |
| |
target.process.file.names | Merged from file_name if not empty |
| |
target.process.parent_process.file.full_path | Value from contents.0.ParentPath |
| |
target.process.parent_process.file.md5 | Value from contents.0.ParentMd5 if lowercase succeeds |
| |
target.process.parent_process.file.names | Merged from parent_file_name if not empty |
| |
target.process.parent_process.pid | Value from contents.0.ParentPID |
| |
target.process.pid | Value from dpid |
| |
target.resource.attribute.labels | Value from DriveType_label, ServiceLevel_label |
| |
target.url | Value from request |
| |
target.user.user_display_name | Value from temp_duser |
| |
target.user.userid | Value from temp_duid if not empty |
Need more help? Get answers from Community members and Google SecOps professionals.

