The following table describes important Backup and DR Service event IDs, event messages, and steps to resolve them:
Event ID | Event message | What to do |
---|---|---|
5022
|
Actifio Connector: Failed in preparing VSS snapshot set | This issue occurs if Windows fails to create a VSS snapshot. To resolve |
|
this issue, do the following: | |
|
||
|
* Check UDSAgent.log | |
|
* Check disk space on protected volumes. 300MB may not be enough. | |
|
* Check Windows Event Logs for VSS related errors. | |
|
* vssadmin list writers may show writers in a bad state. | |
|
||
|
Usually these errors are accompanied by VSS errors reported in the logs | |
|
such as: VSS_E_VOLUME_NOT_SUPPORTED_BY_PROVIDER
|
|
|
VSS_E_UNEXPECTED_PROVIDER_ERROR
|
|
|
||
|
First check if all the VSS writers are in a stable state by going to the | |
|
command line and issuing this command. # vssadmin list writers
|
|
|
||
|
Check output to confirm that all the writers are in a stable state. | |
|
||
|
Restart VSS service and check if the writers are stable. If not you may | |
|
have to reboot the machine. | |
5024
|
Actifio Connector: Failed to create VSS snapshot for backup. Insufficient | This issue occurs if there is insufficient disk space to process a |
|
storage available to create either the shadow copy storage file or other | snapshot. |
|
shadow copy data | |
|
1. Ensure the drive being backed up is not full. | |
|
2. Check if all the VSS writers are in a stable state From the Windows | |
|
command line, run: vssadmin list providers
vssadmin list writers
|
|
|
3. If these services are not running, start them and re-run the job. If | |
|
the writer's State is Not Stable, restart the VSS service. If the | |
|
problem continues after restarting the service, reboot the host. | |
|
||
|
Sometimes the message appears when internal VSS errors occur. | |
|
||
|
Check the Windows Event Logs for VSS related errors. For errors related | |
|
to VSS, search for related Microsoft patches. Additional VSS | |
|
troubleshooting details can be found on Microsoft TechNet. | |
|
||
|
Microsoft recommends at least 320MB on devices specified for saving the | |
|
created VSS snapshot, plus change data that is stored there. | |
|
||
|
Actifio recommends the shadow storage space be set to unbounded | |
|
(unlimited) using these commands: vssadmin list shadowstorage
|
|
|
vssadmin Resize ShadowStorage /On=[drive]: /For=[drive]: / Maxsize=[size]
|
|
|
||
|
To change the storage area size in the Windows UI, refer to | |
|
Configuring Volume Shadow Copy on Windows Server 2008 . | |
|
||
|
Re-run the backup once the VSS state is stable and shadow storage is set | |
|
to unbounded. | |
5046
|
Backup staging LUN is not visible to the Actifio Connector | This issue occurs if the staging LUN is not visible to the UDSAgent on |
|
the application's host and the host is unable to detect the staging LUN | |
|
from the backup/recovery appliance. | |
5049
|
Actifio Connector: Failed identifying logical volume on the backup staging | Actifio Connector couldn't see the staging LUN. This can be caused by a |
|
lun | bad connection or by trouble on the LUN. |
|
||
|
Verify that FC/iSCSI connectivity is good, then make sure it works by | |
|
mapping the VDisk, partitioning it, formatting it, and copying files to | |
|
it. The steps for partitioning and formatting are OS specific. | |
5078
|
Actifio Connector: The staging disk is full | Jobs fail if a file that was modified in the source disk is copied to |
|
the staging disk, but the file is larger than the free space available | |
|
in the staging disk. To fix the issue with full staging disk, increase | |
|
the staging disk. Specify the size of the staging disk in the advanced | |
|
settings for the application. Set the value for staging disk size such | |
|
that it is greater than the sum of size of the source disk and the size | |
|
of the largest file. Note: Changing the staging disk in advanced | |
|
settings provokes a full backup. | |
5087
|
Actifio Connector: Failed to write files during a backup (Source File) | Anti-virus programs or third party drivers may have applied file locks |
|
that cannot be overridden. | |
|
||
|
Check the UDSAgent.log to see which file couldn't be accessed. Attempt | |
|
to find which process is locking the file using lsof on Unix/ Linux, or | |
|
fltmc on Windows. Exclude the file from the antivirus or capture job | |
|
and re-try the capture. | |
|
||
|
The current processes known to Microsoft are listed at: | |
|
Allocated filter altitudes . | |
|
||
|
These errors are rarely found on Unix or Linux, but it is possible that | |
|
a process such as database maintenance or patch install / update has | |
|
created an exclusive lock on a file. | |
|
||
|
Install the latest Actifio Connector. | |
|
||
|
A file system limitation or inconsistency was detected by the host | |
|
operating system. | |
|
||
|
Run the Windows Disk defragmenter on the staging disk. | |
|
||
|
Low I/O throughput from the hosts disks or transport medium, iSCSI or | |
|
FC. | |
|
||
|
Ensure there are no I/O issues in the host's disks or transport medium. | |
|
The transport medium will either be iSCSI or Fibre Channel depending on | |
|
out of band configuration. Consult storage and network administrators as | |
|
needed. | |
5131 - SQL Logs report error 3041
|
SQL log backups on instance fail with error 5131 | To resolve this, enable "Don't forcefully unload the user registry at |
|
user logoff", see User Profile Service functionality . | |
5131 - SQL logs show backup/ recovery appliances error 43901
|
Snapshot jobs fail with error 5131, SQL logs show backup/ recovery | This is because the ODBC login for the database is failing. Fixing the |
|
"Failed snapshot Job" | ODBC login resolves the issue. |
|
appliances error 43901 | |
5136
|
Actifio Connector: The staging volume is not readable | Check /act/logs/UDSAgent.log for details and contact Google support for |
|
the resolution for the issue. | |
5241
|
Actifio Connector: Failed to mount/clone applications from mapped image | Invalid username and password being parsed from the control file. |
|
(Source File) | On the source, review the UDSAgent.log to see if the source is |
|
configured with the correct username/password under Advanced Settings in | |
|
the connector properties. | |
5547
|
Oracle: Failed to backup archivelog (Source File) | Actifio Connector failed to backup the archive log using RMAN archive |
|
backup commands. The likely causes for this failure are: | |
|
||
|
* Connector failed to establish connection to database | |
|
* The archive logs were purged by another application | |
|
* TNS Service name is configured incorrectly, causing backup command | |
|
to be sent to a node where the staging disk isn't mounted | |
|
||
|
Search for ORA- or RMAN- errors in the RMAN log. This is the error | |
|
received from Oracle. Use the preferred Oracle resource as these are not | |
|
Backup and DR Service conditions, and hence cannot be resolved within | |
|
Backup and DR Service. | |
|
* Actifio Connector logs: /var/act/log/UDSAgent.log | |
|
* Oracle RMAN logs: /var/act/log/********_rman.log | |
10032
|
Snapshot pool exceeded warning level | To reduce consumption of the snapshot pool, do the following: |
|
||
|
* Move VMware VMs from a snapshot to a Direct-to-OnVault backup plan. | |
|
Then expire all snapshots to release the space used by the staging | |
|
disks and last snap. This only works for VMware VMs; other | |
|
application types still use some snapshot pool space if protected by | |
|
a Direct-to-OnVault policy. | |
|
* Reduce the number of snaps kept for an application by changing the | |
|
policy template. Applications that have high change rates create | |
|
larger snapshots, so this has the highest benefit for high | |
|
change-rate applications. This does not necessarily lead to a | |
|
different RPO, an OnVault images of each snap can be created before | |
|
they are expired. | |
|
* Delete mounts, clones, and live-clones if they are not needed. | |
10038
|
About to exceed VDisk warning limit | To immediately reduce VDisk consumption, do the following: |
|
||
|
* Ensure expirations are enabled, both at the global and individual | |
|
application level. | |
|
* Group databases from a single host together into a Consistency Group. | |
|
For example, if a host has 9 databases, create one Consistency Group | |
|
for all 9 databases, then protect that consistency group rather than | |
|
the individual databases. | |
|
* Reduce the number of snapshots kept for an application by changing | |
|
the policy template used by a backup plan. | |
|
* Delete unwanted mounts, clones, and live-clones images. | |
|
* Move VMware VMs from a snapshot to a Direct- OnVault backup plan. | |
|
You need to expire all snapshots to release the VDisks used by the | |
|
staging disks. This only lowers the VDisk count for VMware VMs, and | |
|
still uses VDisks when protected by a direct-to-OnVault policy. | |
|
* Change VMware VMDKs that don't need to be protected to independent | |
|
mode as these cannot be protected by VMware snapshots. | |
|
||
|
If this alert repeats daily but the appliance does not reach the maximum | |
|
VDisks, then modify the policies to reduce the number of VDisks used, or | |
|
increase the alert threshold. During a daily snapshot window the VDisk | |
|
count can fluctuate while new VDisks are created for snapshots before | |
|
the old VDisks are removed as a part of snapshot expirations. The daily | |
|
fluctuations varies depending on the number of applications protected. | |
10039
|
Network error reaching storage device | A heartbeat ping to monitored storage has failed due to hardware |
|
failure or network issue. Check the network to resolve the issue. | |
10043
|
A backup plan violation has been detected | Review the backup plan best practices |
|
and optimize your policies. The are common causes for backup plan | |
|
violations. | |
|
||
|
* Job scheduler is not enabled. See to enable the scheduler . | |
|
* The first jobs for new applications can often take a long time: | |
|
Long job times can occur during the first snapshot or dedup job | |
|
for an application. On-ramp settings can be used to prevent | |
|
ingest jobs from locking up slots and locking out ingested | |
|
applications. See [Set priorities for the first new | |
|
applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | |
|
* Applications are inaccessible due to network issues. | |
|
* Policy windows are too small or job run times are too long: While | |
|
you cannot control how long each job takes to run, you can control | |
|
the schedule time for applications that are running. Jobs that run | |
|
for many hours occupy job slots that could be used by other | |
|
applications. Review the [backup plan best | |
|
practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | |
|
and adjust policies accordingly. | |
|
* Replication process sends the data to a remote | |
|
backup/recovery appliance. Ensure that the bandwidth & utilization of | |
|
your replication link is not saturated. | |
10046
|
Performance Pool exceeded safe threshold | To reduce consumption of the snapshot pool, do the following: |
|
||
|
* Move VMware VMs from a snapshot to a Direct-to-OnVault backup plan. | |
|
Then expire all snapshots to release the space used by the staging | |
|
disks and last snap. This only works for VMware VMs; other | |
|
application types still use some snapshot pool space if protected by | |
|
a Direct-to-OnVault policy. | |
|
* Reduce the number of snaps kept for an application by changing the | |
|
policy template. Applications that have high change rates create | |
|
larger snapshots, so this has the highest benefit for high | |
|
change-rate applications. This does not necessarily lead to a | |
|
different RPO, an OnVault images of each snap can be created before | |
|
they are expired. | |
|
* Delete mounts, clones, and live-clones if they are not needed. | |
10055
|
Unable to check remote protection | Each backup/recovery appliance checks the remote appliance hourly for |
|
possible remote protection issues. The appliance communication fails due | |
|
to the following issues: | |
|
||
|
* Network error (temporary or permanent). Temporary network error does | |
|
not mean job to fail; jobs are retried, but the hourly check is not | |
|
updated. | |
|
* Certificate error. To fix the certificate error, you need to | |
|
re-exchange the certificate. | |
10070
|
Udppm scheduler is off for more than 30 minutes. | The scheduler is off. This may have been set for maintenance. If the |
|
maintenance is complete, you can re-enable the scheduler, see to | |
|
enable the scheduler . | |
10084
|
Alert for application (app name) and policy (policyname) job did not run | Review the backup plan best practices |
|
because of unknown reason | and optimize your policies. The are common causes for backup plan |
|
violations. | |
|
||
|
* Job scheduler is not enabled. See to enable the scheduler . | |
|
* The first jobs for new applications can often take a long time: | |
|
Long job times can occur during the first snapshot or dedup job | |
|
for an application. On-ramp settings can be used to prevent | |
|
ingest jobs from locking up slots and locking out ingested | |
|
applications. See [Set priorities for the first new | |
|
applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | |
|
* Applications are inaccessible due to network issues. | |
|
* Policy windows are too small or job run times are too long: While | |
|
you cannot control how long each job takes to run, you can control | |
|
the schedule time for applications that are running. Jobs that run | |
|
for many hours occupy job slots that could be used by other | |
|
applications. Review the [backup plan best | |
|
practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | |
|
and adjust policies accordingly. | |
|
* Replication process sends the data to a remote | |
|
backup/recovery appliance. Ensure that the bandwidth & utilization of | |
|
your replication link is not saturated. | |
10085
|
Backup Plan violation for application (app name) on host (host name) and | Review the backup plan best practices |
|
policy (policy name). Job did not run because of unknown reason. | and optimize your policies. The are common causes for backup plan |
|
violations. | |
|
||
|
* Job scheduler is not enabled. See to enable the scheduler . | |
|
* The first jobs for new applications can often take a long time: | |
|
Long job times can occur during the first snapshot or dedup job | |
|
for an application. On-ramp settings can be used to prevent | |
|
ingest jobs from locking up slots and locking out ingested | |
|
applications. See [Set priorities for the first new | |
|
applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | |
|
* Applications are inaccessible due to network issues. | |
|
* Policy windows are too small or job run times are too long: While | |
|
you cannot control how long each job takes to run, you can control | |
|
the schedule time for applications that are running. Jobs that run | |
|
for many hours occupy job slots that could be used by other | |
|
applications. Review the [backup plan best | |
|
practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | |
|
and adjust policies accordingly. | |
|
* Replication process sends the data to a remote | |
|
backup/recovery appliance. Ensure that the bandwidth & utilization of | |
|
your replication link is not saturated. | |
10120
|
Psrv started successfully | This is an internal event and can be ignored. |
10220
|
NTP Service is not running or not synchronised. | The NTP Service on the backup appliance is not running. The NTP Service |
|
is needed to ensure the backup appliance uses the correct timestamps. A | |
|
Compute Engine appliance should be using metadata.google.internal. | |
|
Follow how to set the NTP server [DNS and NTP | |
|
method](/backup-disaster-recovery/docs/concepts/appliance-network-settings#dns-ntp). | |
10225
|
Udp corefiles are found, filename udpengine.(file name) | Internal processes are unexpectedly logging error files. Contact Google |
|
support to get the resolution for this issue. | |
10229
|
Exceeded storage, System name: (device name) | This is is an internal event and normally can be ignored. |
10237
|
The X job has been running for more than 3 hours. | There are many reasons why a job can run for more than 3 hours. |
11001
|
Backup appliance certificate expires in X days. Power on for 24 hours to | The backup/recovery appliance's certificate last update is more than 15 |
|
renew, or contact support. | days old. If the backup/recovery appliance is down, bring it up. |
11004
|
System components are down. Please contact support in case backups are | Contact support. |
|
affected. | |
11006
|
Not able to sync with host X, regular sync with host required to avoid | The certificate on the host hasn't been updated for more than 7 days. |
|
permanent communication loss between Backup appliance and Host. | Try rebooting and reconnecting to the host. |
20019
|
Insufficient CPU / Memory. Minimum number of core required: (cores) Actual | Backup/recovery appliance has been changed and is not the recommended |
|
cores : (cores). Minimum memory size required (GB): (memory) Actual memory : | size. Contact Google support to get the resolution for this issue. |
|
(memory) | |
20025
|
Swap usage exceeded | This issue occurs when the swap usage is exceeding the configured |
|
threshold limit that is set for the backup/recoevery appliance. Contact | |
|
Google support to get the resolution for this issue. | |
20030
|
tomcat stopped successfully | This is is an internal event and can be ignored. |
20031
|
tomcat started successfully | This is is an internal event and can be ignored. |
22001
|
OMD started Successfully, sltname: , slpname: . | This is is an internal event and can be ignored. |
42356
|
File changes have been detected no deleted files have been detected new | This is is an internal event and can be ignored. |
|
files have been detected. | |
43151
|
couldn't add raw device mappings to virtual machine (VM). Error: VM task | Adding a raw device mapping to a VM "stuns" the VM until ESX has had a |
|
failed A general system error occurred: The system returned an error. | chance to add the new resource. To find out why the raw device mapping |
|
couldn't be added, look at the ESX logs for the VM in question | |
|
(vmware.log). | |
|
||
|
Refer to the VMware documentation and knowledge base for assistance on | |
|
reviewing the logs for error messages. Also, review the VMware article | |
|
for more information on [collecting VMware | |
|
logs](https://kb.vmware.com/s/article/653). | |
43155
|
Error: VM task failed. An error occurred while saving the snapshot: Failed | This is a VMware issue; for additional information, refer to VMware KB |
|
to quiesce the virtual machine. | article - 1015180 . |
|
||
|
Virtual machine quiesce issues are dependent on the OS type. Additional | |
|
investigation, further VMware KBA searches or contact VMware support to | |
|
resolve this issue. | |
43155 - a
|
Error: VM task failed. Device scsi3 couldn't be hot-added. | This usually means that the SCSI device you are trying to add to the VM |
|
is already in use by another VM. | |
43155 - b
|
Error: VM task failed. The virtual disk is either corrupted or not a | This issue occurs if the VM's CTK files are locked, unreadable, or are |
|
supported format. | being committed. To fix this issue, remove and re-create these CTK |
|
files. Refer to the KB article - 2013520 | |
|
for more information. | |
43155 - c
|
Error: VM task failed. The operation is not allowed in the current state | There are two options for formatting a VMware datastore: NFS and VMFS. |
|
of the datastore." progress ="11" status="running" | With NFS, there are some limitations like not being able to do RDM (Raw |
|
Disk Mapping). This means that you cannot mount from the | |
|
backup/recovery appliance to an NFS datastore. Refer to the following KB | |
|
article - 1001856 for | |
|
additional information. | |
43175
|
UDSAgent socket connection got terminated abnormally; while waiting for | The Actifio Connector stops responding between the appliance and a host |
|
the response from agent | with Backup and DR agent is installed. |
|
||
|
1. Restart the UDSAgent Backup and DR agent service on the | |
|
specified host. | |
|
2. Telnet to tcp port 5106 (UDSAgent communication port) | |
|
||
|
``` | |
|
<P></P> | |
|
Expected output: | |
|
||
|
<P></P> | |
|
Trying 10.50.100.67... | |
|
||
|
<P></P> | |
|
Connected to dresx2.accu.local. | |
|
||
|
<P></P> | |
|
Escape character is '^]'. | |
|
||
|
<P></P> | |
|
Connection closed by foreign host. | |
|
``` | |
|
||
|
3. Verify network connectivity between appliance and host doesn't drop. | |
|
If the problem persists, network analysis will be required. | |
43604
|
Failed to verify fingerprint | This occurs when an inconsistency is found between the source and target |
|
data. Contact Google support to get the resolution for this issue. | |
43690
|
Host doesn't have any SAN or iSCSI ports defined. | This issue occurs if the backup/recovery appliance is not configured with |
|
iSCSI connection to the target host. | |
|
||
|
Ensure that the network ports are open for iSCSI and the target host has | |
|
discovered the backup/recovery appliances. | |
43698
|
ESX host is not accessible for NBD mode data movement | The backup/recovery appliance is unable to reach the ESX host over the |
|
network or resolve the ESX hostname using DNS. Contact Google support to | |
|
get the resolution for this issue. | |
43702
|
Backup was aborted because there are too many extra files in the home | This is an alert condition generated by Backup and DR Service and is caused |
|
directory of the VM | by leftover delta files in the VM's datastore. Normally, the delta files |
|
are removed after Backup and DR snapshot is consolidated. In | |
|
some instances, these can be left behind by the VMware consolidation, | |
|
and Backup and DR begins failing jobs to prevent exacerbating | |
|
the issue. | |
|
||
|
This issue is caused by VMware, refer to the knowledge base article - | |
|
1002310 . | |
43755
|
Failed to open VMDK volume; check connectivity to ESX server. | This happens when the ESX server cannot be reached by the controller, |
|
usually because of a physical connection or DNS problem. To fix this | |
|
issue, do the following: | |
|
||
|
* Ensure port 902 is open between the backup/recovery appliance and the | |
|
ESX host. | |
|
* Check the current DNS server and ensure it is current and valid. | |
|
* If the vCenter is virtualized, attempt a backup after migrating the | |
|
vCenter to a different ESX host. | |
|
* Ensure SSL required is set to True on the ESX host in the advanced | |
|
settings. | |
43844
|
Invalid size vmdk detected for the VM | There are two possible solutions for this situation: |
|
||
|
* If consolidation is required for some disks on VM, size is reported | |
|
as zero. To fix this issue, creating and deleting a snapshot of the | |
|
VM. | |
|
* See if the VMDK can be restored from a backup image. | |
43873
|
Disk space usage on datastore has grown beyond the critical threshold | This issue occurs when the remaining space on the datastore is less than |
|
the critical threshold. If more storage is not made available soon, then | |
|
jobs start to fail when the remaining space is inadequate to store them. | |
|
||
|
This alert is created to help you take action to prevent ESX datastores | |
|
from filling with snapshot data. Increase available space by expanding | |
|
the datastore, migrating some VMs, or deleting old data on the | |
|
datastore. | |
|
||
|
Snapshots grow as more change data is added. If a datastore fills up due | |
|
to a growing snapshot, VMs may be taken offline automatically by VMware | |
|
to protect the data. | |
43900
|
Retry pending OnVault (log) (jobname for application (appname) on host | Job retries can be caused by many errors. Each 43900 event message |
|
(hostname) Error: (errorID) (Error Description) | includes an error code and an error message. |
43901
|
Job failure | Job failures can be caused by many errors. Each 43901 event message |
|
includes an error code and an error message. | |
43903
|
Failed expire job | This issue occurs when the image is in use at the time of the |
|
expiration. This can be due to the image is in use by another process or | |
|
operation, such as a mount, clone, restore. The expiration job most | |
|
likely complete successfully on the second attempt. | |
|
Backup and DR does not report the successful completion of | |
|
this second attempt. If you get only one error for an image, it is safe | |
|
to conclude that a second attempt to expire this image was successful. | |
|
If there is a legitimate reason why this image cannot be expired, you | |
|
will get multiple errors related to this image. If you receive more than | |
|
one error, contact Google Support. | |
43905
|
Failed mount job | There are many ways a mount job can fail. The error code that |
|
accompanies the event helps to identify the root cause. | |
43908
|
Failed restore job | Job failures can be caused by many errors. Each 43908 event message |
|
includes an error code and an error message. | |
43915
|
Couldn't connect to backup host. Make sure Backup and DR agent | To initiate backup, the Actifio Connector service must be reachable by |
|
is running on (host) and network port (port) is open | the backup/recovery appliance. This issue occurs, when the required ports |
|
are not open, the incorrect host IP is configured, the | |
|
Backup and DR agent service not running, or the host is out | |
|
of physical resources. To fix this issue, do the following: | |
|
||
|
1. Ensure that the port in use between the host, | |
|
backup/recovery appliance, and Actifio Connector is open. By default, | |
|
the Backup and DR agent uses port 5106 for bi-directional | |
|
communication from the backup/recovery appliance. Make sure your | |
|
firewall permits bi-directional communication through this port. | |
|
2. Ensure that the correct IP is configured for the host **Manage > | |
|
Appliance > Configure Appliance Networking**. | |
|
3. Ensure that the Backup and DR agent service is running on | |
|
the target host and restart, if necessary. | |
|
* On Windows, find the UDS Host Agent service in services.msc and | |
|
click Restart. | |
|
* On Linux, run the command /etc/init.d/udsagent restart
|
|
|
||
|
``` | |
|
<timestamp> GEN-DEBUG [4400] UDSAgent starting up ... | |
|
<timestamp> GEN-INFO [4400] Locale is initialized to C | |
|
<timestamp> GEN-WARN [4400] VdsServiceObject::initialize | |
|
- LoadService for Vds failed with error 0x80080005 | |
|
<timestamp> GEN-WARN [4400] initialize - Failed to initialize Microsoft Disk | |
|
Management Services: Server execution failed [0x80080005] | |
|
<timestamp> GEN-WARN [4400] Failed initializing VDSMgr, | |
|
err = -1, exiting... | |
|
<timestamp> GEN-INFO [4400] Couldn't connect to namespace: root\mscluster | |
|
<timestamp> GEN-INFO [4400] This host is not part of cluster | |
|
<timestamp> GEN-WARN [4400] Failed initializing connectors,exiting -1 | |
|
``` | |
|
||
|
4. Retry the backup. | |
43941
|
Disk space usage on datastore has grown beyond the critical threshold | This issue occurs when the remaining space on the datastore is less than |
|
the critical threshold. If more storage is not made available soon, then | |
|
jobs start to fail when the remaining space is inadequate to store them. | |
|
||
|
This alert is created to help you take action to prevent ESX datastores | |
|
from filling with snapshot data. Increase available space by expanding | |
|
the datastore, migrating some VMs, or deleting old data on the | |
|
datastore. | |
|
||
|
Snapshots grow as more change data is added. If a datastore fills up due | |
|
to a growing snapshot, VMs may be taken offline automatically by VMware | |
|
to protect the data. | |
43954
|
Failed OnVault job | During a mount job, the backup/recovery appliance is unable to connect to |
|
the OnVault pool. This issue can be occurred due to any of the following | |
|
reasons. | |
|
||
|
* No bucket name is specified for the OnVault pool. | |
|
* Invalid credentials-access ID or access key not specified or wrong | |
|
ID for the OnVault pool | |
|
* Invalid bucket in the OnVault pool | |
|
* General authentication issues for the OnVault pool. | |
|
* DNS server in clusters /etc/resolv.conf
is either different or the |
|
|
forward and reverse DNS zones files are changed. | |
43929
|
Snapshot creation of VM failed. Error: VM task failed An error occurred | VM snapshot fails if the ESX server is unable to quiesce the virtual |
|
while saving the snapshot: Failed to quiesce the virtual machine. | machine - either because of too much I/O, or because VMware tools cannot |
|
quiesce the application using VSS in time. Check the event logs on the | |
|
host and check the VM's ESX log (vmware.log). | |
|
||
|
Crash-consistent snapshots and connector-based backups show this | |
|
behavior less often. For more information, refer to the VMware knowledge | |
|
base articles - 1018194 and | |
|
1007696 . | |
43933
|
Failed to find VM with matching BIOS UUID | This issue occurs if the VM's UUID is modified. To fix this issue, |
|
rediscover the VM and check if it was discovered as a new UUID. You can | |
|
confirm this in the management console by comparing the UUID of the | |
|
newly discovered VM and that of the previously discovered VM. If the | |
|
UUIDs don't match, the VM might have been cloned. | |
|
||
|
You can also see this error, if a large number of | |
|
Backup and DR managed VMs are removed from the vCenter. | |
43948
|
The number of images not expired awaiting further processing is (x) | "Event ID 43948 is generated when an application begins halting |
|
images ((x) snapshots, (x) onvaults) from (x) unique | expirations as a part of Image Preservation. 'Image Preservation' |
|
applications. (x) snapshots and (x) OnVaults were added in the | preserves snapshot and OnVault images beyond their expiration dates to |
|
last (x) seconds ((x) hours (x) minutes)., sltname: No | ensure that those images are properly processed by the |
|
specific slt, slpname: No specific slp. | backup/recovery appliance. When a new application enters into a preserved |
|
mode, a Warning alert will be generated. The most common cause of this | |
|
is backup plan violations as documented under event ID 10085". | |
43954
|
Retry OnVault | An OnVault job needed to be retried. Possible issues could include: The |
|
Service Account being used has the wrong role. The Service Account does | |
|
not have authority to write to the bucket. The Cloud Storage bucket | |
|
no longer exists. | |
43960
|
Skipped backing up 6 offline applications for SqlServerWriter application. | Backup of a SQL Server Instance found some databases were offline and |
|
couldn't be backed up. This commonly occurs when the database has been | |
|
deleted on the server side, but is still included on the Backup/DR side. | |
|
The error message contains the names of the offline databases that | |
|
should be investigated. | |
43972
|
Metadata upload to bucket failed. | Metadata writes to an OnVault bucket failed. Possible issues could |
|
include: The Service Account being used has the wrong role. The Service | |
|
Account does not have authority to write to the bucket. The | |
|
Cloud Storage bucket no longer exists. | |
43973
|
udppm started Successfully | This is is an internal event and can be ignored. |
43999
|
Warning: VM is running on a host that is running an outdated version of | Upgrade the VM to a supported version (>=) to ensure the best results. |
|
ESXi , which is not supported by Google. Please upgrade it to a | |
|
supported version (>=) to ensure the best results. | |
44003
|
Succeeded Job_xx-xx-xx
for application application ID
on host |
This is is a successful status event and can be ignored. |
|
host
, sltname: template
, slpname: profile
. |
|
62001
|
Streamsnapd daemon started successfully | This is is an internal event and can be ignored. |
90003
|
A new update (version X) is available for Backup Appliance | A new update is available. Update your backup/recovery appliances as soon |
|
as possible. |
What's next
- To view the event logs created on backup/recovery appliances, see Backup and DR event logs
- To configure log-based alerts for Backup and DR Service, based on sample queries, see create a log-based alert .