The following table describes important Backup and DR Service event IDs, event messages, and steps to resolve them:
| Event ID | Event message | What to do |
|---|---|---|
|
5022
|
Actifio Connector: Failed in preparing VSS snapshot set | This issue occurs if Windows fails to create a VSS snapshot. To resolve |
| |
this issue, do the following: | |
| |
||
| |
* Check UDSAgent.log | |
| |
* Check disk space on protected volumes. 300MB may not be enough. | |
| |
* Check Windows Event Logs for VSS related errors. | |
| |
* vssadmin list writers may show writers in a bad state. | |
| |
||
| |
Usually these errors are accompanied by VSS errors reported in the logs | |
| |
such as: VSS_E_VOLUME_NOT_SUPPORTED_BY_PROVIDER
|
|
| |
VSS_E_UNEXPECTED_PROVIDER_ERROR
|
|
| |
||
| |
First check if all the VSS writers are in a stable state by going to the | |
| |
command line and issuing this command. # vssadmin list writers
|
|
| |
||
| |
Check output to confirm that all the writers are in a stable state. | |
| |
||
| |
Restart VSS service and check if the writers are stable. If not you may | |
| |
have to reboot the machine. | |
|
5024
|
Actifio Connector: Failed to create VSS snapshot for backup. Insufficient | This issue occurs if there is insufficient disk space to process a |
| |
storage available to create either the shadow copy storage file or other | snapshot. |
| |
shadow copy data | |
| |
1. Ensure the drive being backed up is not full. | |
| |
2. Check if all the VSS writers are in a stable state From the Windows | |
| |
command line, run: vssadmin list providers
vssadmin list writers
|
|
| |
3. If these services are not running, start them and re-run the job. If | |
| |
the writer's State is Not Stable, restart the VSS service. If the | |
| |
problem continues after restarting the service, reboot the host. | |
| |
||
| |
Sometimes the message appears when internal VSS errors occur. | |
| |
||
| |
Check the Windows Event Logs for VSS related errors. For errors related | |
| |
to VSS, search for related Microsoft patches. Additional VSS | |
| |
troubleshooting details can be found on Microsoft TechNet. | |
| |
||
| |
Microsoft recommends at least 320MB on devices specified for saving the | |
| |
created VSS snapshot, plus change data that is stored there. | |
| |
||
| |
Actifio recommends the shadow storage space be set to unbounded | |
| |
(unlimited) using these commands: vssadmin list shadowstorage
|
|
| |
vssadmin Resize ShadowStorage /On=[drive]: /For=[drive]: / Maxsize=[size]
|
|
| |
||
| |
To change the storage area size in the Windows UI, refer to | |
| |
Configuring Volume Shadow Copy on Windows Server 2008 . | |
| |
||
| |
Re-run the backup once the VSS state is stable and shadow storage is set | |
| |
to unbounded. | |
|
5046
|
Backup staging LUN is not visible to the Actifio Connector | This issue occurs if the staging LUN is not visible to the UDSAgent on |
| |
the application's host and the host is unable to detect the staging LUN | |
| |
from the backup/recovery appliance. | |
|
5049
|
Actifio Connector: Failed identifying logical volume on the backup staging | Actifio Connector couldn't see the staging LUN. This can be caused by a |
| |
lun | bad connection or by trouble on the LUN. |
| |
||
| |
Verify that FC/iSCSI connectivity is good, then make sure it works by | |
| |
mapping the VDisk, partitioning it, formatting it, and copying files to | |
| |
it. The steps for partitioning and formatting are OS specific. | |
|
5078
|
Actifio Connector: The staging disk is full | Jobs fail if a file that was modified in the source disk is copied to |
| |
the staging disk, but the file is larger than the free space available | |
| |
in the staging disk. To fix the issue with full staging disk, increase | |
| |
the staging disk. Specify the size of the staging disk in the advanced | |
| |
settings for the application. Set the value for staging disk size such | |
| |
that it is greater than the sum of size of the source disk and the size | |
| |
of the largest file. Note: Changing the staging disk in advanced | |
| |
settings provokes a full backup. | |
|
5087
|
Actifio Connector: Failed to write files during a backup (Source File) | Anti-virus programs or third party drivers may have applied file locks |
| |
that cannot be overridden. | |
| |
||
| |
Check the UDSAgent.log to see which file couldn't be accessed. Attempt | |
| |
to find which process is locking the file using lsof on Unix/ Linux, or | |
| |
fltmc on Windows. Exclude the file from the antivirus or capture job | |
| |
and re-try the capture. | |
| |
||
| |
The current processes known to Microsoft are listed at: | |
| |
Allocated filter altitudes . | |
| |
||
| |
These errors are rarely found on Unix or Linux, but it is possible that | |
| |
a process such as database maintenance or patch install / update has | |
| |
created an exclusive lock on a file. | |
| |
||
| |
Install the latest Actifio Connector. | |
| |
||
| |
A file system limitation or inconsistency was detected by the host | |
| |
operating system. | |
| |
||
| |
Run the Windows Disk defragmenter on the staging disk. | |
| |
||
| |
Low I/O throughput from the hosts disks or transport medium, iSCSI or | |
| |
FC. | |
| |
||
| |
Ensure there are no I/O issues in the host's disks or transport medium. | |
| |
The transport medium will either be iSCSI or Fibre Channel depending on | |
| |
out of band configuration. Consult storage and network administrators as | |
| |
needed. | |
|
5131 - SQL Logs report error 3041
|
SQL log backups on instance fail with error 5131 | To resolve this, enable "Don't forcefully unload the user registry at |
| |
user logoff", see User Profile Service functionality . | |
|
5131 - SQL logs show backup/ recovery appliances error 43901
|
Snapshot jobs fail with error 5131, SQL logs show backup/ recovery | This is because the ODBC login for the database is failing. Fixing the |
| |
"Failed snapshot Job" | ODBC login resolves the issue. |
| |
appliances error 43901 | |
|
5136
|
Actifio Connector: The staging volume is not readable | Check /act/logs/UDSAgent.log for details and contact Google support for |
| |
the resolution for the issue. | |
|
5241
|
Actifio Connector: Failed to mount/clone applications from mapped image | Invalid username and password being parsed from the control file. |
| |
(Source File) | On the source, review the UDSAgent.log to see if the source is |
| |
configured with the correct username/password under Advanced Settings in | |
| |
the connector properties. | |
|
5547
|
Oracle: Failed to backup archivelog (Source File) | Actifio Connector failed to backup the archive log using RMAN archive |
| |
backup commands. The likely causes for this failure are: | |
| |
||
| |
* Connector failed to establish connection to database | |
| |
* The archive logs were purged by another application | |
| |
* TNS Service name is configured incorrectly, causing backup command | |
| |
to be sent to a node where the staging disk isn't mounted | |
| |
||
| |
Search for ORA- or RMAN- errors in the RMAN log. This is the error | |
| |
received from Oracle. Use the preferred Oracle resource as these are not | |
| |
Backup and DR Service conditions, and hence cannot be resolved within | |
| |
Backup and DR Service. | |
| |
* Actifio Connector logs: /var/act/log/UDSAgent.log | |
| |
* Oracle RMAN logs: /var/act/log/********_rman.log | |
|
10032
|
Snapshot pool exceeded warning level | To reduce consumption of the snapshot pool, do the following: |
| |
||
| |
* Move VMware VMs from a snapshot to a Direct-to-OnVault backup plan. | |
| |
Then expire all snapshots to release the space used by the staging | |
| |
disks and last snap. This only works for VMware VMs; other | |
| |
application types still use some snapshot pool space if protected by | |
| |
a Direct-to-OnVault policy. | |
| |
* Reduce the number of snaps kept for an application by changing the | |
| |
policy template. Applications that have high change rates create | |
| |
larger snapshots, so this has the highest benefit for high | |
| |
change-rate applications. This does not necessarily lead to a | |
| |
different RPO, an OnVault images of each snap can be created before | |
| |
they are expired. | |
| |
* Delete mounts, clones, and live-clones if they are not needed. | |
|
10038
|
About to exceed VDisk warning limit | To immediately reduce VDisk consumption, do the following: |
| |
||
| |
* Ensure expirations are enabled, both at the global and individual | |
| |
application level. | |
| |
* Group databases from a single host together into a Consistency Group. | |
| |
For example, if a host has 9 databases, create one Consistency Group | |
| |
for all 9 databases, then protect that consistency group rather than | |
| |
the individual databases. | |
| |
* Reduce the number of snapshots kept for an application by changing | |
| |
the policy template used by a backup plan. | |
| |
* Delete unwanted mounts, clones, and live-clones images. | |
| |
* Move VMware VMs from a snapshot to a Direct- OnVault backup plan. | |
| |
You need to expire all snapshots to release the VDisks used by the | |
| |
staging disks. This only lowers the VDisk count for VMware VMs, and | |
| |
still uses VDisks when protected by a direct-to-OnVault policy. | |
| |
* Change VMware VMDKs that don't need to be protected to independent | |
| |
mode as these cannot be protected by VMware snapshots. | |
| |
||
| |
If this alert repeats daily but the appliance does not reach the maximum | |
| |
VDisks, then modify the policies to reduce the number of VDisks used, or | |
| |
increase the alert threshold. During a daily snapshot window the VDisk | |
| |
count can fluctuate while new VDisks are created for snapshots before | |
| |
the old VDisks are removed as a part of snapshot expirations. The daily | |
| |
fluctuations varies depending on the number of applications protected. | |
|
10039
|
Network error reaching storage device | A heartbeat ping to monitored storage has failed due to hardware |
| |
failure or network issue. Check the network to resolve the issue. | |
|
10043
|
A backup plan violation has been detected | Review the backup plan best practices |
| |
and optimize your policies. The are common causes for backup plan | |
| |
violations. | |
| |
||
| |
* Job scheduler is not enabled. See to enable the scheduler . | |
| |
* The first jobs for new applications can often take a long time: | |
| |
Long job times can occur during the first snapshot or dedup job | |
| |
for an application. On-ramp settings can be used to prevent | |
| |
ingest jobs from locking up slots and locking out ingested | |
| |
applications. See [Set priorities for the first new | |
| |
applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | |
| |
* Applications are inaccessible due to network issues. | |
| |
* Policy windows are too small or job run times are too long: While | |
| |
you cannot control how long each job takes to run, you can control | |
| |
the schedule time for applications that are running. Jobs that run | |
| |
for many hours occupy job slots that could be used by other | |
| |
applications. Review the [backup plan best | |
| |
practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | |
| |
and adjust policies accordingly. | |
| |
* Replication process sends the data to a remote | |
| |
backup/recovery appliance. Ensure that the bandwidth & utilization of | |
| |
your replication link is not saturated. | |
|
10046
|
Performance Pool exceeded safe threshold | To reduce consumption of the snapshot pool, do the following: |
| |
||
| |
* Move VMware VMs from a snapshot to a Direct-to-OnVault backup plan. | |
| |
Then expire all snapshots to release the space used by the staging | |
| |
disks and last snap. This only works for VMware VMs; other | |
| |
application types still use some snapshot pool space if protected by | |
| |
a Direct-to-OnVault policy. | |
| |
* Reduce the number of snaps kept for an application by changing the | |
| |
policy template. Applications that have high change rates create | |
| |
larger snapshots, so this has the highest benefit for high | |
| |
change-rate applications. This does not necessarily lead to a | |
| |
different RPO, an OnVault images of each snap can be created before | |
| |
they are expired. | |
| |
* Delete mounts, clones, and live-clones if they are not needed. | |
|
10055
|
Unable to check remote protection | Each backup/recovery appliance checks the remote appliance hourly for |
| |
possible remote protection issues. The appliance communication fails due | |
| |
to the following issues: | |
| |
||
| |
* Network error (temporary or permanent). Temporary network error does | |
| |
not mean job to fail; jobs are retried, but the hourly check is not | |
| |
updated. | |
| |
* Certificate error. To fix the certificate error, you need to | |
| |
re-exchange the certificate. | |
|
10070
|
Udppm scheduler is off for more than 30 minutes. | The scheduler is off. This may have been set for maintenance. If the |
| |
maintenance is complete, you can re-enable the scheduler, see to | |
| |
enable the scheduler . | |
|
10084
|
Alert for application (app name) and policy (policyname) job did not run | Review the backup plan best practices |
| |
because of unknown reason | and optimize your policies. The are common causes for backup plan |
| |
violations. | |
| |
||
| |
* Job scheduler is not enabled. See to enable the scheduler . | |
| |
* The first jobs for new applications can often take a long time: | |
| |
Long job times can occur during the first snapshot or dedup job | |
| |
for an application. On-ramp settings can be used to prevent | |
| |
ingest jobs from locking up slots and locking out ingested | |
| |
applications. See [Set priorities for the first new | |
| |
applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | |
| |
* Applications are inaccessible due to network issues. | |
| |
* Policy windows are too small or job run times are too long: While | |
| |
you cannot control how long each job takes to run, you can control | |
| |
the schedule time for applications that are running. Jobs that run | |
| |
for many hours occupy job slots that could be used by other | |
| |
applications. Review the [backup plan best | |
| |
practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | |
| |
and adjust policies accordingly. | |
| |
* Replication process sends the data to a remote | |
| |
backup/recovery appliance. Ensure that the bandwidth & utilization of | |
| |
your replication link is not saturated. | |
|
10085
|
Backup Plan violation for application (app name) on host (host name) and | Review the backup plan best practices |
| |
policy (policy name). Job did not run because of unknown reason. | and optimize your policies. The are common causes for backup plan |
| |
violations. | |
| |
||
| |
* Job scheduler is not enabled. See to enable the scheduler . | |
| |
* The first jobs for new applications can often take a long time: | |
| |
Long job times can occur during the first snapshot or dedup job | |
| |
for an application. On-ramp settings can be used to prevent | |
| |
ingest jobs from locking up slots and locking out ingested | |
| |
applications. See [Set priorities for the first new | |
| |
applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | |
| |
* Applications are inaccessible due to network issues. | |
| |
* Policy windows are too small or job run times are too long: While | |
| |
you cannot control how long each job takes to run, you can control | |
| |
the schedule time for applications that are running. Jobs that run | |
| |
for many hours occupy job slots that could be used by other | |
| |
applications. Review the [backup plan best | |
| |
practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | |
| |
and adjust policies accordingly. | |
| |
* Replication process sends the data to a remote | |
| |
backup/recovery appliance. Ensure that the bandwidth & utilization of | |
| |
your replication link is not saturated. | |
|
10120
|
Psrv started successfully | This is an internal event and can be ignored. |
|
10220
|
NTP Service is not running or not synchronised. | The NTP Service on the backup appliance is not running. The NTP Service |
| |
is needed to ensure the backup appliance uses the correct timestamps. A | |
| |
Compute Engine appliance should be using metadata.google.internal. | |
| |
Follow how to set the NTP server [DNS and NTP | |
| |
method](/backup-disaster-recovery/docs/concepts/appliance-network-settings#dns-ntp). | |
|
10225
|
Udp corefiles are found, filename udpengine.(file name) | Internal processes are unexpectedly logging error files. Contact Google |
| |
support to get the resolution for this issue. | |
|
10229
|
Exceeded storage, System name: (device name) | This is is an internal event and normally can be ignored. |
|
10237
|
The X job has been running for more than 3 hours. | There are many reasons why a job can run for more than 3 hours. |
|
11001
|
Backup appliance certificate expires in X days. Power on for 24 hours to | The backup/recovery appliance's certificate last update is more than 15 |
| |
renew, or contact support. | days old. If the backup/recovery appliance is down, bring it up. |
|
11004
|
System components are down. Please contact support in case backups are | Contact support. |
| |
affected. | |
|
11006
|
Not able to sync with host X, regular sync with host required to avoid | The certificate on the host hasn't been updated for more than 7 days. |
| |
permanent communication loss between Backup appliance and Host. | Try rebooting and reconnecting to the host. |
|
20019
|
Insufficient CPU / Memory. Minimum number of core required: (cores) Actual | Backup/recovery appliance has been changed and is not the recommended |
| |
cores : (cores). Minimum memory size required (GB): (memory) Actual memory : | size. Contact Google support to get the resolution for this issue. |
| |
(memory) | |
|
20025
|
Swap usage exceeded | This issue occurs when the swap usage is exceeding the configured |
| |
threshold limit that is set for the backup/recoevery appliance. Contact | |
| |
Google support to get the resolution for this issue. | |
|
20030
|
tomcat stopped successfully | This is is an internal event and can be ignored. |
|
20031
|
tomcat started successfully | This is is an internal event and can be ignored. |
|
22001
|
OMD started Successfully, sltname: , slpname: . | This is is an internal event and can be ignored. |
|
42356
|
File changes have been detected no deleted files have been detected new | This is is an internal event and can be ignored. |
| |
files have been detected. | |
|
43151
|
couldn't add raw device mappings to virtual machine (VM). Error: VM task | Adding a raw device mapping to a VM "stuns" the VM until ESX has had a |
| |
failed A general system error occurred: The system returned an error. | chance to add the new resource. To find out why the raw device mapping |
| |
couldn't be added, look at the ESX logs for the VM in question | |
| |
(vmware.log). | |
| |
||
| |
Refer to the VMware documentation and knowledge base for assistance on | |
| |
reviewing the logs for error messages. Also, review the VMware article | |
| |
for more information on [collecting VMware | |
| |
logs](https://kb.vmware.com/s/article/653). | |
|
43155
|
Error: VM task failed. An error occurred while saving the snapshot: Failed | This is a VMware issue; for additional information, refer to VMware KB |
| |
to quiesce the virtual machine. | article - 1015180 . |
| |
||
| |
Virtual machine quiesce issues are dependent on the OS type. Additional | |
| |
investigation, further VMware KBA searches or contact VMware support to | |
| |
resolve this issue. | |
|
43155 - a
|
Error: VM task failed. Device scsi3 couldn't be hot-added. | This usually means that the SCSI device you are trying to add to the VM |
| |
is already in use by another VM. | |
|
43155 - b
|
Error: VM task failed. The virtual disk is either corrupted or not a | This issue occurs if the VM's CTK files are locked, unreadable, or are |
| |
supported format. | being committed. To fix this issue, remove and re-create these CTK |
| |
files. Refer to the KB article - 2013520 | |
| |
for more information. | |
|
43155 - c
|
Error: VM task failed. The operation is not allowed in the current state | There are two options for formatting a VMware datastore: NFS and VMFS. |
| |
of the datastore." progress ="11" status="running" | With NFS, there are some limitations like not being able to do RDM (Raw |
| |
Disk Mapping). This means that you cannot mount from the | |
| |
backup/recovery appliance to an NFS datastore. Refer to the following KB | |
| |
article - 1001856 for | |
| |
additional information. | |
|
43175
|
UDSAgent socket connection got terminated abnormally; while waiting for | The Actifio Connector stops responding between the appliance and a host |
| |
the response from agent | with Backup and DR agent is installed. |
| |
||
| |
1. Restart the UDSAgent Backup and DR agent service on the | |
| |
specified host. | |
| |
2. Telnet to tcp port 5106 (UDSAgent communication port) | |
| |
||
| |
``` | |
| |
<P></P> | |
| |
Expected output: | |
| |
||
| |
<P></P> | |
| |
Trying 10.50.100.67... | |
| |
||
| |
<P></P> | |
| |
Connected to dresx2.accu.local. | |
| |
||
| |
<P></P> | |
| |
Escape character is '^]'. | |
| |
||
| |
<P></P> | |
| |
Connection closed by foreign host. | |
| |
``` | |
| |
||
| |
3. Verify network connectivity between appliance and host doesn't drop. | |
| |
If the problem persists, network analysis will be required. | |
|
43604
|
Failed to verify fingerprint | This occurs when an inconsistency is found between the source and target |
| |
data. Contact Google support to get the resolution for this issue. | |
|
43690
|
Host doesn't have any SAN or iSCSI ports defined. | This issue occurs if the backup/recovery appliance is not configured with |
| |
iSCSI connection to the target host. | |
| |
||
| |
Ensure that the network ports are open for iSCSI and the target host has | |
| |
discovered the backup/recovery appliances. | |
|
43698
|
ESX host is not accessible for NBD mode data movement | The backup/recovery appliance is unable to reach the ESX host over the |
| |
network or resolve the ESX hostname using DNS. Contact Google support to | |
| |
get the resolution for this issue. | |
|
43702
|
Backup was aborted because there are too many extra files in the home | This is an alert condition generated by Backup and DR Service and is caused |
| |
directory of the VM | by leftover delta files in the VM's datastore. Normally, the delta files |
| |
are removed after Backup and DR snapshot is consolidated. In | |
| |
some instances, these can be left behind by the VMware consolidation, | |
| |
and Backup and DR begins failing jobs to prevent exacerbating | |
| |
the issue. | |
| |
||
| |
This issue is caused by VMware, refer to the knowledge base article - | |
| |
1002310 . | |
|
43755
|
Failed to open VMDK volume; check connectivity to ESX server. | This happens when the ESX server cannot be reached by the controller, |
| |
usually because of a physical connection or DNS problem. To fix this | |
| |
issue, do the following: | |
| |
||
| |
* Ensure port 902 is open between the backup/recovery appliance and the | |
| |
ESX host. | |
| |
* Check the current DNS server and ensure it is current and valid. | |
| |
* If the vCenter is virtualized, attempt a backup after migrating the | |
| |
vCenter to a different ESX host. | |
| |
* Ensure SSL required is set to True on the ESX host in the advanced | |
| |
settings. | |
|
43844
|
Invalid size vmdk detected for the VM | There are two possible solutions for this situation: |
| |
||
| |
* If consolidation is required for some disks on VM, size is reported | |
| |
as zero. To fix this issue, creating and deleting a snapshot of the | |
| |
VM. | |
| |
* See if the VMDK can be restored from a backup image. | |
|
43873
|
Disk space usage on datastore has grown beyond the critical threshold | This issue occurs when the remaining space on the datastore is less than |
| |
the critical threshold. If more storage is not made available soon, then | |
| |
jobs start to fail when the remaining space is inadequate to store them. | |
| |
||
| |
This alert is created to help you take action to prevent ESX datastores | |
| |
from filling with snapshot data. Increase available space by expanding | |
| |
the datastore, migrating some VMs, or deleting old data on the | |
| |
datastore. | |
| |
||
| |
Snapshots grow as more change data is added. If a datastore fills up due | |
| |
to a growing snapshot, VMs may be taken offline automatically by VMware | |
| |
to protect the data. | |
|
43900
|
Retry pending OnVault (log) (jobname for application (appname) on host | Job retries can be caused by many errors. Each 43900 event message |
| |
(hostname) Error: (errorID) (Error Description) | includes an error code and an error message. |
|
43901
|
Job failure | Job failures can be caused by many errors. Each 43901 event message |
| |
includes an error code and an error message. | |
|
43903
|
Failed expire job | This issue occurs when the image is in use at the time of the |
| |
expiration. This can be due to the image is in use by another process or | |
| |
operation, such as a mount, clone, restore. The expiration job most | |
| |
likely complete successfully on the second attempt. | |
| |
Backup and DR does not report the successful completion of | |
| |
this second attempt. If you get only one error for an image, it is safe | |
| |
to conclude that a second attempt to expire this image was successful. | |
| |
If there is a legitimate reason why this image cannot be expired, you | |
| |
will get multiple errors related to this image. If you receive more than | |
| |
one error, contact Google Support. | |
|
43905
|
Failed mount job | There are many ways a mount job can fail. The error code that |
| |
accompanies the event helps to identify the root cause. | |
|
43908
|
Failed restore job | Job failures can be caused by many errors. Each 43908 event message |
| |
includes an error code and an error message. | |
|
43915
|
Couldn't connect to backup host. Make sure Backup and DR agent | To initiate backup, the Actifio Connector service must be reachable by |
| |
is running on (host) and network port (port) is open | the backup/recovery appliance. This issue occurs, when the required ports |
| |
are not open, the incorrect host IP is configured, the | |
| |
Backup and DR agent service not running, or the host is out | |
| |
of physical resources. To fix this issue, do the following: | |
| |
||
| |
1. Ensure that the port in use between the host, | |
| |
backup/recovery appliance, and Actifio Connector is open. By default, | |
| |
the Backup and DR agent uses port 5106 for bi-directional | |
| |
communication from the backup/recovery appliance. Make sure your | |
| |
firewall permits bi-directional communication through this port. | |
| |
2. Ensure that the correct IP is configured for the host **Manage > | |
| |
Appliance > Configure Appliance Networking**. | |
| |
3. Ensure that the Backup and DR agent service is running on | |
| |
the target host and restart, if necessary. | |
| |
* On Windows, find the UDS Host Agent service in services.msc and | |
| |
click Restart. | |
| |
* On Linux, run the command /etc/init.d/udsagent restart
|
|
| |
||
| |
``` | |
| |
<timestamp> GEN-DEBUG [4400] UDSAgent starting up ... | |
| |
<timestamp> GEN-INFO [4400] Locale is initialized to C | |
| |
<timestamp> GEN-WARN [4400] VdsServiceObject::initialize | |
| |
- LoadService for Vds failed with error 0x80080005 | |
| |
<timestamp> GEN-WARN [4400] initialize - Failed to initialize Microsoft Disk | |
| |
Management Services: Server execution failed [0x80080005] | |
| |
<timestamp> GEN-WARN [4400] Failed initializing VDSMgr, | |
| |
err = -1, exiting... | |
| |
<timestamp> GEN-INFO [4400] Couldn't connect to namespace: root\mscluster | |
| |
<timestamp> GEN-INFO [4400] This host is not part of cluster | |
| |
<timestamp> GEN-WARN [4400] Failed initializing connectors,exiting -1 | |
| |
``` | |
| |
||
| |
4. Retry the backup. | |
|
43941
|
Disk space usage on datastore has grown beyond the critical threshold | This issue occurs when the remaining space on the datastore is less than |
| |
the critical threshold. If more storage is not made available soon, then | |
| |
jobs start to fail when the remaining space is inadequate to store them. | |
| |
||
| |
This alert is created to help you take action to prevent ESX datastores | |
| |
from filling with snapshot data. Increase available space by expanding | |
| |
the datastore, migrating some VMs, or deleting old data on the | |
| |
datastore. | |
| |
||
| |
Snapshots grow as more change data is added. If a datastore fills up due | |
| |
to a growing snapshot, VMs may be taken offline automatically by VMware | |
| |
to protect the data. | |
|
43954
|
Failed OnVault job | During a mount job, the backup/recovery appliance is unable to connect to |
| |
the OnVault pool. This issue can be occurred due to any of the following | |
| |
reasons. | |
| |
||
| |
* No bucket name is specified for the OnVault pool. | |
| |
* Invalid credentials-access ID or access key not specified or wrong | |
| |
ID for the OnVault pool | |
| |
* Invalid bucket in the OnVault pool | |
| |
* General authentication issues for the OnVault pool. | |
| |
* DNS server in clusters /etc/resolv.conf
is either different or the |
|
| |
forward and reverse DNS zones files are changed. | |
|
43929
|
Snapshot creation of VM failed. Error: VM task failed An error occurred | VM snapshot fails if the ESX server is unable to quiesce the virtual |
| |
while saving the snapshot: Failed to quiesce the virtual machine. | machine - either because of too much I/O, or because VMware tools cannot |
| |
quiesce the application using VSS in time. Check the event logs on the | |
| |
host and check the VM's ESX log (vmware.log). | |
| |
||
| |
Crash-consistent snapshots and connector-based backups show this | |
| |
behavior less often. For more information, refer to the VMware knowledge | |
| |
base articles - 1018194 and | |
| |
1007696 . | |
|
43933
|
Failed to find VM with matching BIOS UUID | This issue occurs if the VM's UUID is modified. To fix this issue, |
| |
rediscover the VM and check if it was discovered as a new UUID. You can | |
| |
confirm this in the management console by comparing the UUID of the | |
| |
newly discovered VM and that of the previously discovered VM. If the | |
| |
UUIDs don't match, the VM might have been cloned. | |
| |
||
| |
You can also see this error, if a large number of | |
| |
Backup and DR managed VMs are removed from the vCenter. | |
|
43948
|
The number of images not expired awaiting further processing is (x) | "Event ID 43948 is generated when an application begins halting |
| |
images ((x) snapshots, (x) onvaults) from (x) unique | expirations as a part of Image Preservation. 'Image Preservation' |
| |
applications. (x) snapshots and (x) OnVaults were added in the | preserves snapshot and OnVault images beyond their expiration dates to |
| |
last (x) seconds ((x) hours (x) minutes)., sltname: No | ensure that those images are properly processed by the |
| |
specific slt, slpname: No specific slp. | backup/recovery appliance. When a new application enters into a preserved |
| |
mode, a Warning alert will be generated. The most common cause of this | |
| |
is backup plan violations as documented under event ID 10085". | |
|
43954
|
Retry OnVault | An OnVault job needed to be retried. Possible issues could include: The |
| |
Service Account being used has the wrong role. The Service Account does | |
| |
not have authority to write to the bucket. The Cloud Storage bucket | |
| |
no longer exists. | |
|
43960
|
Skipped backing up 6 offline applications for SqlServerWriter application. | Backup of a SQL Server Instance found some databases were offline and |
| |
couldn't be backed up. This commonly occurs when the database has been | |
| |
deleted on the server side, but is still included on the Backup/DR side. | |
| |
The error message contains the names of the offline databases that | |
| |
should be investigated. | |
|
43972
|
Metadata upload to bucket failed. | Metadata writes to an OnVault bucket failed. Possible issues could |
| |
include: The Service Account being used has the wrong role. The Service | |
| |
Account does not have authority to write to the bucket. The | |
| |
Cloud Storage bucket no longer exists. | |
|
43973
|
udppm started Successfully | This is is an internal event and can be ignored. |
|
43999
|
Warning: VM is running on a host that is running an outdated version of | Upgrade the VM to a supported version (>=) to ensure the best results. |
| |
ESXi , which is not supported by Google. Please upgrade it to a | |
| |
supported version (>=) to ensure the best results. | |
|
44003
|
Succeeded Job_xx-xx-xx
for application application ID
on host |
This is is a successful status event and can be ignored. |
| |
host
, sltname: template
, slpname: profile
. |
|
|
62001
|
Streamsnapd daemon started successfully | This is is an internal event and can be ignored. |
|
90003
|
A new update (version X) is available for Backup Appliance | A new update is available. Update your backup/recovery appliances as soon |
| |
as possible. |
What's next
- To view the event logs created on backup/recovery appliances, see Backup and DR event logs
- To configure log-based alerts for Backup and DR Service, based on sample queries, see create a log-based alert .

