Article number
000003553
Affected Versions
All
Source Hypervisor
All
Target Hypervisor
All

Error: “An unresolvable error has occurred with VPG {0}. The VPG must be deleted {1}.”

Viewed 858 times

Summary

An administrator notices a VPG is suddenly in an Error status where the error suggests this is unrecoverable and must recreate the VPG.

Root Cause

When reviewing the Recovery VRA logs, the following messages will be present shortly before the time the alert was activated:

ERR: Error in func SfsStream::checkMdEntryValidity - stream<MIRROR-ID,0> Possible corrupt meta-data. Invalid metadata type type=0xb757f31a [UNKNOWN] aborting mirror traversal. SFS Item: type 0xb757f31a [UNKNOWN]:

ERR: Error in func SfsStreamBundle::recover - bundle<MIRROR-ID> Recover returned RC_CORRUPTED_OBJECT. Most likely meta-data corruption in journal.

These errors, as stated right in the log event, reflect corruption in the journal storage itself.

Symptoms

A VPG is in a red error status with the following error:

An unresolvable error has occurred with VPG {0}. The VPG must be deleted {1}

Solution

Follow the below steps to resole the error:

  1. The VPG(s) affected must be recreated.

  2. Follow the Preseeding Volumes KB to recreate the VPG with preseed disks to avoid initial sync.

​​​​​​​NOTE: Preseed to AWS is not supported. Manual preseed to Azure is also not supported (must be done via the Import option in the Zerto Diagnostics Utility).

To resolve the issue permanently, follow the below steps:

  1. The administrator should engage their storage team/vendor to review the datastore that owned the affected journal disk(s) and resolve whatever issue caused the corruption to avoid this from occurring again.

  2. Creating a new datastore to use as a journal datastore could also avoid this problem in future as well.

NOTE: When storage teams/vendors review for datastore issues, it's encouraged to do a deeper review, as this issue has been seen as the result of various storage related components, including physical storage array issues, outdated HBA firmware/drivers on hosts, and SAN connectivity issues.