Troubleshooting known issues
You should be aware of some known issues that might occur when you use SnapManager, and how to work around them.
SnapManager for SAP fails to identify the profile as a Cluster-Mode profile
If the Cluster-Mode profile name is not present in the cmode_profiles.config
file in the SnapManager for SAP installation directory, the following error message might trigger:
Please configure DFM server using snapdrive config set -dfm user_name appliance_name.
Also, while upgrading the SnapManager for SAP, if you delete the /opt/NetApp/smsap/*
folder, then the cmode_profiles.config
file that has the Cluster-Mode profile names also get deleted. This issue also triggers the same error message.
Workaround
Update the profile: smsap profile update-profile <profile_name>
If SnapManager for SAP is installed in the /opt/NetApp/smsap/ path, then the file location will be /opt/NetApp/smsap/cmode_profile/cmode_profiles.config.
|
The server fails to start
When starting the server, you might see an error message similar to the following:
SMSAP-01104: Error invoking command: SMSAP-17107: SnapManager Server failed to start on port 8074 because of the following errors: java.net.BindException: Address already in use
This might be because the SnapManager listening ports (27314 and 27315, by default) are currently in use by another application.
This error can also occur if the smsap_server
command is already running, but SnapManager does not detect the existing process.
Workaround
You can reconfigure either SnapManager or the other application to use different ports.
To reconfigure SnapManager, edit the following file: /opt/NTAP/smsap/properties/smsap.config
You assign the following values:
-
SMSAP Server.port=27314
-
SMSAP Server.rmiRegistry.port=27315
-
remote.registry.ocijdbc.port= 27315
The remote.registry.ocijdbc.port must be the same as Server.rmiRegistry.port.
To start the SnapManager server, enter the following command: smsap_server start
An error message is displayed if the server is already running. |
If the server is already running, perform the following steps:
-
Stop the server by entering the following command:
smsap_server stop
-
Restart the server by entering the following command:
smsap_server start
Terminating a currently running SnapManager operation
If SnapManager server hangs and you cannot execute any operations successfully, you can terminate SnapManager and its operations.
Workaround
SnapManager works with both SnapManager and Protection Manager. You must perform the following steps to list the different processes running and stop the last process running.
-
List all SnapDrive processes that are running:
ps
ps
Example:
ps | rep snapdrive
-
Stop the SnapDrive process or processes:
kill <pid>
pid
is the list of processes you found using theps
command.Do not stop all SnapDrive processes. You might want to end only the last process that is running. -
If one of the operations involves restoring a protected backup from secondary storage, open the Protection Manager console and perform the following:
-
From the System menu, select Jobs.
-
Select Restore.
-
Check for the name of the dataset that matches the one in the SnapManager profile.
-
Right-click and select Cancel.
-
-
List the SnapManager processes:
-
Log in as a root user.
-
List the processes by using the ps command.
Example:
ps | grep java
-
-
End the SnapManager process:
kill <pid>
Unable to delete or free the last protected backup
When you create the first backup for a profile on secondary storage, SnapManager sends all the information about the backup to Protection Manager. For subsequent backups related to this profile, SnapManager sends only the modified information. If you remove the last protected backup, SnapManager loses the ability to identify the differences between backups and must to find a way to rebaseline these relationships. Therefore, attempting to delete the last protected backup results in an error message being displayed.
Workaround
You can delete the profile or only the profile backup.
To delete the profile, perform the following steps:
-
Delete the profile's backups.
-
Update the profile and disable protection in the profile.
This deletes the dataset.
-
Delete the last protected backup.
-
Delete the profile.
To delete only the backup, perform the following steps:
-
Create another backup copy of the profile.
-
Transfer that backup copy to secondary storage.
-
Delete the previous backup copy.
Unable to manage archive log file destination names if the destination names are part of other destination names
While creating an archive log backup, if the user excludes a destination that is part of other destination names, then the other destination names are also excluded.
For example, assume that there are three destinations available to be excluded: /dest, /dest1,
and /dest2.
While creating the archive log file backup, if you exclude /dest
by using the command
smsap backup create -profile almsamp1 -data -online -archivelogs -exclude-dest /dest
, SnapManager for SAP excludes all the destinations starting with /dest.
Workaround
-
Add a path separator after destinations are configured in
v$archive_dest
. For example, change the/dest
to/dest/
. -
While creating a backup, include destinations instead of excluding any destination.
Restoring control files that are multiplexed on Automatic Storage Management (ASM) and non-ASM storage fails
When the control files are multiplexed on ASM and non-ASM storage, the backup operation is successful. However, when you try to restore control files from that successful backup, the restore operation fails.
SnapManager clone operation fails
When you clone a backup in SnapManager, the DataFabric Manager server might fail to discover volumes, and display the following error message:
SMSAP-13032: Cannot perform operation: Clone Create. Root cause: SMSAP-11007: Error cloning from snapshot: FLOW-11019: Failure in ExecuteConnectionSteps: SD-00018: Error discovering storage for /mnt/datafile_clone3: SD-10016: Error executing snapdrive command "/usr/sbin/snapdrive storage show -fs /mnt/datafile_clone3": 0002-719 Warning: Could not check SD.Storage.Read access on volume filer:/vol/SnapManager_20091122235002515_vol1 for user user-vm5\oracle on Operations Manager servers x.x.x.x
Reason: Invalid resource specified. Unable to find its Id on Operations Manager server 10.x.x.x
This occurs if the storage system has large number of volumes.
Workaround
You must perform one of the following:
-
From the Data Fabric Manager server, run
dfm host discover storage_system
You can also add the command in a shell script file and schedule a job in the DataFabric Manager server to run the script at a frequent interval.
-
Increase the value of
dfm-rbac-retries
in theSnapdrive.conf
file.SnapDrive uses the default refresh interval value and default number of retries. The default value of
dfm-rbac-retry-sleep-secs
is 15 seconds anddfm-rbac-retries
is 12 iterations.The Operations Manager refresh interval depends on the number of storage systems, number of storage objects in the storage system, and the load on the DataFabric Manager server.
As a recommendation, perform the following:
-
From the DataFabric Manager server, manually run the following command for all the secondary storage systems associated with the dataset:
dfm host discover storage_system
-
Double the time taken to perform the host discovery operation and assign that value to
dfm-rbac-retry-sleep-secs
.For example, if the operation took 11 seconds, you can set the value of
dfm-rbac-retry-sleep-secs
to 22 (11*2).
Repository database size grows with time and not with the number of backups
The repository database size grows with time because SnapManager operations insert or delete data within the schema in the repository database tables, which results in high index space usage.
Workaround
You must monitor and rebuild the indexes according to the Oracle guidelines to control the space consumed by the repository schema.
The SnapManager GUI cannot be accessed and SnapManager operations fail when the repository database is down
SnapManager operations fail and you cannot access the GUI when the repository database is down.
The following table lists the different actions you might want to perform, and their exceptions:
Operations | Exceptions |
---|---|
Opening a closed repository |
The following error message is logged in |
Refreshing an opened repository by pressing F5 |
A repository exception is displayed in the GUI and also logs a NullPointerException in the |
Refreshing the host server |
A NullPointerException is logged in the |
Creating a new profile |
A NullPointerException is displayed in the Profile Configuration window. |
Refreshing a profile |
The following SQL exception is logged in |
Accessing a backup |
The following error message is logged in |
Viewing clone properties |
The following error message is logged in |
Workaround
You must ensure that the repository database is running when you want to access the GUI or want to perform any SnapManager operations.
Unable to create temporary files for the cloned database
When temporary tablespace files of the target database are placed in mount points different from the mount point of the data files, the clone create operation is successful but SnapManager fails to create temporary files for the cloned database.
Workaround
You must perform either of the following:
-
Ensure that the target database is laid out so that temporary files are placed in the same mount point as that of the data files.
-
Manually create or add temporary files in the cloned database.
Unable to migrate the protocol from NFSv3 to NFSv4
You can migrate the protocol from NFSv3 to NFSv4 by enabling the enable-migrate-nfs-version
parameter in the snapdrive.conf
file. During the migration, SnapDrive considers only the protocol version, irrespective of the mount point options such as rw, largefiles, nosuid,
and so on.
However, after migrating the protocol to NFSv4, when you restore the backup that was created by using NFSv3, the following occurs:
-
If NFSv3 and NFSv4 are enabled at the storage level, the restore operation is successful but is mounted with the mount point options that were available during backup.
-
If only NFSv4 is enabled at the storage level, the restore operation is successful and only the protocol version (NFSv4) is retained.
However, the other mount point options such as
rw, largefiles, nosuid,
and so on are not retained.
Workaround
You must manually shut down the database, unmount the database mount points, and mount with the options available prior to the restore.
Back up of Data Guard Standby database fails
If any archive log location is configured with the service name of the primary database, the back up of Data Guard Standby database fails.
Workaround
In the GUI, you must clear Specify External Archive Log location corresponding to the service name of the primary database.