The v1.2.0 release of our Check NetApp-REST product contains a new monitoring plugin called check_netapp_volume usage. It provides the monitoring of:
the used space of each volume (in Bytes) the total used space of several volumes the average used space of several volumes the min/max of used space of several volumes A typical output would look like:
$ check_netapp_volume usage -H filer -w 30GiB -c 50GiB NETAPP VOLUME USAGE OK - 5 volumes checked vserv_b.
The first stable release (v1.1.0) of our brand new Check E-Series product brings an automated health check for e-series nodes.
check_eseries_health will run the health checks regularly and automatically report any errors found.
Example $ ./check_eseries_health --host=10.1.1.125 --system-id 1 [...] NETAPP ESERIES HEALTH CRITICAL - 16 health checks checked, 1 CRITICAL StorageGRID-XG-102.missingVolumes: notCompleted (CRITICAL) StorageGRID-XG-102.integratedHealthCheck: ok StorageGRID-XG-102.dbSubRecordsValidation: ok StorageGRID-XG-102.melEventCheck: ok StorageGRID-XG-102.validPassword: ok StorageGRID-XG-102.failedDrivesPresent: ok StorageGRID-XG-102.exclusiveOperations: ok StorageGRID-XG-102.driveCheck: ok StorageGRID-XG-102.nvsramDisableCfwDownloads: ok StorageGRID-XG-102.
We just pushed the first RC for the upcoming 6.0.0 version to the Distributors Portal. Customers who would like to test are very much appreciated! (Pls. check with the distributor to get access to unstable releases as well, if you do not see them on your account.) The 6.0.0 version brings a completely dockered and therefore dependency free installation plus new checks and options. Please read the release history for details.
Since 4.0.0 every over-all check has a feature, which we call “instance-name contains node-name". We have introduced this as an replacement for the sporadic available --node|--vserver parameter (not available any more). So since then each instance-name has the name of its node prefixed. This allows to use the --include|--exclude parameters to filter for all instances of a specific node.
E.g. if you have a new-node with a different hardware you may want to check only the ports of this new-node:
The 5.3.0 version, released yesterday includes a new check to monitor the EMS(Event Management System)-Logfile. This allows several new checks among them one to send an alarm, if an unusal high number of autogrow-events take place.
Let’s have some examples:
# check myfiler for autoSize events and alarm according to the defaults check_netapp_ems -H myfiler --name=wafl.vol.autoSize.done # same as above, but check only on node 01 check_netapp_ems -H myfiler --node=myfiler-01 --name=wafl.
Our quite new SnapCenter check can filter for resource-groups and policies. Until now one had to precisely define which resource-groups should get checked.
In some situations the resource-group name changes. To allow checking such resource-groups without having to constantly change the filters in the service-check configuration we are introducing regexmatching for these filters.
Example Let me show you an example:
If a resource-group sometimes changes from …
AT01_HA_D01 to AT01_HA_D01_SR (and vice versa),
We have one report, that changing the setup from a chain setup (netapp01 → netapp02 → netapp03) to a star setup (netapp01 → netapp02 ➕ netapp01 → netapp03) makes stoping some collectors (e.g. get_netapp_cm.pl) with an error of “Instance ‘NETAPP01-SVM01:NETAPP01_SVM01_xxxx01_vol’ already exists - can not continue!”
If you see this error in your setup please get in contact with us.
We are happy to announce that the next major release of check_netapp_pro is now available on the Q-Portal for download.
The breaking changes for this release are:
A new directory layout which will ease the creation of RPM-packages. Also we have introduced a new getter for DataONTAP 9.7s new API plus several new checks. Some of these are already available as a compiled binary with zero dependencies. Several of the getters and checks now require a license-file, which you will find at the Q-Portal.
The new UnprotectedVolume check searches for volumes not protected by a SnapMirror.
An alarm is generated if one or more unprotected volumes are found. Specific volumes and vservers can be checked by using the standard check_netapp_pro --include= and --exclude= arguments. A simple example follows using the default arguments and a single host
./check\_netapp\_pro.pl UnprotectedVolume -H filer NETAPP\_PRO UNPROTECTEDVOLUME CRITICAL - 7 volumes checked, 6 critical and 0 warning mycluster-01:vol0: unprotected (CRITICAL) vserv\_a:vol0: unprotected (CRITICAL) vserv\_b:vol0: unprotected (CRITICAL) vserv\_b:vol1: unprotected (CRITICAL) vserv\_b:vol1\_mirror: unprotected (CRITICAL) vserv\_b:vol2: unprotected (CRITICAL) vserv\_a:vol1: protected (vserv\_b:vol1\_mirror, data\_protection) This check is part of the upcoming 5.
The check_netapp_quotas script returns a list of filers where quotas are in
place and where quotas are exceeded.
This script has a nifty new feature, which is available from check_netapp_pro version 5 onwards, to list quota limits per user. Enabling the --list_users switch will make the check_netapp_quotas script print a human-readable view of both soft- and hard-limits, which are overused, per user. An example will make this clearer. First, the standard check_netapp_quotas behaviour: