ontap

ONTAP 9 – How to review logs

In Clustered Data ONTAP 8.x and ONTAP 9 there are multiple log files. If you plan to review logs, the good place to start is EMS log file. The Event Management System (EMS) collects events that are noticed by ONTAP. It also provides filtering mechanism for easy review. EMS events can be viewed from the clustershell – You can review node’s log by executing command

cluster1::> event log show 

This command will print all events from the newest ones to the oldest from the current period (last 4000 events). You can also specify a date range or time for the events, for example to print all events from last 10 minutes:

cluster1::> event log show -time >10m

Continue reading

NetApp – Basic Administration (7-mode) screen-cast

It’s been a while since I wrote anything about NetApp. Instead of going deep into any subject I decided to create a ~40 minute screen-cast with some very basic overview of a normal NetApp administration. This video is intended for rather less advanced admins (or even newbies to NetApp world) and I’m focusing mostly on GUI – OnCommand System Manager.

I started with showing how you can access NetApp filer – via GUI (Graphical User Interface) or CLI (Command Line). Later on I give a very short introduction to command line, and move on to the GUI – OnCommand System Manager.

Within GUI I’m starting with creating and explaining what is an aggregate. Later on creating flexible volumeĀ  that resides on this aggregate, qtree and set up some quota. Once this is ready I’m configuring CIFS Server and create CIFS shares plus additional NFS exports. Mounting those from both Windows and Linux host.

At the very end I give a very short introduction to iSCSI LUN as well, but for that topic I promiss to deliver a seperate video.

As always – all comments are most welcomed!

NetApp – Fix the “Bad Label” issue

Recently I came across with a Bad Label error during the change of failed disk. My company changed the support company for one of our systems and the new one sent disks to replace the failed ones from the system. Normally the DC tech make a swap and assign the disks to the system, but this time he called me with an issue (from /etc/messages):

Thu May 22 13:02:54 CEST [NETAPP: raid.config.disk.bad.label:error]: Disk 9.10 Shelf 6 Bay 9 [NETAPP X291_S15KXXX0F15 NA01] S/N [3QQ312Y2XXXPBW] has bad label.
Thu May 22 13:02:54 CEST [NETAPP: raid.config.disk.bad.label:error]: Disk 6.70 Shelf 4 Bay 7 [NETAPP X291_S15KXXX0F15 NA01] S/N [3QQ3097KXXX5VU] has bad label.

To fix the issue I did:

NETAPP> priv set advanced
Warning: These advanced commands are potentially dangerous; use
them only when directed to do so by NetApp
personnel.
NETAPP*> vol status -f

Broken disks

RAID Disk Device HA SHELF BAY CHAN Pool Type RPM Used (MB/blks) Phys (MB/blks)
--------- ------ ------------- ---- ---- ---- ----- -------------- --------------
bad label 6.70 0d 4 7 FC:A 1 FCAL 15000 418000/856064000 420156/860480768
bad label 9.10 0d 6 9 FC:A 1 FCAL 15000 418000/856064000 420156/860480768
NETAPP*> disk unfail -s 6.70
disk unfail: unfailing disk 6.70...
NETAPP*> Fri May 23 08:42:47 CEST [NETAPP: raid.disk.unfail.done:info]: Disk 6.70 Shelf 4 Bay 7 [NETAPP X291_S15XXX0F15 NA01] S/N [3QQ3097KXXX5VU] unfailed, and is now a spare

NETAPP*> disk unfail -s 9.10
disk unfail: unfailing disk 9.10...
NETAPP*> Fri May 23 08:43:04 CEST [NETAPP: raid.disk.unfail.done:info]: Disk 9.10 Shelf 6 Bay 9 [NETAPP X291_S15XXX0F15 NA01] S/N [3QQ312Y2XXXPBW] unfailed, and is now a spare

NETAPP*> vol status -f

Broken disks (empty)
NETAPP*> vol status -s

Pool1 spare disks

RAID Disk Device HA SHELF BAY CHAN Pool Type RPM Used (MB/blks) Phys (MB/blks)
--------- ------ ------------- ---- ---- ---- ----- -------------- --------------
Spare disks for block or zoned checksum traditional volumes or aggregates
spare 6.70 0d 4 7 FC:A 1 FCAL 15000 418000/856064000 420156/860480768 (not zeroed)
spare 9.10 0d 6 9 FC:A 1 FCAL 15000 418000/856064000 420156/860480768 (not zeroed)
NETAPP* > priv set
NETAPP> disk zero spares
NETAPP>