NetApp Basic Concepts Quickstart Guide
NetApp Basic Concepts Quickstart Guide
Table of Contents
Storage Terminology
Current NetApp Hardware Disks, RAID4, Aggregates and Space Calculation
Qtrees
LUNs Network Configuration: VIFs
Uptime ServiceDesk
http://now.netapp.com/
serial number = linked to hardware, changes when hardware is replaced system ID = set in software, doesn't change when hardware is replaced
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
NetApp Products
Filer
FAS20x0, FAS30x0, FAS60x0, ... FC disks, SATA disks, SAS disks FC & SAS disks faster (random access), SATA slower (archiving)
Gateway products
NetCache (sold)
FAS6040
840 TB 840 spindles
FAS3070 FAS3040
126 TB 252 spindles 252 TB 504 spindles
FAS6070 FAS6030
420 TB 840 spindles 504 TB 1008 spindles
FAS2050
69 TB 104 spindles
FAS3020
84 TB 168 spindles
FAS3050
168 TB 336 spindles
Data ONTAP
One architecture One application interface One management interface Total interoperability Learn one; know them all
FAS2020
24 TB 40 spindles
N7000
N7700 (FAS6030) 2 Gbps FC ports, 64-bit N7900 (FAS6070) 2 Gbps FC ports, 64-bit N7600 (FAS6040) 4 Gbps FC ports, 64-bit N7800 (FAS6080) 4 Gbps FC ports, 64-bit N5300 (FAS3040) 2 Gbps FC ports, 32-bit N5600 (FAS3070) 2 Gbps FC ports, 32-bit
N5000
N3000
N3700 (FAS270)
FAS200 series
FAS800 series FAS900 series
Present
FAS2000 series (has recently (june 2007) replaced FAS200 series) FAS3000 series (replaces FAS800 & FAS900 series) FAS6000 series (very high-end)
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
Storage Terminology
SAN vs. NAS
FILE FILE
FILE
SCSI
FCAL/FCP (encapsulated SCSI) iSCSI (encapsulated SCSI)
LUN
LUN
SAN (Block)
NetApp FAS
NAS (File)
NAS
Network-Attached Storage File-level data storage connected to a computer network providing data access to heterogeneous network client Computing architecture implemented over a computer network, allows devices to share files and resources Two most commonly used NAS protocols A CIFS server makes data available via shares, a Unix server makes data available via exports CIFS clients typically map a network drive to access data stored on a server, Unix clients typically mount the remote resource
Client/server
Share, export
SAN
Storage Area Network Device from storage manufacturer that provides centralized storage for server systems Logical Unit Number A disk, presented by a SAN, to a host OS (Windows, Unix, ...) that looks like a locally attached disk to the host OS The machine that offers a disk (LUN) to another machine, in other words, the SAN
LUN
Target
Initiator
The machine that expects to see a disk (LUN), in other words, the host OS
Typically, the host will only see LUNs after the appropriate initiator software has been installed, eg. FC/iSCSI drivers
Fabric
One or more fibre channel switches with target(s) and initiator(s) connected to them are referred to as a fabric Well-known verndors of fibre channel switches: Brocade, Cisco (Example on next slide)
HBA
Multipathing (MPIO)
fabric
fabric
Zoning
The partitioning of a fabric (or storage area network) into smaller subsets to restrict interference, add security, and to simplify management. If a SAN hosts several hundred disk drives, each system connected to the SAN doesn't need to see all of them Compare this to VLANs in networking
Put boot disk(s) of server on SAN Requires special HBA features FCP and iSCSI
Table of Contents
NetApp Terminology
Current NetApp Hardware Where to Get Help ?
Snapshots
Qtrees LUNs
NetApp Terminology
Some NetApp-specific Terms ...
Data ONTAP
= Operating system on Network Appliance filers and nearstores, borrows ideas from Unix (BSD)
eg. /etc/ directory on vol0 eg. inodes
Same OS on every Filer/NearStore model, different architectures supported (Intel x86, AMD, 64-bit, MIPS, ... depending on hardware model)
All features are activated through licenses, eg. clustering Recent version: 7.2.5(.1), 7.0.7 Release early, release often-philosophy
RC = release candidate GA = general availability, supported but not fully tested GD = general deployment, factory-installed on machines
Head/filer
Contains motherboard, network connections, fiber connections, console connections, connections to disks, ...
Contain disks
(Disk) Shelf
Motherboard and first disk shelf are integrated (disk shelf can be turned into filer and vice versa)
Disk firmwares
Non-disruptively
Shelf firmwares
Non-disruptively for FCAL shelves Disruptively for (S)ATA shelves
Data ONTAP
Requires reboot
RLM/BMC (Remote LAN Module/Baseboard Management Controller) No reboot required
When to Upgrade ?
(snmp, ndmp)
Most day-to-day activities can be performed via the web interface Command-line interface: not-so-commonly-used commands, eg snap restore + many more commands 2 most commonly used commands: sysconfig & options
Browses through man(ual) pages of command documentation sysconfig (-a, -r, -c, -t, -m) Shows information about filer (hardware, disks, aggregates & RAID groups, ...) options Queries or changes values for various registry options setup Walks through initial setup questions: filer name, IP addresses, etc., but does not erase any data cifs setup Walks through CIFS setup questions: domain/workgroup membership etc. sysstat -x -s 1
Reads a text file and prints contents to console (Unix cat) wrfile Reads from console and sends output to text file (Unix Cat >) snap (restore) Performs snapshot operations, eg. restore from snapshot snapvault snapmirror Manipulates/controls SnapMirror/SnapVault/OSSV operations from the command-line
Controls clustering, eg. enabling/disabling, forcing takeover & giveback aggr Creates/expands/destroys/manipulates aggregates, eg. change options vol Create/resizes/destroys/manipulates volumes, eg. change options
df
Shows free disk space (volumes, aggregates, also inodes) qtree Creates/manipulates qtrees (=special directories)
Creates/destroys/manipulates virtual network interfaces (eg. team interfaces for failover or load-balancing) ifconfig Sets network IP configuration (put in /etc/rc to survive reboots) ifstat Shows network interface statistics netdiag Performs basic network diagnostic testing ndmpd ndmpcopy Manipulates NDMP settings, or use ndmpcopy to copy files via NDMP priv set/priv set advanced/priv set diag Goes into advanced/diagnostics mode
What ?
Provides capability to configure filers to initiate automatic sending of email notifications to NetApp technical support and other designated addressees when specific events occur
Helps solve problems proactively Via email (smtp, preferred) or the web (http(s)) Configure via the web interface ... ... or via the 'options' command can only be changed via command-line
Why ?
How ?
filer> options autosupport.support.transport smtp filer> options autosupport.doit testing123 if email subject contains the word test, Netapp will send a reply to all email recipients
FCP (Fiber) fast, expensive, on all models, originally in filers SATA (Serial ATA) slower, cheaper, on all models, originally on nearstores
SAS (Serial Attached SCSI) fast, expensive, currently only on FAS20x0 series, poised to replace FCP in the long run
Recent models can combine FC, SATA, & SAS disks
Now:
Aggregate
FlexVol1 File1 File2 ...
FlexVol2 LUN
Unix-based, hence terms like inodes, but allows NTFSpermissions (NTFS-security style) Formatting disks ? No: zeroing disks
Aggregate
Aggregate can contain multiple volumes Contain actual data (files and LUNs) Can grow and shrink
Data disks
Spare disks
Parity disks Double parity disks (Broken disks) (Partner disks) When dealing with storage, ONTAP 7 will try to hide as much as possible from this from the storage administrator
On our filer, we will create an aggregate of many data disks + parity. This aggregate can be expanded. On this aggregate we will create flexible volumes that can grow and shrink in size, and that will contain actual data (files and LUNs)
RAID1 (mirroring)
controller
DATA DISK
DATA DISK
RAID5
parity data data data data data parity data date data data data parity data data data data data parity data data data data data parity
NetApp uses RAID4 (or RAID-DP double parity) as the only underlying RAID level (no RAID1 or RAID5)
This is an aggregate
Data Disks
You are STRONGLY encouraged to use RAID-DP instead of RAID-4 for better fauilt tolerance
RAID4 advantages
Combined with WAFL filesystem: on-the-fly expansion of storage (no shrinking) without parity recalculation
Existing parity on disks
Battery-backed RAM (eg. 128 Mb on FAS270), split in two parts (four on cluster)
All write operations are written to NVRAM When 50% of NVRAM is full, flush writes to disk: CP: Consistency Point: spreads out writes over all disks
NVRAM will also be flushed every 10 seconds when not 50% full
writes
NVRAM
Aggregate = collection of RAID-4/RAID-DP disks with parity or double parity, consists of one or more RAID groups
Typically, all disks except spare disks in same aggregate
Exception: different disk sizes are in different aggregates (performance) Exception: disks of different types (eg. FC vs. SATA) Exception: Max 16 Tb raw capacity
When disks are added to the system, aggregates can be expanded on-the-fly Aggregates cannot shrink, however! Aggregates do not contain data directly
You must create flexible volumes (flexvols) inside aggregates. Volumes will contain data
RAID group
parity disk
data disk
aggregate
spare disk
Not like eg. Windows cluster: No shared resources 2 nodes only, model name contains c or ha, eg.
Hardware: a heartbeat cable between both systems Hardware: systems have connection to disks of other system
In case of failure of one system: other system will notice, take over the disks and boot the first system from its disks
CIFS users will get disconnected, but generally client will recover Same story for LUNs: retransmissions Note: NVRAM synchronization on clusters
Every head and every shelf have plastic quick reference cards that can be pulled out from under the front panel eg. disk numbering
Table of Contents
NetApp Products Storage Terminology NetApp Terminology NetApp Hardware Essentials Where to Get Help ? Disks, RAID4, Aggregates and Space Calculation When Disks Go Bad ... Volumes Snapshots Qtrees LUNs Network Configuration: VIFs
Table of Contents
Shelf Modules
LRC (Loop Redundancy Circuit): older hardware ESH, ESH2, ESH4 (Embedded Switched Hub): newer technology, ESH2+ has autotermination AT-FC & AT-FC2: Used in R150/R200, single-path only
Disk Shelves
DS20-ESAS
This is a FAS270 shrunken head module. It transforms a regular shelf into a FAS270 or FAS270c (if 2 FAS270 modules are used)
ESH2 shelf module runs at 2 Gbps, is autoterminated ESH4 shelf module (4 Gbps) looks very similar
1x module
2x modules
shelf in out ID
Data Cables
DB9 (serial or console) cable is needed for connection to a controller It is required during initial setup when there is no network connection to the filer
Ethernet network cables are needed for network connectivity (filer management, iSCSI, NFS, CIFS access, ...)
FCP cables used can be optical or copper Optical cables require the use of SFPs (Small Form-factor Pluggable Transceiver) on filer or shelf Copper is via SFP connections or HSSDC2 for connection of shelves to FAS270
InfiniBand for cluster interconnect, now MTP cable is used with converter
(Disk) Loop
0a
When adding shelves, one can either add shelves to existing loops, or create additional loops
0a
0c
Two FCAL loops, why ? - when backplane speed is maxed out (2 Gbps) - don't mix FC and SATA disks in same loop
ESH2 module ESH module (autotermination) Backplane Speed Switch (1/2/4 Gbps)
ESH
ESH2 (modern)
shelf ID Fibre (FC) connection for tape backup 2x Gigabit Connection NICs, can for serial be teamed console cable (VIF)
shelf ID Fibre (FC) Fibre (FC) 2x Gigabit Connection connection connection NICs, can for serial for SAN & to additional be teamed console cable tape backup disk shelves (VIF)
Connection 4x Fibre (FC) 1x RLM for serial connection NIC console cable for disk (Remote shelves or FC LAN SAN: Module) 0a, 0b, 0c, 0d
6x Gigabit NICs, can be teamed (VIF): e0a, e0b, e0c, e0d, e0e, e0f
8x Fibre (FC) connection for disk shelves or FC SAN: 0a, 0b, 0c, 0d, 0e, 0f, 0g, 0h
Two FCAL loops, why ? - when backplane speed is maxed out (2 Gbps) - don't mix FC and SATA disks in same loop
0a
0c
Note: this is from looking at one machine, it may very well be a clustered system !
Filer Cabling: FAS3020/3050, double connected disk loops for redundancy & load balancing
Two FCAL loops, two filers Who owns disks ? Software-based disk ownership: disk assign command
0a
0c
0a
0c
Four FCAL loops - disks identified on 0a loop of one filer can be seen on 0c loop of partner
Cluster Filer Cabling: FAS3020/3050 Standard Cluster Important Notes about Previous Slide
Software-based ownership is still possible (eg. upgrade from FAS270c): in this case, disks can be literally anywhere in shelves Typically: Hardware-based:
The filers that connects to the TOP module of a shelf controls the disks in that shelf under normal (ie. non-failover) circumstances
Whoever connects to this module, owns the disks in this shelf under normal circumstances
So-called mailbox disks (and backup mailbox disks) act as quorum If upgrading from FAS270c, remove ownership (disk command from maintenance mode) and put disks in correct shelves
0b/0d for additional shelves It's not always 0a/0c Check FC Config Guide on NOW site Different scenarios are possible, eg. 0a/0b & 0c/0d in FC SAN configs! Can get very complicated in dual fabric SAN environments Additional FC cards in PCI(-Express) slots are possible !
For a more thorough overview of all supported cabling configurations, you must read (and follow) the FC Config Guide, available from the NOW site !
So far, we are not protected against complete shelf failures. We need a Metrocluster design to provide this feature
0a
0b 0c
0d
0a
0b 0c
0d
Table of Contents
NetApp Terminology
NetApp Hardware Essentials Where to Get Help ?
Snapshots
Qtrees LUNs
Syslog translator:
https://now.netapp.com/eservice/ems
SnapManager for SQL (latest version) SnapManager for Exchange (latest version)
On the NOW site, look up the toolchest On the NOW site, locate the System Configuration Guide On the NOW site, download the latest firmware for your filer
On the NOW site, locate your NetApp products and serial numbers
On the NOW site, locate the Autosupport analysis for your filer On the NOW site, look up replacements parts for your filer On the NOW site, search the Bugs Online database
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
no control
control
A disk that is 100% full has terrible write performance (fragmentation); by limiting the disk utilization to 90% of its full size, a full WAFL data disk still has somewhat decent performance
WAFL metadata (eg. related to snapshot overhead) is stored in the 10% reserved space Bad blocks do not necessarily mean that a disk should be failed. However, the bad blocks should be relocated to a different disk location WAFL 10% is used
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
RAID4 protects against single-disk failures in a RAID group RAID_DP protects against double-disk failures in a RAID group Plex/shelf mirroring (SyncMirror license, as in Mirrored Cluster or MetroCluster) protects against multiple-disk failures in same RAID group System will use one or more spare disks to recover in the background, while system keeps on running In all other failure scenarios, your only resort for data recovery is to quickly pick up a religion of choice and start praying really hard (I recommend Voodoo has been known to work in
some cases)
If a system is in a state where one more loss of a disk in a RAID group will result in DATA loss for that RAID group (i.e. there are not enough spare disks), the system is in DEGRADED mode and will shut down in 24 hours as a data protection measure
Idem when low NVRAM battery (will not boot until battery is sufficiently reloaded)
Notes:
Timeout is configurable via options raid.timeout (default 24 hours) System with no spares but at least one parity/dparity disk in RAID groups is NOT degraded Again: no spare disks available does not necessarily mean degraded
Normally, failure light should be on If not, make disk LED blink If LED is broken, make disk to the left & right blink
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
Volumes
Contain actual data WAFL (Write Anywhere Filesystem Layout) filesystem (NOT NTFS, ...)
2 types:
Volumes (cont.)
Why Flexvols ?
Virtualization
Volumes (cont.)
Why FlexVols ? (cont.) Regular Volumes
FlexVol Volumes
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
Snapshots
Notes:
Up to 255 snapshots per volume Can be scheduled Maximum space occupied can be specified (default 20%)
Snapshots (cont.)
A Bird's Eye View at Snapshots & SnapRestore
Snapshots
file a
file b
...
file z
D A B C
E F G ? ? H
Free Blocks
Snapshots (cont.)
A Bird's Eye View at Snapshots & SnapRestore
Snapshots
Snapshot
file a
file b
...
file z
file a
file b
...
file z
D A B C
E F G ? ? H
Free Blocks
Snapshots (cont.)
A Bird's Eye View at Snapshots & SnapRestore
Snapshots
Snapshot
file a
file b
...
file z
file a
file b
...
file z
D A B C
E F G ? ? H
Active file system changes = WAFL never overwrites data blocks, old (freed) data blocks are referenced to from the snapshot
Disk Data Blocks
C'
Note that there is no performance impact file a now consists of blocks A, B, C'
Free Blocks
Snapshots (cont.)
A Bird's Eye View at Snapshots & SnapRestore (cont.)
Snapshots (cont.)
A Bird's Eye View at Snapshots & SnapRestore
Snapshots
Snapshot
file a
file b
...
file z
file a
file b
...
file z
D A B C
E F G ? ? H
C'
Free Blocks
Snapshots (cont.)
A Bird's Eye View at Snapshots & SnapRestore
Snapshots
Snapshot
file a
file b
...
file z
file a
file b
...
file z
D A B C
E F G ? ? H
C'
Any changes from after the restored file system (C') are irrevocably lost !
Free Blocks
Snapshots (cont.)
Accessing Snapshots from Clients
NFS clients
.snapshot directory
CIFS clients
Snapshots (cont.)
The Problem of Consistent Snapshots
NAS clients modify files The NetApp filer manages WAFL metadata and buffers in-memory
Clients
WAFL metadata
buffers
FlexVol
file file dir file file
Snapshots (cont.)
The Problem of Consistent Snapshots (cont.)
During snapshot creation, the necessary buffers are flushed to disk, then user I/O is suspended to a volume
Clients
After snapshot creation, user I/O operations to the volume are resumed
Filer Memory GOOD: WAFL will make sure volume meta-data is consistent on-disk
WAFL metadata
buffers
BAD: WAFL ignores any consistency issues in files (eg. open Access databases, .pst files) Do we really care about this ? No, this is the best we can do anyway
Snapshots (cont.)
The Problem of Consistent Snapshots (cont.)
Server SAN complicates things! A server will have its own buffers and filesystem metadata (eg. NTFS on Windows systems)
buffers filesystem metadata
Filer Memory
A WAFL-consistent snapshot will not necessary be consistent from the server's filesystem's point of view, as filesystem metadata and buffers have not been flushed & synced correctly
WAFL metadata
buffers
FlexVol
file file dir file file
Solution: we need some software (driver) on the server to talk to the local filesystem (eg. NTFS) and freeze/thaw it before/after a NetApp snapshot creation
LUN
SnapDrive !
Snapshots (cont.)
The Problem of Consistent Snapshots (cont.)
Server
buffers
filesystem metadata
Talk to NTFS to suspend server I/O during snapshot creation Filer Memory Talk to filer to take a snapshot of the impacted volumes
WAFL metadata
buffers
LUN
Snapshots (cont.)
The Problem of Consistent Snapshots (cont.)
Server
database database buffers & metadata buffers filesystem metadata
Running a database on your SAN complicates things even more if you want to take snapshot backups of your data
A special application that talks to the database's backup API is necessary
Filer Memory
SnapManager !
WAFL metadata
buffers
FlexVol
file file dir file file
LUN
Snapshots (cont.)
The Problem of Consistent Snapshots (cont.)
Server
database
SnapManager
SnapManager talks to SnapDrive to take buffers filesystem SnapDrive talks to metadata snapshots of the LUNs NTFS to suspend containing database(s) server I/O during and transaction logfiles snapshot creation Filer Memory SnapManager packages all SnapDrive talks to filer to take this in an application with a snapshots of affected volumes nice management GUI and WAFL metadata buffers takes care of snapshot Filer takes consistent snapshots management (eg. snapshot of affected volumes renaming & deleting, ...) FlexVol SnapManager performs the steps file dir file described above when backing file file up a database via NetApp LUN snapshots
SnapDrive
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
Qtrees
Security style & oplocks settings can be different than rest of volume
Nowadays, we have FlexVols, can have different security styles & oplocks settings
Less important now BUT: Still important when dealing with replication technologies:
SnapMirror = we can replicate whole volumes OR qtrees SnapVault = we can only replicate qtrees OSSV (Open Systems SnapVault) = we can only replicate directories to qtrees
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
LUNs
Look like big files on WAFL filesystem Are formatted and handled by host OS Mapped via FCP or iSCSI See SnapDrive & SAN course for more info
Table of Contents
Qtrees
LUNs Network Configuration: VIFs
Network Configuration
Give IP address to physical interfaces, or ... ... create VIFs and give IP address to VIF
Single-mode VIF
Multi-mode VIF
Multi-mode VIF:
Table of Contents
NetApp Products Storage Terminology NetApp Terminology NetApp Hardware Essentials Where to Get Help ? Disks, RAID4, Aggregates and Space Calculation When Disks Go Bad ... Volumes Snapshots Qtrees LUNs Network Configuration: VIFs
Replication Technologies
SnapMirror, SnapVault (and OSSV), SyncMirror
Name SnapMirror Type ASync Mirror (> 1 minute) Protocol IP (WAN/LAN) Mode Active/Active Filer Type : Mix of models Distance no limit Solutions Long distance DR Data consolidation
Name SnapVault Type ASync Mirror (> 1 hour) Protocol IP (WAN/LAN) Mode Active/Active Filer Type : Mix of models SV for Open systems (win 2K NT Unix) Distance no limit Solutions disk-to-disk backup,restore HSM
Name SyncMirror Type Synchronous Protocol Fibre Channel or DWDM Mode Active/Active Filer Type : Clustered filers Same models Distance Max. 35 Km. Solutions Real Time replication of data
SnapMirror (Disaster Recovery) SnapVault (Backup/Restore) SyncMirror is not a replication technology (=data redundancy)
Overview
SnapMirror
Replicate:
Volumes Qtrees
SnapVault
Backup technology (restore data from a remote location) Source & destination must be NetApps (Source needs SnapVault Primary license, destination needs SnapVault Secondary license you need two different NetApps ! Asynchronous replication (=periodic updates) Pull based (destination contacts source and asks for changes) Replicate:
Backup technology, see SnapVault Source must be a server (Windows, Unix, Linux)
SyncMirror
Not a replication technology Allows to synchronously mirror an aggregate Used eg. in Metrocluster for data redundancy
NetApp Filer
snapshots