Untitled
Untitled
Network Design
Edward Mazurek
Technical Lead Data Center Storage Networking
[email protected]
@TheRealEdMaz
BRKSAN-2883
Agenda
• Introduction
• Technology Overview
• Design Principles
• Storage Fabric Design Considerations
• Data Center SAN Topologies
• Intelligent SAN Services
• Q&A
3
Introduction
6
An Era of Massive Data Growth
Creating New Business Imperatives for IT
By 2020
40% of Data Will Be “Touched” by Cloud
7
Evolution of Storage Networking….
Enterprise Apps: OLTP, VDI, etc. Big Data, Scale-Out NAS Cloud Storage (Object)
Compute Nodes
REST API
Fabric
Fabric
60
• Price/performance making
50
15K rpm drives SSD more affordable
(8 drives)
40 • Solid state drives dramatically
30 Enterprise Flash increase IOPS that a given
Drives (8 drives) array can support
20
15
Fibre Channel – Foundations
Based on SCSI
N_Port-to-N_Port connection
Host Disk
• Logical node connection point (Initiator) (Target)
Flow controlled
• Buffer-to-buffer credits and end-to-end basis Transmitter Receiver
N_port
17
Fibre Channel Addressing
Dual Port HBA Every Fibre Channel port and node has two
10:00:00:00:c9:6e:a8:16 64 bit hard-coded addresses called World
10:00:00:00:c9:6e:a8:17
Wide Names (WWN)
50:0a:09:83:9d:53:43:54 • NWWN(node) uniquely identify devices
• PWWN(port) uniquely identify each port in a
device
• Allocated to manufacturer by IEEE
Host Switch Disk
phx2-9513# show int fc 1/1 • Coded into each device when manufactured
fc1/1 is up
Hardware is Fibre Channel, SFP is short wave laser Switch Name Server maps PWWN to FCID
Port WWN is 20:01:00:05:9b:29:e8:80
4 bits 12 bits 24 bits 24 bits
N-port or IEEE Organizational Unique ID
0002 Locally Assigned Identifier
F_port Identifier (OUI)
Format Identifier Port Identifier Assigned to each vendor Vendor-Unique Assignment
18
Port Initialization – FLOGI and PLOGIGIs/PLOGIs Target
Step 1: Fabric Login (FLOGI)
• Determines the presence or absence of a Fabric 3
FC Fabric
• Exchanges Service Parameters with the Fabric
• Switch identifies the WWN in the service parameters
of the accept frame and assigns a Fibre Channel ID
(FCID)
• Initializes the buffer-to-buffer credits E_Port
Initiator
19
FC_ID Address Model
• FC_ID address models help speed up FC routing
• Switches assign FC_ID addresses to N_Ports
• Some addresses are reserved for fabric services
• Private loop devices only understand 8-bit address (0x0000xx)
• FL_Port can provide proxy service for public address translation
• Maximum switch domains = 239 (based on standard)
8 Bits 8 Bits 8 Bits
Switch
Switch Topology Model Area Device
Domain
Private Loop Device Arbitrated Loop
00 00 Physical Address (AL_PA)
Address Model
Public Loop Device Switch Arbitrated Loop
Area Physical Address (AL_PA)
Address Model Domain
20
FSPF
Fabric Shortest Path First
• Provides routing services within any FC fabric
• Supports multipath routing
• Bases path status on a link state protocol similar to OSPF
• Routes hop by hop, based only on the domain ID
• Runs on E ports or TE ports and provides a loop free topology
• Runs on a per VSAN basis. Connectivity in a given VSAN in a fabric is guaranteed only for the switches
configured in that VSAN.
• Uses a topology database to keep track of the state of the links on all switches in the fabric and
associates a cost with each link
• Fibre Channel standard ANSI T11 FC-SW2
21
FSPF
phx2-5548-3# show fsp database vsan 12
FCoE
IEEE 802.1
T11 DCB
FC on
FC on Other
other
Network
network
Media
PFC ETS DCBX
media
Status Published Fall 2011 Published Fall 2011 Published Fall 2011
Published in May, 2010 25
• VLAN Tag enables 8 priorities
FCoE Flow Control for Ethernet traffic
IEEE 802.1Qbb Priority Flow Control
3.3ms • PFC enables Flow Control on a
Per-Priority basis using
Resume
PAUSE frames (IEEE 802.1p)
• Receiving device/switch sends
Pause frame when receiving
buffer passes threshold
• Two types of pause frames
• Quanta = 65535 = 3.3ms
• Quanta = 0 = Immediate resume
Ethernet Wire
27
FCoE Is Really Two Different Protocols
FIP (FCoE Initialization Protocol) FCoE Itself
• It is the control plane protocol • Is the data plane protocol
• It is used to discover the FC entities • It is used to carry most of the FC
connected to an Ethernet cloud frames and all the SCSI traffic
• It is also used to login to and logout • Ethertype 0x8906
from the FC fabric
• Uses unique BIA on CNA for MAC
• Ethertype 0x8914
The Two Protocols Have
• Two different Ethertypes
• Two different frame formats
• Both are defined in FC-BB-5
28
http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9670/white_paper_c11-560403.html
FPMA - Fabric Provided MAC Address
Fibre Channel over Ethernet Addressing Scheme Domain ID
FC-MAP FC-ID
FPMA (0E-FC-xx) 10.00.01
29
What is an FCoE Switch?
• FCF (Fibre Channel Forwarder) accepts a Fibre Channel frame
encapsulated in an Ethernet packet and forwards that packet
over a VLAN across an Ethernet network to a remote FCoE end FCoE
device Attached
Storage
• FCF is a logical FC switch inside an FCoE switch
• Fibre Channel login happens at the FCF Nexus
• Contains an FCF-MAC address
FCF
• Consumes a Domain ID
FC
30
FCoE is Operationally Identical
• Supports both FC and FCoE
• FCoE is treated exactly the same as FC
• After zoning device perform registration and then performs discovery
33
After Link Is Up, Accessing Storage
FIP and FCoE Login Process
Target
Step 1: FIP Discovery Process FC or FCoE
• Enables FCoE adapters to discover which VLAN to
transmit & receive FCoE frames
Fabric
• Enables FCoE adapters and FCoE switches to discover
other FCoE capable devices
• Occurs over Lossless Ethernet
E_Ports or
Step 2: FIP Login Process VE_Port
• Similar to existing Fibre Channel Login (FLOGI) process
– Sent to upstream FCF VF_Port
• FCF assigns the host a FCID and FPMA to be used for
FCoE forwarding VN_Port
• Returns the FCID and the Fabric Provided MAC Address
(FPMA) to the ENode
FC-MAC CNA
FCIP
TCP TCP
TCP
FC
F Co
oEE IP IP
Lossless
Ethernet
Ethernet Ethernet
Physical Wire
35
Connectivity Types
FC FCoE
N F F N VN VF VF VN
E E
F NP VE VE VF VNP
TE TE
Fabric NPV
E_Port E_Port F_Port NP_Port
Switch Switch
37
The Story of Interface Speeds
Clocking Encoding Data Rate • Comparing speeds is more
Protocol
Gbps Data/Sent Gbps MB/s complex than just the
8G FC 8.500 8b/10b 6.8 850 “apparent” speed
• Data throughput is based on
10G FC 10.51875 64b/66b 10.2 1,275
both the interface clocking
10G FCoE 10.3125 64b/66b 10.0 1,250 (how fast the interface
transmits) and how efficient
16G FC 14.025 64b/66b 13.6 1,700 the interface transmits (how
much encoding overhead)
32G FC 28.050 64b/66b 27.2 3,400
38
38
Design Principles
39
VSANs
Introduced in 2002
• A Virtual SAN (VSAN) Provides a Method to Allocate Ports within a
Physical Fabric and Create Virtual Fabrics
• Analogous to VLANs in Ethernet Per Port Allocation
• Virtual fabrics created from larger cost-effective redundant physical
fabric
• Reduces wasted ports of a SAN island approach
• Fabric events are isolated per VSAN which gives further isolation
for High Availability
• FC Features can be configured on a per VSAN basis.
• ANSI T.11 committee and is now part of Fibre Channel standards
as Virtual Fabrics
40
• Assign ports to VSANs
VSAN • Logically separate fabrics
• Hardware enforced
• Prevents fabric disruptions
• RSCN sent within fabric only
Disk2
2. Configure zones within each VSAN
Disk3 • A zone consists of multiple zone members
Zone A Host1 Disk1
Zone C
Zone B 3. Assign zones to zoneset
Disk4 Host2
• Each VSAN has its own zoneset
Zoneset 1
4. Activate zoneset in VSAN
VSAN 3
• Members in a zone can access each other;
Zone A
Host4 members in different zones cannot access
Zone B
Host3 Disk5 each other
Disk6
Zoneset 1
• Devices can belong to more than one zone
42
Zoning examples
• Non-zoned devices are members of zone name AS01_NetApp vsan 42
the default zone member pwwn 20:03:00:25:b5:0a:00:06
member pwwn 50:0a:09:84:9d:53:43:54
• A physical fabric can have a maximum
of 16,000 zones (9700-only network)
device-alias name AS01
• Attributes can include pWWN, FC pwwn 20:03:00:25:b5:0a:00:06
alias, FCID, FWWN, Switch Interface device-alias name NTAP
member pwwn 50:0a:09:84:9d:53:43:54
fc x/y, Symbolic node name, Device zone name AS01_NetApp vsan 42
alias member device-alias AS01
member device-alias NTAP
43
The Trouble with sizable Zoning
All Zone Members are Created Equal
100
60
10
20
30
40
50
70
80
90
Number of Members
44
Smart Zoning
Operation Today – 1:1Operation
Zoning Today – Many
Operation
- Many Smart Zoning
8xI
Zones Cmds ACLs Zones Cmds ACLs Zones Cmds ACLs
Create Create Create
4xT
32 96 64 1 13 132 1 13 64
zones(s) zones(s) zones(s)
Add an +4 +12 Add +8
an +1 Add+24an +1 +8
initiator initiator initiator
Add a +8 +24 Add a
+16 +1 Add a
+24 +1 +16
target target target
• Allows storage admins to create larger zones while still keeping premise of single initiator & single target
46
Zoning Best Practices
• zone mode enhanced
• Acquires lock on all switches while zoning changes are underway
• Enables full zoneset distribution
• zone confirm-commit
• Causes zoning changes to be displayed during zone commit
• zoneset overwrite-control – New in NX-OS 6.2(13)
• Prevents a different zoneset than the currently activated zoneset from being
inadvertently activated
48
IVR - Inter-VSAN Routing
• Enables devices in different VSANs to
VSAN 2
communicate
Disk2
Zone A
Disk3
Disk1
• Allows selective routing between specific
Host1
Zone C members of two or
Zone B
Disk4 Host2 more VSANs
• Traffic flow between selective devices
Zoneset 1
VSAN1 VSAN1 Traffic engineering with pruning VSANs on/off the trunk
VSAN2 VSAN2
Efficient use of ISL bandwidth
VSAN3 VSAN3
• Physical servers connect to the NPV switch and login to the upstream NPIV core switch
• FC edge switch in NPV mode does not take up a domain ID phx2-9513 (config)# feature npiv
• Helps to alleviate domain ID exhaustion in large fabrics
Server3
FC1/3 F_Port
N_Port_ID 3
52
N-Port
Comparison Between NPIV and NPV
NPIV (N-Port ID Virtualization) NPV (N-Port Virtualizer)
•Used by HBA and FC •Used by FC (MDS 9124, 9148,
9148S, etc.), FCOE switches
switches (Nexus 5K), blade switches and
•Enables multiple logins on a Cisco UCS Fabric InterConnects
single interface (UCS6100)
•Allows SAN to control and •Aggregate multiple physical/logical
monitor virtual machines logins to the core switch
(VMs) •Addresses the explosion of number
of FC switches
•Used for VMWare, MS Virtual
Server and Linux Xen •Used for server consolidation
applications
applications
53
NPV Uplink Selection
NPV supports automatic selection of NP uplinks. When a server interface is brought up, the NP uplink
interface with the minimum load is selected from the available NP uplinks in the same VSAN as the
server interface.
When a new NP uplink interface becomes operational, the existing load is not redistributed automatically
to include the newly available uplink. Server interfaces that become operational after the NP uplink can
select the new NP uplink.
Manual method with NPV Traffic-Maps associates one or more NP uplink interfaces with a server
interface.
Note: Use of parallel NPV links will pin traffic to one NPV link. Use of SAN Portchannels with NPV actual
traffic will be load balanced.
http://www.cisco.com/c/en/us/td/docs/switches/datacenter/nexus5000/sw/configuration/guide/cli_rel_4_0_1a/CLIConfigurationGuide/npv.html#wp1534672
54
NPV Uplink Selection – UCS Example
• NPV uplink selection can be automatic or manual
• With UCS autoselection, the vHBAs will be uniformly assigned to the available
uplinks depending on the number of logins on each uplink
Cisco UCS FC NPIV
Blade Server NPV Switch NP-Port Core Switch
F-Port
FC1/1
FC1/2 F_Port
FC1/3
FC1/4 F_Port
FC1/5
FC1/6
58
Uplink Port Failure
• Failure of an uplink moves pinned hosts from failed port to up port(s)
• Path selection is the same as when new hosts join NPV switch and pathing
decision is made
2 devices re-login
Cisco UCS
Blade Server
FC NPIV
NPV Switch NP-Port
Core Switch
F-Port
FC1/1
FC1/2 F_Port
FC1/3 Port is Down
FC1/4 F_Port
FC1/5
FC1/6
59
Uplink Port Recovery
• No automatic redistribution of hosts to recovered NP port
Cisco UCS
Blade Server FC NPIV
NPV Switch NP-Port
Core Switch
F-Port
FC1/1
FC1/2 F_Port
FC1/3 Port is Up
FC1/4 F_Port
FC1/5
FC1/6
60
New F-Port Attached Host
• New host entering fabric is automatically pinned to recovered NP_Port
• Previously pinned hosts are still not automatically redistributed
Cisco UCS
Blade Server
FC NPIV
NPV Switch NP-Port
Core Switch
F-Port
FC1/1
FC1/2 F_Port
FC1/3
FC1/4 F_Port
FC1/5
FC1/6
61
New NP_Port & New F-Port Attached Host
• NPV continues to distribute new hosts joining fabric
Cisco UCS
Blade Server
FC NPIV
NPV Switch NP-Port
Core Switch
F-Port
FC1/1
FC1/2 F_Port
FC1/3
FC1/4 F_Port
FC1/5
FC1/6 F_Port
New Port Added
62
Auto-Load-Balance
Disruptive load balance works independent of automatic selection of interfaces and a configured traffic map of external
interfaces. This feature forces reinitialization of the server interfaces to achieve load balance when this feature is
enabled and whenever a new external interface comes up. To avoid flapping the server interfaces too often, enable this
feature once and then disable it whenever the needed load balance is achieved.
If disruptive load balance is not enabled, you need to manually flap the server interface to move some of the load to a
new external interface.
http://www.cisco.com/c/en/us/td/docs/switches/datacenter/mds9000/sw/6_2/configuration/guides/interf
aces/nx-os/cli_interfaces/npv.html#pgfId-1072790
63
F-Port Port Channel and F-Port Trunking
Enhanced Blade Switch Resiliency
F-Port Port Channel F-Port Port Channel w/ NPV
VSAN1
Blade N Servers
Blade 2 VSAN2
Separate management domains
Blade 1 VSAN3
Separate fault isolation domains
N-Port F-Port Differentiated services: QoS, Security
64
Port Channeling & Trunking - Configuration
phx2-5548-3# show run interface san-port-channel 1
interface san-port-channel 1
switchport trunk allowed vsan 1
switchport trunk allowed vsan add 12
interface fc2/13
channel-group 1 force Nexus MDS
no shutdown 5548 9148
fc2/13 fc1/1
interface fc2/14 fc2/14 1 fc1/2
channel-group 1 force
no shutdown
D2 D3
67
Port Channeling & Trunking - Configuration
phx2-9148-2# show run interface port-channel 1
interface port-channel1
switchport trunk allowed vsan 1
switchport trunk allowed vsan add 12
interface fc1/1
channel-group 1 force Nexus MDS
no shutdown 5548 9148
fc2/13 fc1/1
interface fc1/2 fc2/14 1 fc1/2
channel-group 1 force
no shutdown
D2 D3
68
Port Channel – Nexus switch config
phx2-5548-3# show run int san-port-channel 3 Nexus
5548
interface san-port-channel 3
channel mode active
switchport mode F
switchport trunk allowed vsan 1
switchport trunk allowed vsan add 12 D2
phx2-5548-3# show run int fc 2/9-10 fc2/10
fc2/9
interface fc2/9 3
switchport mode F
channel-group 3 force
fc2/1 fc2/2
no shutdown
interface fc2/10
switchport mode F Fabric
channel-group 3 force
no shutdown Interconnect
71
Port Channel – FI Config
5548
D2
fc2/9 fc2/10
3
fc2/1 fc2/2
Fabric
Interconnect
72
FLOGI – Before Port Channel
phx2-5548-3# show flogi database 5548
--------------------------------------------------------------------------------
INTERFACE VSAN FCID PORT NAME NODE NAME
--------------------------------------------------------------------------------
fc2/9 12 0x020000 20:41:00:0d:ec:fd:9e:00 20:0c:00:0d:ec:fd:9e:01
fc2/9 12 0x020001 20:02:00:25:b5:0b:00:02 20:02:00:25:b5:00:00:02
fc2/9 12 0x020002 20:02:00:25:b5:0b:00:04 20:02:00:25:b5:00:00:04
D2
fc2/9 12 0x020003 20:02:00:25:b5:0b:00:01 20:02:00:25:b5:00:00:01
fc2/10 12 0x020020 20:42:00:0d:ec:fd:9e:00 20:0c:00:0d:ec:fd:9e:01 fc2/9 fc2/10
fc2/10 12 0x020021 20:02:00:25:b5:0b:00:03 20:02:00:25:b5:00:00:03
fc2/10 12 0x020022 20:02:00:25:b5:0b:00:00 20:02:00:25:b5:00:00:00
Fabric
Interconnect
73
FLOGI- After port channel
phx2-5548-3# show flogi database 5548
--------------------------------------------------------------------------------
INTERFACE VSAN FCID PORT NAME NODE NAME
--------------------------------------------------------------------------------
San-po3 12 0x020040 24:0c:00:0d:ec:fd:9e:00 20:0c:00:0d:ec:fd:9e:01
San-po3 12 0x020001 20:02:00:25:b5:0b:00:02 20:02:00:25:b5:00:00:02
San-po3 12 0x020002 20:02:00:25:b5:0b:00:04 20:02:00:25:b5:00:00:04 D2
San-po3 12 0x020003 20:02:00:25:b5:0b:00:01 20:02:00:25:b5:00:00:01
San-po3 12 0x020021 20:02:00:25:b5:0b:00:03 20:02:00:25:b5:00:00:03 2/9 2/10
San-po3 12 0x020022 20:02:00:25:b5:0b:00:00 20:02:00:25:b5:00:00:00
Fabric
Interconnect
74
Port-channel design considerations
• All types of switches
• Name port-channels the same on both sides
• Common port allocation in both fabrics
• ISL speeds should be >= edge device speeds
• Maximum 16 members per port-channel allowed
• Multiple port-channels to same adjacent switch should be equal cost
• Member of VSAN 1 + trunk other VSANs
• Check TCAM usage:
• show system internal acl tcam-usage
75
port-channel design considerations
• Director class
• Split port-channel members across multiple line cards
• When possible use same port on each LC:
• Ex. fc1/5, fc2/5, fc3/5, fc4/5, etc.
• If multiple members per linecard distribute across port-groups
• show port-resources module x
76
Port-channel design considerations
• Fabric switches
• Ensure enough credits for distance
• Can “rob” buffers from other ports in port-group that are “out-of-service”
• Split port-channel member across different forwarding engines to distribute
ACLTCAM
• For F port-channels to NPV switches (like UCS FIs)
• Each device’s zoning ACLTCAM programming will be repeated on each PC member
• For E port-channels using IVR
• Each host/target session that gets translated will take up ACLTCAM on each member
• Use following table:
• Ex. On a 9148S a six member port-channel could be allocated across the 3 fwd engines as
follows:
• fc1/1, fc1/2, fc1/17, fc1/18, fc1/33 and fc1/34
78
F port-channel design considerations
Fwd Zoning Region Bottom Region
Switch/Module Port Range(s) Fwd-Eng Number
Engines Entries Entries
MDS 9396S 12 1-8 0 49136 19664
9-16 1 49136 19664
17-24 2 49136 19664
25-32 3 49136 19664
33-40 4 49136 19664
41-48 5 49136 19664
49-56 6 49136 19664
57-64 7 49136 19664
65-72 8 49136 19664
73-80 9 49136 19664
81-88 10 49136 19664
89-96 11 49136 19664
79
F port-channel design considerations
Fwd Zoning Region Bottom Region
Switch/Module Port Range(s) Fwd-Eng Number
Engines Entries Entries
DS-X9248-48K9 1 1-48 0 27168 2680
DS-X9248-96K9 2 1-24 0 27168 2680
25-48 1 27168 2680
DS-X9224-96K9 2 1-12 0 27168 2680
13-24 1 27168 2680
DS-X9232-256K9 4 1-8 0 49136 19664
9-16 1 49136 19664
17-24 2 49136 19664
25-32 3 49136 19664
DS-X9248-256K9 4 1-12 0 49136 19664
13-24 1 49136 19664
25-36 2 49136 19664
37-48 3 49136 19664
80
F port-channel design considerations
Fwd Zoning Region Bottom Region
Switch/Module Port Range(s) Fwd-Eng Number
Engines Entries Entries
DS-X9448-768K9 6 1-8 0 49136 19664
9-16 1 49136 19664
17-24 2 49136 19664
25-32 3 49136 19664
33-40 4 49136 19664
41-48 5 49136 19664
81
Internal CRC handling
• New feature to handle frames internally corrupted due to bad HW
• Frames that are received corrupted are dropped at the ingress port
• These frames are not included in this feature
• In rare cases frames can get corrupted internally due to bad hardware
• These are then dropped
• Sometimes difficult to detect
82
Internal CRC handling
• Stages of Internal CRC Detection and Isolation
84
Device-alias
• device-alias(DA) is a way of naming PWWNs
• DAs are distributed on a fabric basis via CFS
• device-alias database is independent of VSANs
• If a device is moved from one VSAN to another no DA changes are needed
• device-alias can run in two modes:
• Basic – device-alias names can be used but PWWNs are substituted in config
• Enhanced – device-alias names exist in configuration natively – Allows rename without
zoneset re-activations
• device-alias are used in zoning, IVR zoning and port-security
• copy running-config startup-config fabric after making changes!
85
Device-alias
• device-alias confirm-commit
• Displays the changes and prompts for confirmation
86
Device-alias
• Note: To prevent problems the same device-alias is only allowed once per
commit.
• Example:
MDS9148s-1(config)# device-alias database
MDS9148s-1(config-device-alias-db)# device-alias name test pwwn 1122334455667788
MDS9148s-1(config-device-alias-db)# device-alias rename test test1
Command rejected. Device-alias reused in current session :test
Please use 'show device-alias session rejected' to display the rejected set of commands and for the
device-alias best-practices recommendation.
87
Cisco Prime Data Center Network Manager
Feature Support and User Interface
VMpath Analysis provides VM connectivity to
network and storage across Unified Compute
and Unified Fabric
• Visibility past physical access (switch)
layer
• Standard & Custom Reports
• On Nexus and MDS platforms
• Dynamic Topology Views
• Rule-based event filtering and
forwarding
• Threshold Alerting
• Integration via vCenter API
88
SAN Design Security Challenges
SAN design security is often overlooked as an area of concern
• Application integrity and security is addressed, but not back-end storage network carrying actual data
• SAN extension solutions now push SANs outside datacenter boundaries
SAN design security is only one part of complete data center solution
• Host access security—one-time passwords, auditing, VPNs
• Storage security—data-at-rest encryption, LUN security
Privilege Escalation/ Theft
External DOS
or Other Unintended Privilege
Unauthorized Data
Intrusion Connections Tampering
Application
Tampering (Internal)
(Trojans, etc.)
SAN
LAN 89
SAN Security
Device/SAN
Secure management access Management
Security Via SSH,
• Role-based access control
SFTP, SNMPv3, and
• CLI, SNMP, and web access User Roles RADIUS or
TACACS+ or LDAP
Secure management protocols Server for
Authentication
• SSH, SFTP, and SNMPv3
91
Storage Fabric Topology
Considerations
92
The Importance of “Architecture”
93
SAN Major Design Factors High
Performance
Port density Crossbar
• How many now, how many later?
2
• Topology to accommodate port
requirements Large Port
QoS, Count
Congestion Directors
Network performance
Control,
• What is acceptable? Unavoidable? Reduce FSPF 3
Routes 1
Traffic management
8 8 8 8 8 8 8 8 8 8 8 8
• Preferential routing or resource allocation
Fault isolation
• Consolidation while maintaining isolation
Management
• Secure, simplified management
4
Failure of One Device Has
No Impact on Others 94
94
Scalability—Port Density
Topology Requirements
Considerations
• Number of ports for end devices
Large Port
• How many ports are needed now? Count
Directors
• What is the expected life of the
SAN?
Best Practice
• Design to cater for future requirements
• Doesn’t imply “build it all now,” but means “cater for it” and
avoids costly retrofits tomorrow
95
Scalability—Port Density – MDS Switch selection
• MDS 9148S – 48 ports 16G FC
• MDS 9250i – 40 ports 16G FC + 8 port 10G FCoE + 2 FCIP ports
• MDS 9396S – 96 ports 16G FC
• MDS 9706 – Up to 192 ports 16G FC and/or 10G FCoE and/or 40G FCoE
• MDS 9710 – Up to 384ports 16G FC and/or 10G FCoE and/or 40G FCoE
• MDS 9718 – Up to 768 ports 16G FC and/or 10G FCoE and/or 40G FCoE
• All MDS 97xx chassis are 32G ready!
• All 16G MDS platforms are full line rate
96
Scalability—Port Density – Nexus Switch selection
• Nexus 55xx – Up to 96 ports 10G FCoE and/or 8G FC ports
• Nexus 5672UP – Up to 48 10G FCoE and/or 16 8G FC ports
• Nexus 5672UP-16G – Up to 48 10G FCoE and/or 24 16G FC ports
• Nexus5624Q – 12 ports 40G or 48 ports 10G FCoE
• Nexus5648Q – 24 ports 40G or 96 ports 10G FCoE
• Nexus5696Q – Up to 32 ports 100G / 96 ports 40G / 384 ports 10G FCoE or 60
8G FC
• Nexus 56128P – Up to 96 10G FCoE and/or 48 8G FC ports
• All Nexus platforms are full line rate
97
Traffic Management
Do different apps/servers have different
performance requirements?
• Should bandwidth be
reserved for specific applications? QoS,
Congestion
• Is preferential treatment/ Control,
QoS necessary? Reduce FSPF
Routes
8 8 8 8 8 8 8 8 8 8 8 8
Given two alternate paths for traffic
between data centers, should traffic
use one path in preference to the other?
• Preferential routes
98
Network Performance
Oversubscription Design Considerations
All SAN Designs Have Some Degree of
Oversubscription
• Without oversubscription, SANs would Tape Oversubscription
be too costly Disk Oversubscription
Disk do not sustain wire-rate I/O Need to sustain close to
with ‘realistic’ I/O mixtures maximum data rate
• Oversubscription is introduced at LTO-6 Native Transfer
Vendors may recommend a 6:1 to
multiple points as high as 20:1 host to disk Rate ~ 160 MBps
fan-out ratio
• Switches are rarely the bottleneck Highly application dependent
in SAN implementations
99
Fault Isolation
Consolidation of Storage
• Single Fabric = Increased Storage Utilization +
Reduced Administration Overhead
Major Drawback
• Faults Are No Longer Isolated Physical SAN Islands Are
Virtualized onto Common
• Technologies such as VSANs enable consolidation SAN Infrastructure
and scalability while maintaining security and
stability
• VSANs constrain fault impacts
Fabric
• Faults in one virtual fabric (VSAN) are contained #3
and do not impact other virtual fabrics Fabric
#1 Fabric
#2
100
Data Center SAN
Topologies
101
Denser Server Cabinets
What are the implications?
Uplinks change from 40 GE servers Vertical Horizontal
to 4x 10G servers Cabling Cabling EoR X-Connect
DC Infrastructure Changes
Denser: cabinets, cross-connects cable runs
From 42U to ~58U Horizontal Cabling: from 10G, through 40G to 100G – longer distances
Vertical Cable: match appropriate server connectivity choice
Is SAN EoR economical now? 102
Structured Cabling
Supporting new EoR & ToR designs
103
Core-Edge
Highly Scalable Network Design
MDS 9710
• Easy expansion
Ports Deployed 3456 per fabric 6,912 total
106
Very Large Edge-Core/End-of-Row Design
“A” Fabric Shown,
Very Large Edge/Core/Edge 576(288 per switch) Repeat for “B” Fabric
(6144 End Device Ports per Fabric) Storage ports at 16Gb
MDS 9718
• Traditional Core-Edge design Is ideal for very
large centralized services and consistent host-
disk performance regardless of location
• Full line rate ports, no fabric oversubscription
• 16Gb hosts and targets
• Services consolidated in the core
• Easy expansion
768(48 per switch) 24
Ports Deployed 12,288 ISLs from host edge to
core @ 16Gb
Used Ports 10,368 @ 16Gb
107
SAN Top of Rack – MDS 9148S
SAN Top of Rack
(5,376 Usable Ports)
352 Storage ports at 16Gb
• Ideal for centralized services while reducing cabling MDS 9710
requirements
• Consistent host/target performance regardless of
location in rack
• 8Gb hosts & 16Gb targets
• Easy edge expansion A B
• Massive cabling infrastructure avoided as compared
to EoR designs 4 ISLs from each
edge to core @ 16Gb
• Additional efficiencies with in rack IO convergence
MDS 9148S
Ports Deployed 5,376
4,224 hosts @ 16Gb
Used Ports 5,344
110
POD SAN Design
POD SAN Design
Ideal for centralized services
36-48 Storage
• Consistent host/target performance regardless of ports at 16Gb
location in blade enclosure or rack
• 10/16Gb hosts & 16Gb targets
• Need to manage more SAN Edge switches/Blade MDS 9396S MDS 9396S
Switches
• NPV attachment reduces fabric complexity A B
• Add blade server ISLs to reduce fabric
6 ISLs from each edge 8 ISLs from each edge
oversubscription to core @ 16Gb to core @ 8Gb
111
252 hosts @ 16Gb or 288 hosts @ 10Gb
FI 6332-16UP, FI 6332 UCS SAN Design
FI 6332-16UP Use Case FI 6332 Use Case
40G 40G
Nexus Nexus
7K/9K 7K/9K
16G FC 40G FCoE
FI 6332-16UP FI 6332
40G 40G 40G 40G
Storage Storage
MDS Array MDS Array
UCS UCS 9700 UCS UCS 9700
B-Series C-Series B-Series C-Series
B200 C220 B200 C220
B260 C240 B260 C240
B460 C460 B460 C460
and and
IOM 2304 IOM 2304
40G 40G
16G FC 40G FCoE
112
Intelligent SAN Services
113
Enhancing SAN Design with Services
Extend Fabrics
• FCIP
• Extended Buffer to Buffer credits
• Encrypt the pipe
IOA IOA
MAN/WAN
IOA IOA
IOA IOA
MAN/WAN
IOA IOA
92% throughput
FCIP increase
Highly resilient– Clustering of IOA engines allows for load balancing and failover
Improved Scalability- Scale without increasing management overhead
Significant reutilization of existing infrastructure- All chassis and common
equipment re-utilized
Flat VSAN topology- Simple capacity and availability planning
118
SAN Extension – FC over long distance
BB_Credits and Distance
~1 km per Frame
2 Gbps FC
16 Km
phx2-9513(config)# feature fcrxbbcredit extended • BB_Credits are used to ensure enough FC frames in flight
phx2-9513(config)# interface 1/1
phx2-9513(config-if)# switchport fcrxbbcredit extended 1000 • A full (2112 byte) FC frame is approx 1 km long @ 2 Gbps,
phx2-9513# show interface 1/1 ½ km long @ 4 Gbps ¼ km long at 8 Gbps
fc1/1 is up
….. • As distance increases, the number of available BB_Credits
Transmit B2B Credit is 128 need to increase as well
Receive B2B Credit is 1000 • Insufficient BB_Credits will throttle performance - no data 119
will be transmitted until R_RDY is returned
SAN Extension – FCoE over long distance
FCoE Flow Control
For long distance FCoE, receiving switch Ingress Buffer must be large enough to absorb all
packets in flight from the time the Pause frame is sent to the to time the Pause Frame is
received
Buffer Threshold
• A 10GE, 50 km link can hold ~300 frames
• That means 600+ frames could be either in flight or will be transmitted by the time the receiver
detects buffer congestion and sends a Pause frame to the time the Pause frame is received and the
sender stops transmitting
Pause
• DMM offers
• Online migration of heterogeneous arrays
Data Mobility Manager
• Simultaneous migration of multiple LUNs
Application Data • Unequal size LUN migration
I/O Migration
• Rate adjusted migration
• Verification of migrated data
• Dual fabric support
• CLI and wizard-based management with Cisco Fabric Manager
• Not metered on no. of terabytes migrated or no. of arrays
Old New • Requires no SAN reconfiguration or rewiring
Array Array
• Uses FC Redirect 121
SAN Extension - CWDM
Course Wavelength Division Multiplexing
TX RX
Transmission
TX RX
TX Optical fiber pair RX
TX RX
Optical OADM Optical
transmitters receivers
• 8 channels WDM using 20nm spacing
• Colored CWDM SFPs used in FC switch
• Optical multiplexing done in OADM
• Passive device
122
SAN Extension - DWDM
Dense Wavelength Division Multiplexing
TX Transmission RX
TX RX
Optical Splitter Protection
TX RX
Optical fiber pair
TX RX
Optical DWDM devices Optical
transmitters receivers
MDS ONS
Array
124
DWDM CWDM
Summary
Drivers in DC are forcing change Many design options
• 10G convergence & server virtualization • Optimized for performance
• It's not just about FCP anymore. FCoE, NFS, iSCSI are • Some for management
being adopted • Others for cable plant optimization
Proper SAN design is holistic in the approach
• Performance, Scale, Management attributes all play critical roles
125
Additional Relevant Sessions
Storage Networking – Cisco Live Berlin
126
Call to Action
• Visit the World of Solutions for:
• Multiprotocol Storage Networking booth
• See the MDS 9718, Nexus 5672UP, 2348UPQ, and MDS 40G FCoE blade
• Data Center Switching Whisper Suite
• Strategy & Roadmap (Product portfolio includes: Cisco Nexus 2K, 5K, 6K, 7K, and MDS products).
• Technical Solution Clinics
• Meet the Engineer
• Available Tuesday and Thursday
128
Complete Your Online Session Evaluation
• Please complete your online session
evaluations after each session.
Complete 4 session evaluations
& the Overall Conference Evaluation
(available from Thursday)
to receive your Cisco Live T-shirt.
129
Thank you
130